hexsha
string
size
int64
ext
string
lang
string
max_stars_repo_path
string
max_stars_repo_name
string
max_stars_repo_head_hexsha
string
max_stars_repo_licenses
list
max_stars_count
int64
max_stars_repo_stars_event_min_datetime
string
max_stars_repo_stars_event_max_datetime
string
max_issues_repo_path
string
max_issues_repo_name
string
max_issues_repo_head_hexsha
string
max_issues_repo_licenses
list
max_issues_count
int64
max_issues_repo_issues_event_min_datetime
string
max_issues_repo_issues_event_max_datetime
string
max_forks_repo_path
string
max_forks_repo_name
string
max_forks_repo_head_hexsha
string
max_forks_repo_licenses
list
max_forks_count
int64
max_forks_repo_forks_event_min_datetime
string
max_forks_repo_forks_event_max_datetime
string
content
string
avg_line_length
float64
max_line_length
int64
alphanum_fraction
float64
794d386805036f115e96769717b98ec508bf7e53
1,178
py
Python
tests/openbb_terminal/stocks/technical_analysis/test_finbrain_model.py
tehcoderer/GamestonkTerminal
54a1b6f545a0016c576e9e00eef5c003d229dacf
[ "MIT" ]
255
2022-03-29T16:43:51.000Z
2022-03-31T23:57:08.000Z
tests/openbb_terminal/stocks/technical_analysis/test_finbrain_model.py
tehcoderer/GamestonkTerminal
54a1b6f545a0016c576e9e00eef5c003d229dacf
[ "MIT" ]
14
2022-03-29T14:20:33.000Z
2022-03-31T23:39:20.000Z
tests/openbb_terminal/stocks/technical_analysis/test_finbrain_model.py
tehcoderer/GamestonkTerminal
54a1b6f545a0016c576e9e00eef5c003d229dacf
[ "MIT" ]
24
2022-03-29T15:28:56.000Z
2022-03-31T23:54:15.000Z
# IMPORTATION STANDARD # IMPORTATION THIRDPARTY import pytest import requests # IMPORTATION INTERNAL from openbb_terminal.stocks.technical_analysis import finbrain_model @pytest.mark.vcr def test_get_technical_summary_report(recorder): result = finbrain_model.get_technical_summary_report(ticker="PM") recorder.capture(result) @pytest.mark.vcr(record_mode="none") def test_get_technical_summary_report_invalid_status(mocker): mock_response = requests.Response() mock_response.status_code = 400 mocker.patch(target="requests.get", new=mocker.Mock(return_value=mock_response)) result = finbrain_model.get_technical_summary_report(ticker="TSLA") assert result == "" @pytest.mark.vcr(record_mode="none") def test_get_technical_summary_report_invalid_json(mocker): mock_response = requests.Response() mock_response.status_code = 200 mocker.patch.object( target=mock_response, attribute="json", side_effect=lambda: dict(), ) mocker.patch(target="requests.get", new=mocker.Mock(return_value=mock_response)) result = finbrain_model.get_technical_summary_report(ticker="TSLA") assert result == ""
29.45
84
0.767402
794d38d417a827f9f66891884a9c1a8ab3cfd49c
338
py
Python
Article/admin.py
yqji/DJNeverDie
6126ca4093f8d06b5a044c330e832432fec0a0c2
[ "Apache-2.0" ]
null
null
null
Article/admin.py
yqji/DJNeverDie
6126ca4093f8d06b5a044c330e832432fec0a0c2
[ "Apache-2.0" ]
null
null
null
Article/admin.py
yqji/DJNeverDie
6126ca4093f8d06b5a044c330e832432fec0a0c2
[ "Apache-2.0" ]
null
null
null
from django.contrib import admin # Register your models here. from .models import Article class ArticleAdmin(admin.ModelAdmin): list_display = ('title', 'author', 'channel', 'views', 'comments') search_fields = ('title', 'author', 'channel') list_filter = ('channel', 'status') admin.site.register(Article, ArticleAdmin)
24.142857
70
0.707101
794d39043eb2435ed937a555210ffa15f3729d2f
3,330
py
Python
hy_corporate_bond_curve_prediction_early/hy_corporate_bond.py
chinthojuprajwal/IE517_ML_course
4b9a1eae9d5100d3573607c46d0fa37fe35074a9
[ "Unlicense" ]
null
null
null
hy_corporate_bond_curve_prediction_early/hy_corporate_bond.py
chinthojuprajwal/IE517_ML_course
4b9a1eae9d5100d3573607c46d0fa37fe35074a9
[ "Unlicense" ]
null
null
null
hy_corporate_bond_curve_prediction_early/hy_corporate_bond.py
chinthojuprajwal/IE517_ML_course
4b9a1eae9d5100d3573607c46d0fa37fe35074a9
[ "Unlicense" ]
null
null
null
import pandas as pd import numpy as np import seaborn as sns import matplotlib.pyplot as plt df=pd.read_csv('D:/UIUC_courses/IE517/IE517_FY21_HW3/HY_Universe_corporate_bond.csv') print(df.describe()) df=df.values df[:,14]=df[:,14].astype(str) df[:,12]=df[:,12].astype(str) df[:,5:9]=df[:,5:9].astype(str) df[:,15]=df[:,15].astype(np.float) score=np.zeros(len(df[:,1])) avg_score=np.zeros(len(df[:,1])) count=np.zeros(len(df[:,1])) for i in range(len(df[:,1])): for j in range(4): if df[i,j+5][0]=='A': df[i,j+5]=2 count[i]=count[i]+1 score[i]=score[i]+2; elif df[i,j+5][0]=='B': df[i,j+5]=1 count[i]=count[i]+1 score[i]=score[i]+1; elif df[i,j+5][0]=='C': df[i,j+5]=0 count[i]=count[i]+1 score[i]=score[i]+0; else: df[i,j+5]='NaN' avg_score[i]=(score[i])/count[i] volume_traded_score_raw=np.transpose(np.array([avg_score,df[:,22]])) volume_traded_score=[] #volume_traded_score1=volume_traded_score[:, ~np.isnan(volume_traded_score[:,0].astype(np.float))] for i in range(len(df[:,1])): if ~np.isnan(avg_score[i]): volume_traded_score.append([round(avg_score[i],2)*5,df[i,22]]) volume_traded_score_np=np.array(volume_traded_score) volume_traded_score_df=pd.DataFrame(np.array(volume_traded_score),columns=['score','volume_traded']) sns.scatterplot(data=volume_traded_score_df,x="volume_traded",y="score") plt.xlabel("volume_traded") plt.ylabel("Average rating between 0 and 10") plt.show() coupon_score_raw=np.transpose(np.array([avg_score,df[:,9]])) coupon_score=[] #coupon_score1=coupon_score[:, ~np.isnan(coupon_score[:,0].astype(np.float))] for i in range(len(df[:,1])): if ~np.isnan(avg_score[i]): coupon_score.append([round(avg_score[i],2)*5,df[i,9]]) coupon_score_np=np.array(coupon_score) coupon_score_df=pd.DataFrame(np.array(coupon_score),columns=['score','coupon']) sns.scatterplot(data=coupon_score_df,x="coupon",y="score") plt.xlabel("coupon") plt.ylabel("Average rating between 0 and 10") plt.show() In_ETF_score_raw=np.transpose(np.array([avg_score,df[:,19]])) In_ETF_score=[] #In_ETF_score1=In_ETF_score[:, ~np.isnan(In_ETF_score[:,0].astype(np.float))] for i in range(len(df[:,1])): if ~np.isnan(avg_score[i]): In_ETF_score.append([round(avg_score[i]*5,2),df[i,19]]) In_ETF_score_np=np.array(In_ETF_score) In_ETF_score_df=pd.DataFrame(np.array(In_ETF_score),columns=['score','In_ETF']) sns.swarmplot(data=In_ETF_score_df,x="In_ETF",y="score") plt.xlabel("In_ETF") plt.ylabel("Average rating between 0 and 10") plt.show() sector_score_raw=np.transpose(np.array([avg_score,df[:,14]])) sector_score=[] #sector_score1=sector_score[:, ~np.isnan(sector_score[:,0].astype(np.float))] for i in range(len(df[:,1])): if ~np.isnan(avg_score[i]): sector_score.append([round(avg_score[i]*5,2),df[i,14]]) sector_score_np=np.array(sector_score) sector_score_df=pd.DataFrame(np.array(sector_score),columns=['score','sector']) g = sns.FacetGrid(sector_score_df, col="sector",col_wrap=8) g.map(sns.histplot, "score") plt.xlabel("sector") plt.ylabel("Average rating between 0 and 10") plt.show()
33.979592
100
0.654655
794d3af425e416e2939c54c9c56f391f4e882fca
634
py
Python
app/manage.py
justTill/whereToLunch
555ffd9c987cbbb17bd4b69ab48fe23b58691bbd
[ "Unlicense" ]
1
2020-05-21T15:12:19.000Z
2020-05-21T15:12:19.000Z
app/manage.py
justTill/whereToLunch
555ffd9c987cbbb17bd4b69ab48fe23b58691bbd
[ "Unlicense" ]
10
2020-04-08T14:44:17.000Z
2021-06-10T20:00:10.000Z
app/manage.py
justTill/whereToLunch
555ffd9c987cbbb17bd4b69ab48fe23b58691bbd
[ "Unlicense" ]
null
null
null
#!/usr/bin/env python """Django's command-line utility for administrative tasks.""" import os import sys def main(): os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'LightningLunch.settings') try: from django.core.management import execute_from_command_line except ImportError as exc: raise ImportError( "Couldn't import Django. Are you sure it's installed and " "available on your PYTHONPATH environment variable? Did you " "forget to activate a virtual environment?" ) from exc execute_from_command_line(sys.argv) if __name__ == '__main__': main()
28.818182
78
0.68612
794d3b0d117ecaad7916dadae5a0e633c09ef3b5
413
py
Python
museosMadrid/migrations/0007_usuario_count.py
AlbertoCoding/X-Serv-Practica-Museos
5d3e3c99b8750ece9973f4e04ae3c3bfe77f3946
[ "Apache-2.0" ]
null
null
null
museosMadrid/migrations/0007_usuario_count.py
AlbertoCoding/X-Serv-Practica-Museos
5d3e3c99b8750ece9973f4e04ae3c3bfe77f3946
[ "Apache-2.0" ]
null
null
null
museosMadrid/migrations/0007_usuario_count.py
AlbertoCoding/X-Serv-Practica-Museos
5d3e3c99b8750ece9973f4e04ae3c3bfe77f3946
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ('museosMadrid', '0006_usuario_titulo'), ] operations = [ migrations.AddField( model_name='usuario', name='count', field=models.PositiveSmallIntegerField(default=0), ), ]
20.65
62
0.615012
794d3b1a0ea829069972129a15a1fefb5e4b428c
6,779
py
Python
pydyn/operations/geometry.py
vkotaru/pydyn_on_manifolds
7ce6e4f4be5baff6422b3f8e8523fa644adc85f5
[ "BSD-3-Clause" ]
2
2021-02-20T15:19:37.000Z
2021-07-30T03:56:57.000Z
pydyn/operations/geometry.py
vkotaru/pydyn_on_manifolds
7ce6e4f4be5baff6422b3f8e8523fa644adc85f5
[ "BSD-3-Clause" ]
null
null
null
pydyn/operations/geometry.py
vkotaru/pydyn_on_manifolds
7ce6e4f4be5baff6422b3f8e8523fa644adc85f5
[ "BSD-3-Clause" ]
null
null
null
from abc import ABC from pydyn.operations.transpose import Transpose from pydyn.base.matrices import MatrixExpr from pydyn.base.scalars import ScalarExpr, Scalar from pydyn.base.vectors import VectorExpr, Vector from pydyn.base.expr import Expr, Expression from pydyn.base.nodes import UnaryNode, BinaryNode from pydyn.utils.errors import UndefinedCaseError, ExpressionMismatchError class Delta(UnaryNode, Expr, ABC): """ Variation operator """ def __init__(self, expr): super().__init__() self.expr = expr self.type = expr.type def __str__(self): return '\delta{' + self.expr.__str__() + '}' def __mul__(self, other): from pydyn.operations.multiplication import Mul, SVMul, SMMul, VVMul, MVMul, MMMul if type(other) == float or type(other) == int: other = Scalar('(' + str(other) + ')', value=other, attr=['Constant']) if self.type == Expression.SCALAR: if other.type == Expression.SCALAR: return Mul(self, other) elif other.type == Expression.VECTOR: return SVMul(other, self) elif other.type == Expression.MATRIX: return SMMul(other, self) else: raise UndefinedCaseError elif self.type == Expression.VECTOR: if other.type == Expression.SCALAR: return SVMul(self, other) elif other.type == Expression.VECTOR: return VVMul(self, other) elif other.type == Expression.MATRIX: return MVMul(self, other) else: return UndefinedCaseError elif self.type == Expression.MATRIX: if other.type == Expression.SCALAR: return SMMul(self, other) elif other.type == Expression.VECTOR: if type(other) == type(Transpose(None)): raise ExpressionMismatchError else: return MVMul(self, other) elif other.type == Expression.MATRIX: return MMMul(self, other) else: raise UndefinedCaseError else: raise UndefinedCaseError def __add__(self, other): return self.expr.__add__(other) def diff(self): return Delta(self.expr.diff()) def integrate(self): return Delta(self.expr.integrate()) class Dot(BinaryNode, ScalarExpr, ABC): """ Dot product of vectors """ def __init__(self, l, r): super().__init__() if l.type == Expression.VECTOR and r.type == Expression.VECTOR: self.left = l self.right = r self.type = Expression.SCALAR else: raise ExpressionMismatchError self.isConstant = self.left.isConstant and self.right.isConstant if self.left.isZero or self.right.isZero: self.isConstant = True self.isZero = self.left.isZero or self.right.isZero def __str__(self): return 'Dot(' + self.left.__str__() + ',' + self.right.__str__() + ')' def delta(self): from pydyn.operations.multiplication import MVMul from pydyn.operations.addition import Add if isinstance(self.right, MVMul): if self.right.left.isSymmetric and self.left == self.right.right: return Add(Dot(self.left.delta(), self.right) * 2, Dot(self.left, MVMul(self.right.left.delta(), self.right.right))) else: return Add(Dot(self.left.delta(), self.right), Dot(self.left, self.right.delta())) else: if self.left.isConstant and not self.right.isConstant: return Dot(self.left, self.right.delta()) elif not self.left.isConstant and self.right.isConstant: return Dot(self.left.delta(), self.right) elif self.left.isConstant and self.right.isConstant: return Vector('0', attr=['Constant', 'Zero']) elif self.left == self.right: return Dot(self.left.delta(), self.right) * 2 else: return Add(Dot(self.left.delta(), self.right), Dot(self.left, self.right.delta())) class Cross(BinaryNode, VectorExpr, ABC): """ Cross product of 3x1 vectors """ def __init__(self, l, r): super().__init__() if l.type == Expression.VECTOR and r.type == Expression.VECTOR: self.left = l self.right = r self.type = Expression.VECTOR else: raise ExpressionMismatchError self.isZero = self.left.isZero or self.left.isZero def __str__(self): return 'Cross(' + self.left.__str__() + ',' + self.right.__str__() + ')' def delta(self): if self.left.isConstant and not self.right.isConstant: return Cross(self.left, self.right.delta()) elif not self.left.isConstant and self.right.isConstant: return Cross(self.left.delta(), self.right) elif self.left.isConstant and self.right.isConstant: return Vector('0', attr=['Constant', 'Zero']) else: return Cross(self.left.delta(), self.right) + Cross(self.left, self.right.delta()) def diff(self): if self.left.isConstant and not self.right.isConstant: return Cross(self.left, self.right.diff()) elif not self.left.isConstant and self.right.isConstant: return Cross(self.left.diff(), self.right) elif self.left.isConstant and self.right.isConstant: return Vector('0', attr=['Constant', 'Zero']) else: return Cross(self.left.diff(), self.right) + Cross(self.left, self.right.diff()) class Hat(UnaryNode, MatrixExpr, ABC): """ Hat map: Rn to LieGroup G """ def __init__(self, expr): super().__init__() if expr.type == Expression.VECTOR: self.expr = expr self.type = Expression.MATRIX else: from pydyn.utils.errors import ExpressionMismatchError raise ExpressionMismatchError def __str__(self): return 'Hat(' + self.expr.__str__() + ')' def delta(self): return Hat(self.expr.delta()) class Vee(UnaryNode, MatrixExpr, ABC): """Vee map: LieGroup G to Rn""" def __init__(self, expr): super().__init__() if expr.type == Expression.MATRIX: # ideally this should be skew-symmetric matrix self.expr = expr self.type = Expression.VECTOR else: from pydyn.utils.errors import ExpressionMismatchError raise ExpressionMismatchError def __str__(self): return 'Vee(' + self.expr.__str__() + ')'
35.492147
98
0.591238
794d3c082b095d0e80de3985176595571316a9aa
2,636
py
Python
setup.py
Syslog777/psak
da9fd58a2e90adf8882019a04ddbdc2eede872cd
[ "BSD-3-Clause" ]
28
2017-12-12T12:16:13.000Z
2022-02-23T04:34:35.000Z
setup.py
Syslog777/psak
da9fd58a2e90adf8882019a04ddbdc2eede872cd
[ "BSD-3-Clause" ]
7
2017-12-12T11:53:20.000Z
2020-01-25T10:13:40.000Z
setup.py
Syslog777/psak
da9fd58a2e90adf8882019a04ddbdc2eede872cd
[ "BSD-3-Clause" ]
10
2018-01-02T20:30:07.000Z
2020-05-06T20:26:08.000Z
""" Copyright (c) 2017, Syslog777 All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of Desktop nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. """ from setuptools import find_packages from setuptools import setup """ https://docs.python.org/3/distutils/setupscript.html#installing-package-data EX. setup.py src/ mypkg/ __init__.py module.py data/ tables.dat spoons.dat forks.dat setup(..., packages=['mypkg'], package_dir={'mypkg': 'src/mypkg'}, package_data={'mypkg': ['data/*.dat']}, ) """ """ src=src mypkg=psak_core """ setup( name='psak', version='0.6.0.1', packages=find_packages('psak_src'), # include all packages under psak_src package_dir={'': 'psak_src'}, # tell distutils packages are under psak_src entry_points={"console_scripts": ["psak=psak_src.psak:main"]}, include_package_data=True, # include everything in source control, url='https://github.com/Syslog777/psak', license='BSD 3', author='Syslog777', author_email='thejavapi26@gmail.com', description='The Pentester\'s Swiss Army Knife ', keywords=["pentesting", "psak", "dos", "ddos", "ultimate", "hacking"] ) print(find_packages())
37.126761
81
0.72003
794d3e8f2fcf0033d4dd979bf67d7256308b97e7
508
py
Python
addLabelToStories.py
marcosdiez/pytracker
82b88e74600ed69b8401aee51e7e924ba72ceec5
[ "Apache-2.0" ]
1
2016-05-11T19:41:43.000Z
2016-05-11T19:41:43.000Z
addLabelToStories.py
marcosdiez/pytracker
82b88e74600ed69b8401aee51e7e924ba72ceec5
[ "Apache-2.0" ]
1
2016-05-30T16:45:16.000Z
2016-05-30T16:45:16.000Z
addLabelToStories.py
marcosdiez/pytracker
82b88e74600ed69b8401aee51e7e924ba72ceec5
[ "Apache-2.0" ]
2
2015-12-17T14:49:33.000Z
2018-11-26T11:33:46.000Z
#!/usr/bin/env python3 import sys from pytracker import Tracker import settings def main(): if len(sys.argv) < 3: print("usage {} label story1 [story2] [story N]".format(sys.argv[0])) return label = sys.argv[1] stories = sys.argv[2:] tracker = Tracker(settings.project_id, settings.token) for story_id in stories: story = tracker.GetStory(story_id) story.AddLabel(label) tracker.UpdateStory(story) print("{}-{}", story.GetStoryId(), story.GetName()) if __name__ == "__main__": main()
20.32
71
0.69685
794d3f8b6524d862dd15c774af71ba72157e8fc9
9,048
py
Python
streamCrawler.py
rishivinod7/TwitterCrawler
c1f4594b03e2472aabf7d56e397a796e4a868ae1
[ "Unlicense" ]
null
null
null
streamCrawler.py
rishivinod7/TwitterCrawler
c1f4594b03e2472aabf7d56e397a796e4a868ae1
[ "Unlicense" ]
null
null
null
streamCrawler.py
rishivinod7/TwitterCrawler
c1f4594b03e2472aabf7d56e397a796e4a868ae1
[ "Unlicense" ]
null
null
null
import tweepy import json from pymongo import MongoClient from datetime import datetime import time import sys import emoji import re #Author: 2331751v, Rishi Vinod # please put your credentials below - very important consumer_key = "" consumer_secret ="" access_token ="" access_token_secret ="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret ) auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) if (not api): print('Can\'t authenticate') print('failed cosumeer id ----------: ', consumer_key ) # set DB DETAILS # this is to setup local Mongodb client = MongoClient('localhost',27017) #is assigned local port dbName = "TwitterDump" # set-up a MongoDatabase db = client[dbName] collName = 'TweetColl' # here we create a collection dataColl = 'DataColl' collection = db[collName] # This is for the Collection put in the DB coll2 = db[dataColl] rt_count = 0 # tweetcount = db.collection.count() # print ("number of tweets collected:", tweetcount) def strip_emoji(text): # copied from web - don't remember the actual link new_text = re.sub(emoji.get_emoji_regexp(), r"", text) return new_text def cleanList(text): # copied from web - don't remember the actual link #remove emoji it works text = strip_emoji(text) # Remove URLs text = re.sub(r'(https|http)?:\/\/(\w|\.|\/|\?|\=|\&|\%)*\b', '', str(text), flags=re.MULTILINE) # Remove new lines text = re.sub(r'/\r?\n|\r/', '', str(text), flags=re.MULTILINE) # Remove broken symbols text = re.sub(r'&amp;', '', str(text), flags=re.MULTILINE) text.encode("ascii", errors="ignore").decode() return text def processTweets(tweet): # this module is for cleaning text and also extracting relevant twitter feilds # initialise placeholders place_countrycode =None place_name =None place_country =None place_coordinates =None source =None exactcoord =None place=None rt = False is_image = False is_vid = False media_type =None media_url = None quoted = False # print(t) # Pull important data from the tweet to store in the database. try: created = tweet['created_at'] quoted = tweet['is_quote_status'] tweet_id = tweet['id_str'] # The Tweet ID from Twitter in string format username = tweet['user']['screen_name'] # The username of the Tweet author verified = tweet['user']['verified'] text = tweet['text'] # The entire body of the Tweet entities = tweet['entities'] source = tweet['source'] geoenabled = tweet['user']['geo_enabled'] location = tweet['user']['location'] exactcoord = tweet['coordinates'] except Exception as e: # if this happens, there is something wrong with JSON, so ignore this tweet print("exc",e) return None try: # // deal with truncated if(tweet['truncated'] == True): text = tweet['extended_tweet']['full_text'] elif(text.startswith('RT') == True): # print(' tweet starts with RT **********') rt = True # print(text) try: if( tweet['retweeted_status']['truncated'] == True): # print("in .... tweet.retweeted_status.truncated == True ") text = tweet['retweeted_status']['extended_tweet']['full_text'] # print(text) else: text = tweet['retweeted_status']['full_text'] except Exception as e: pass except Exception as e: print(e) text = cleanList(text) if "media" in entities: media_url = entities["media"][0]["media_url"] media_type = media_url[-4:] mentions =entities['user_mentions'] mList = [] for x in mentions: mList.append(x['screen_name']) hashtags = entities['hashtags'] # Any hashtags used in the Tweet hList =[] for x in hashtags: hList.append(x['text']) # if hashtags == []: # hashtags ='' # else: # hashtags = str(hashtags).strip('[]') coordinates = None if(exactcoord): coordinates = exactcoord['coordinates'] if ((geoenabled) and (text.startswith('RT') == False)): try: if(tweet['place']): # print(tweet['place']) place_name = tweet['place']['full_name'] place_country = tweet['place']['country'] place_countrycode = tweet['place']['country_code'] place_coordinates = tweet['place']['bounding_box']['coordinates'] except Exception as e: print(e) print('error from place details - maybe AttributeError: ... NoneType ... object has no attribute ..full_name ...') tweet1 = {'_id' : tweet_id, 'date': created, 'username': username, 'text' : text, 'geoenabled' : geoenabled, 'coordinates' : coordinates, 'location' : location, 'place_name' : place_name, 'place_country' : place_country, 'country_code': place_countrycode, 'place_coordinates' : place_coordinates, 'hashtags' : hList , 'mentions' : mList, 'source' : source, 'retweet' : rt, 'quoted' : quoted, 'verified':verified,'media_type' : media_type, 'media_url':media_url} return tweet1 def search_trends(api): ## Shows the 10 most recent trends on twitter in the UK trends = api.trends_place(23424975)[0]["trends"] # WOE ID for UK is 23424975 trend_keywords = [] for trend in trends: trend_keywords.append(trend["name"]) print("Trending keywords:", trend_keywords, "\n") return trend_keywords class StreamListener(tweepy.StreamListener): #This is a class provided by tweepy to access the Twitter Streaming API. global geoEnabled global geoDisabled time_limit = 300 def on_connect(self): self.start_time = time.time() # Called initially to connect to the Streaming API print("You are now connected to the streaming API.") def on_error(self, status_code): # On error - if an error occurs, display the error / status code print('An Error has occured: ' + repr(status_code)) return False def on_data(self, data): if (time.time() - self.start_time) < self.time_limit: #This is where each tweet is collected # let us load the json data t = json.loads(data) # now let us process the tweet so that we will deal with cleaned and extracted JSON tweet = processTweets(t) print(tweet) # now insert it # for this to work you need to start a local mongodb server try: collection.insert_one(tweet) except Exception as e: print("error",e) # this means some Mongo db insertion error else: print("Time limit reached.") #Set up the listener. The 'wait_on_rate_limit=True' is needed to help with Twitter API rate limiting. Loc_UK = [-10.392627, 49.681847, 1.055039, 61.122019] # UK and Ireland # Words_UK =["Boris", "Prime Minister", "Tories", "UK", "London", "England", "Manchester", "Sheffield", "York", "Southampton", \ # "Wales", "Cardiff", "Swansea" ,"Banff", "Bristol", "Oxford", "Birmingham" ,"Scotland", "Glasgow", "Edinburgh", "Dundee", "Aberdeen", "Highlands" \ # "Inverness", "Perth", "St Andrews", "Dumfries", "Ayr" \ # "Ireland", "Dublin", "Cork", "Limerick", "Galway", "Belfast"," Derry", "Armagh" \ # "BoJo", "Labour", "Liberal Democrats", "SNP", "Conservatives", "First Minister", "Surgeon", "Chancelor" \ # "Boris Johnson", "BoJo", "Keith Stramer"] Words_UK = search_trends(api) print("Tracking: " + str(Words_UK)) # here we ste the listener object listener = StreamListener(api=tweepy.API(wait_on_rate_limit=True)) streamer = tweepy.Stream(auth=auth, listener=listener) streamer.filter(locations= Loc_UK, track = Words_UK, languages = ['en'], is_async=True) #locations= Loc_UK, track = Words_UK, # the following line is for pure 1% sample # we can only use filter or sample - not both together # streamer.sample(languages = ['en']) Place = 'London' Lat = '51.450798' Long = '-0.137842' geoTerm=Lat+','+Long+','+'10km' # last_id = None counter =0 sinceID = None results = True while results: # print(geoTerm) if (counter < 180 ): try: results = api.search(geocode=geoTerm, count=100, lang="en", tweet_mode='extended', max_id=last_id) # media_files = set() # for status in results: # media = status.entities.get('media', []) # if (len(media) > 0): # media_files.add(media[0]['media_url']) print("results: ",results) except Exception as e: print(e) counter += 1 else: # the following let the crawler to sleep for 15 minutes; to meet the Twitter 15 minute restriction time.sleep(15*60) #
33.264706
325
0.618811
794d3fe2b5b8f1416376a60686f2372bea5f321a
387
py
Python
swf/workfair/migrations/0003_auto_20200321_0023.py
CerealWhay/swf
5c86a20b947c8e935cac323443bb27f9edaaab25
[ "bzip2-1.0.6" ]
null
null
null
swf/workfair/migrations/0003_auto_20200321_0023.py
CerealWhay/swf
5c86a20b947c8e935cac323443bb27f9edaaab25
[ "bzip2-1.0.6" ]
null
null
null
swf/workfair/migrations/0003_auto_20200321_0023.py
CerealWhay/swf
5c86a20b947c8e935cac323443bb27f9edaaab25
[ "bzip2-1.0.6" ]
1
2021-03-17T06:59:01.000Z
2021-03-17T06:59:01.000Z
# Generated by Django 3.0.4 on 2020-03-20 14:23 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('workfair', '0002_user_username'), ] operations = [ migrations.AlterField( model_name='user', name='isEmployer', field=models.BooleanField(default=False), ), ]
20.368421
53
0.599483
794d426d02daa1d711771d2e42de7ef2045a930e
4,219
py
Python
lab/06_dht-solution/simpleDHT.py
andreastedile/unitn-m-ds2
138c47f2f857e2b4a08c01edcbf0daa00b281077
[ "MIT" ]
1
2021-11-29T09:49:22.000Z
2021-11-29T09:49:22.000Z
lab/06_dht-solution/simpleDHT.py
andreastedile/unitn-m-ds2
138c47f2f857e2b4a08c01edcbf0daa00b281077
[ "MIT" ]
1
2021-09-26T13:25:48.000Z
2021-09-26T13:25:48.000Z
lab/06_dht-solution/simpleDHT.py
andreastedile/unitn-m-ds2
138c47f2f857e2b4a08c01edcbf0daa00b281077
[ "MIT" ]
1
2021-11-29T09:18:41.000Z
2021-11-29T09:18:41.000Z
import hashlib from tabulate import tabulate import code # code.interact(local=dict(globals(), **locals())) import random BITLENGTH = 8 def compute_key(string, bitlength=BITLENGTH): digest = hashlib.sha256(bytes(string, 'utf-8')).digest() subdigest = digest[:bitlength//8] return int.from_bytes(subdigest, 'little') def clock_dist(a, b, maxnum=2 ** BITLENGTH - 1): if a == b: return 0 elif a < b: return b - a else: return maxnum - (a - b) class NodeDHT(): def __init__(self, name): self.name = name self.id = compute_key(name) self.succ = None self.pred = None self.ht = dict() def setSuccessor(self, node): self.succ = node def setPredecessor(self, node): self.pred = node def join(self, requester): requesterID = requester.id dist = clock_dist(self.id, requesterID) distFromMySuccessor = clock_dist(self.succ.id, requesterID) if dist > distFromMySuccessor: self.succ.join(requester) else: requester.set_predecessor(self) requester.set_successor(self.succ) self.succ.set_predecessor(requester) self.setSuccessor(requester) def store(self, key, value): dist = clock_dist(self.id, key) distFromMySuccessor = clock_dist(self.succ.id, key) if dist > distFromMySuccessor: # Forward going closer to responsible node self.succ.store(key, value) else: self.ht[key] = value def lookup(self, key): print("node: {} looking for key: {}".format(self.id, key)) dist = clock_dist(self.id, key) distFromMySuccessor = clock_dist(self.succ.id, key) if dist > distFromMySuccessor: # Forward going closer to responsible node self.succ.lookup(key) else: try: value = self.ht[key] print("key: {} found! Value = {}".format(key, value)) return value except KeyError: print("{} not available in DHT".format(key)) return None def printRing(startnode): nodelist = [startnode.id] nextNode = startnode.succ while (nextNode != startnode): nodelist.append(nextNode.id) nextNode = nextNode.succ nodelist = sorted(nodelist) print(" -> ".join([str(x) for x in nodelist])) def printDHT(startnode): node2content = {startnode.id: startnode.storage} nextNode = startnode.succ while (nextNode != startnode): node2content[nextNode.id] = nextNode.storage nextNode = nextNode.succ tabulable = {k: sorted(list(v.items())) for k, v in sorted(node2content.items())} print(tabulate(tabulable, headers='keys')) def main(): # start a network with 4 random nodes # start first 2 nodes setting manually succ and pred. print("Creating first two nodes") n1 = NodeDHT("node1") n2 = NodeDHT("node2") n1.setSuccessor(n2) n1.setPredecessor(n2) n2.setSuccessor(n1) n2.setPredecessor(n1) print("Creating further two nodes") # add 2 more nodes by join procedure n3 = NodeDHT("node3") n4 = NodeDHT("node4") print("n3 joining from n2") n2.join(n3) print("n4 joining from n1") n1.join(n4) nodes = [n1, n2, n3, n4] # Inject some contents niceQuote = "nel mezzo del cammin di nostra vita mi ritrovai per una selva oscura che la diritta via era smarrita" contents = niceQuote.split(" ") for word in contents: key = compute_key(word) node = random.choice(nodes) node.store(key, word) printDHT(n1) code.interact(local=dict(globals(), **locals())) # Test lookups queries # Correct queries for _ in range(4): c = random.choice(contents) node = random.choice(nodes) print("Searching for \'{}\'".format(c)) node.lookup(compute_key(c)) print("\n") # Wrong query c = "NOT INSERTED UNLESS HASH CONFLICT XD" print("Searching for \'{}\'".format(c)) n1.lookup(compute_key(c)) if __name__ == "__main__": main()
26.534591
118
0.602749
794d4356a08c7e81f13f0694c6e7b3314a07c4f0
512
py
Python
config/urls.py
acuestap/cride-acp
a9920eaf032cd8f4481374c72d25bf9bc1b2e77b
[ "MIT" ]
null
null
null
config/urls.py
acuestap/cride-acp
a9920eaf032cd8f4481374c72d25bf9bc1b2e77b
[ "MIT" ]
null
null
null
config/urls.py
acuestap/cride-acp
a9920eaf032cd8f4481374c72d25bf9bc1b2e77b
[ "MIT" ]
null
null
null
"""Main URLs module.""" from django.conf import settings from django.urls import include, path from django.conf.urls.static import static from django.contrib import admin """import ipdb; ipdb.set_trace()""" urlpatterns = [ # Django Admin path(settings.ADMIN_URL, admin.site.urls), path('', include(('cride.circles.urls', 'circles'), namespace='circles')), path('', include(('cride.users.urls', 'users'), namespace='users')), ] + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
28.444444
78
0.710938
794d44a2cc74842f8b8d00f81d2ce675f076304a
5,043
py
Python
coot/data/ht100m_dataset.py
Jabb0/coot-videotext
2da20a3f3a50b69677e59869b02cbd72945913d9
[ "Apache-2.0" ]
null
null
null
coot/data/ht100m_dataset.py
Jabb0/coot-videotext
2da20a3f3a50b69677e59869b02cbd72945913d9
[ "Apache-2.0" ]
null
null
null
coot/data/ht100m_dataset.py
Jabb0/coot-videotext
2da20a3f3a50b69677e59869b02cbd72945913d9
[ "Apache-2.0" ]
null
null
null
import json import pandas as pd import numpy as np from typing import Union, List from pathlib import Path from timeit import default_timer as timer from nntrainer import data as nn_data def _time_to_seconds(time_column): return pd.to_timedelta(time_column).dt.total_seconds() class HT100MBaseDataset: """ Dataloader for HowTo100M dataset. Based on the index csv file of the HT100M dataset this builds a wrapper around the file structure to return individual files. """ def __init__(self, dataset_root: Union[str, Path], metadata_name: str, split=None): """ Setup the dataset Args: dataset_root: path to the dataset folder metadata_name: identifier of the metadata to use. Will select the files we want to use. split: identifier of the split to use or "ALL"/None to use all data """ dataset_root = Path(dataset_root) # Read the CSV file containing information about the videos # Format is: # video_id, category_1, category_2, rank, task_id # This is used as lookup table of the existing videos csv = dataset_root.joinpath(f"meta_{metadata_name}.csv") self._metadata_csv = pd.read_csv(csv, usecols=["video_id", "split"], index_col="video_id") if split is not None and split != nn_data.DataSplitConst.ALL: self._metadata_csv = self._metadata_csv[self._metadata_csv["split"] == split] metadata_path = dataset_root.joinpath("metadata.json") if not metadata_path.exists(): raise RuntimeError(f"metadata.json for HT100M dataset not found! Path: {dataset_root}") self._metadata = json.load(metadata_path.open("rt", encoding="utf8")) self._fps = self._metadata["fps"] self._caption_root = dataset_root.joinpath("captions") # Get all available caption files self._keys = self._metadata_csv.index.to_list() # Check the dataset integrity. I.e. if all caption csv files for every index are available if not self.check_integrity(): raise RuntimeError("HT100MDataset: There are data_keys for which the features are not available!") def check_integrity(self) -> bool: """ Checks if caption files for all keys exist. This is crucial for the integrity of the dataset. Returns: True if dataset integrity is correct. """ timer_start = timer() available_keys = set([x.stem for x in self._caption_root.glob("*.csv")]) print(f"Took {timer() - timer_start:.1f} seconds for scanning caption directory. " f"Found {len(self._keys)} videos.") missing_keys = set(self._keys).difference(available_keys) keys_are_missing = len(missing_keys) != 0 if keys_are_missing: print(f"There are {len(missing_keys)} missing keys. First 10: {list(missing_keys)[:10]}") return not keys_are_missing def _read_caption_csv(self, video_id: str) -> (List[str], List[float], List[float]): cap_csv = pd.read_csv(self._caption_root.joinpath(video_id + ".csv"), usecols=["start", "end", "text"], keep_default_na=False) cap_csv = cap_csv[ # Drop clips that have no subtitles/captions (cap_csv["text"].str.len() > 0) ] return (cap_csv['text'].tolist(), _time_to_seconds(cap_csv["start"]).tolist(), _time_to_seconds(cap_csv["end"]).tolist()) def __getitem__(self, video_id: str) -> List[str]: raise NotImplementedError("GetItem cannot be called on BaseDataset") def __len__(self): """ Returns len of dataset. I.e. number of videos. """ return len(self._keys) def keys(self): return self._keys def data_keys(self): return self._keys class HT100MCaptionDataset(HT100MBaseDataset): def __getitem__(self, video_id: str) -> List[str]: sentences, _, _ = self._read_caption_csv(video_id) return sentences class HT100MDataset(HT100MBaseDataset): def __init__(self, dataset_root: Union[str, Path], metadata_name: str, split: str, max_datapoints: int = -1): super(HT100MDataset, self).__init__(dataset_root, metadata_name, split=split) # reduce dataset size if request if max_datapoints > -1: self._keys = self._keys[:max_datapoints] print(f"Reduced number of datapoints to {len(self._keys)}") def __getitem__(self, key: str): sentences, starts, stops = self._read_caption_csv(key) # Drop the same items based on the filter as before return { "fps": self._fps, "data_key": key, "segments": [ { "text": text, "start_sec": start, "stop_sec": end } for (text, start, end) in zip(sentences, starts, stops) ] }
37.917293
113
0.628594
794d45aa93f96e192077fff65315a8ade2992b76
10,994
py
Python
AirC2Weather/Tools/MultidimensionSupplementalTools/Scripts/mds/tools/make_netcdf_regularpointslayer.py
esride-jts/air-c2-cop-python
692d18203ace18384b3f956ac393f41bb431d757
[ "Apache-2.0" ]
14
2017-10-04T01:19:07.000Z
2021-12-09T02:43:49.000Z
AirC2Weather/Tools/MultidimensionSupplementalTools/Scripts/mds/tools/make_netcdf_regularpointslayer.py
esride-jts/air-c2-cop-python
692d18203ace18384b3f956ac393f41bb431d757
[ "Apache-2.0" ]
14
2017-09-19T21:27:55.000Z
2020-07-23T23:34:20.000Z
AirC2Weather/Tools/MultidimensionSupplementalTools/Scripts/mds/tools/make_netcdf_regularpointslayer.py
esride-jts/air-c2-cop-python
692d18203ace18384b3f956ac393f41bb431d757
[ "Apache-2.0" ]
19
2017-11-27T14:40:40.000Z
2021-03-11T17:29:07.000Z
import arcpy debug = False class MakeNetCDFRegularPointsLayer(object): def __init__(self): """Define the tool (tool name is the name of the class).""" self.label = "Make NetCDF Regular Points Layer" self.description = "Makes a point feature layer from a netCDF file " \ "which contains regularly spaced points." self.canRunInBackground = False return def getParameterInfo(self): """Define parameter definitions""" #Define parameter definitions # First parameter param0 = arcpy.Parameter( displayName="Input netCDF File", name="in_netcdf", datatype="File", parameterType="Required", direction="Input") # Second parameter param1 = arcpy.Parameter( displayName="Variables", name="in_variables", datatype="String", multiValue = True, parameterType="Required", direction="Input") # Third parameter param2 = arcpy.Parameter( displayName="Output Feature Layer", name="out_feature_layer", datatype="String", parameterType="Required", direction="Input") # Dummpy Parameter to Force Output NetCDF Layer to be added to the # display. param3 = arcpy.Parameter( displayName="Output Feature Layer", name="dummy_feature_layer", datatype="GPFeatureLayer", parameterType="Derived", direction="Output") params = [param0, param1, param2, param3] return params def isLicensed(self): """Set whether tool is licensed to execute.""" return True def updateParameters(self, parameters): """Modify the values and properties of parameters before internal validation is performed. This method is called whenever a parameter has been changed.""" # Create list of possible variables for second parameter. if parameters[0].value and arcpy.Exists(parameters[0].valueAsText): ncFP = arcpy.NetCDFFileProperties(parameters[0].valueAsText) parameters[1].filter.list = ncFP.getVariables() # Populate the output layer name parameter with a reasonable default # (i.e. the name of the first variable chosen with _layer appended # to it.) if parameters[1].value and not parameters[2].altered: variable_name_list = str(parameters[1].valueAsText).split(';') out_layer_name = variable_name_list[0] + "_Layer" # Generate a unique layer name i = 1 temp_name = out_layer_name while arcpy.Exists(temp_name): temp_name = out_layer_name + str(i) i += 1 out_layer_name = temp_name parameters[2].value = out_layer_name return def updateMessages(self, parameters): """Modify the messages created by internal validation for each tool parameter. This method is called after internal validation.""" # Multivalue parameter 1 contains variable names. Split them into a list. selectedVars = [] if parameters[1].value: ncFP = arcpy.NetCDFFileProperties(parameters[0].valueAsText) selectedVars = [] selectedVars = str(parameters[1].valueAsText).split(';') # For each selected variable, get its dimension. Build a list dimList = [] [dimList.append(ncFP.getDimensionsByVariable(v.replace("'", ""))) \ for v in selectedVars] # Verify the all variables have the same exact dimensions if not all(x == dimList[0] for x in dimList): parameters[1].setErrorMessage("All selected variables must share " + "the same dimensions.") return def execute(self, parameters, messages): """The source code of the tool.""" # Describe the supplied netCDF File. ncFP = arcpy.NetCDFFileProperties(parameters[0].valueAsText) selectedVars = parameters[1].valueAsText.split(';') # Coordinate variables can be identified via values of their standard # name attribute or values of their untis attribute. XCoordNamesList = ["longitude", "projection_x_coordinate", "grid_longitude"] YCoordNamesList = ["latitude", "projection_y_coordinate", "grid_longitude"] XUnitNamesList = ["degrees_east", "degree_east", "degree_E", "degrees_E", "degreeE", "degreesE" ] YUnitNamesList = ["degrees_north", "degree_north", "degree_N", "degrees_N", "degreeN", "degreesN"] XAxisNameList = ["X"] YAxisNameList = ["Y"] # Case 1: Independent latitude, Longitude, Vertical and Time Axes # Get the dimensions associated with the first variable. We can use a # single variable because tool validation guarentees that all the # selected variables have the same dimensions. dimNames = ncFP.getDimensionsByVariable(str(selectedVars[0])) x_variable = "" y_variable = "" for dimension in dimNames: # Per CF 1.6 (Chapter 5, preamble): "All of a variable's dimensions # that are latitude, longitude, vertical, or time dimensions must # have corresponding coordinate variables, i.e., one-dimensional # variables with the same name as the dimension." # Identify the coordinate variable by its standard name or units. try: SNattributeValue = ncFP.getAttributeValue(dimension, "standard_name") except: SNattributeValue = "missing" try: UNattributeValue = ncFP.getAttributeValue(dimension, "units") except: UNattributeValue = "missing" try: AXattributeValue = ncFP.getAttributeValue(dimension,"axis") except: AXattributeValue = "missing" if SNattributeValue in XCoordNamesList: x_variable = dimension if SNattributeValue in YCoordNamesList: y_variable = dimension if UNattributeValue in XUnitNamesList: x_variable = dimension if UNattributeValue in YUnitNamesList: y_variable = dimension if AXattributeValue in XAxisNameList: x_variable = dimension if AXattributeValue in YAxisNameList: y_variable = dimension # Set the row dimensions parameter of the Make NetCDF Feature Layer tool # to the names of the latitude and longitude dimensions rowDimensions = "" rowDimensions = x_variable + ";" + y_variable # Case #2: 2 dimensional lat/long coordinate variable # If unsuccessful in locating variable for x and y coordinates based on # latitude and longitude, check to see if variable of interest is # georeferenced via a 2 dimensional lat/long coordinate variable # Coordinate Variable Method will only work if CDL conforms to the # CF 1.6 convention (section 2.4) that "All other dimensions should, # whenever possible, be placed to the left of the spatiotemporal # dimensions." if (not x_variable) or (not y_variable): try: coordAttributeValue = ncFP.getAttributeValue(str(selectedVars[0]), "coordinates") coordVariables = coordAttributeValue.split(' ') for element in coordVariables: try: SNattributeValue = ncFP.getAttributeValue(element, "standard_name") except: SNattributeValue = "missing" try: UNattributeValue = ncFP.getAttributeValue(element, "units") except: UNattributeValue = "missing" try: AXattributeValue = ncFP.getAttributeValue(element, "axis") except: AXattributeValue = "missing" if SNattributeValue in XCoordNamesList: x_variable = element if SNattributeValue in YCoordNamesList: y_variable = element if UNattributeValue in XUnitNamesList: x_variable = element if UNattributeValue in YUnitNamesList: y_variable = element if AXattributeValue in XAxisNameList: x_variable = element if AXattributeValue in YAxisNameList: y_variable = element # Set the row dimensions parameter of the Make NetCDF Feature # Layer tool to the names of the coordinate variables dimNames = ncFP.getDimensionsByVariable(str(selectedVars[0])) rowDimensions = str(dimNames[-1]) + ";" + str(dimNames[-2]) except: CoordAttributeValue = "missing" # Convert the python list of selected variables into a single # (comma delimited string if necessary. if selectedVars.count > 1: variable_list = ','.join([str(x) for x in selectedVars]) else: variable_list = selectedVars[0] if (x_variable) and (y_variable): if (debug): arcpy.AddWarning("netCDFFile Name: %s" \ % parameters[0].valueAsText) arcpy.AddWarning("Variable List: %s" % variable_list) arcpy.AddWarning("x_variable: %s" % x_variable) arcpy.AddWarning("y_variable: %s" % y_variable) arcpy.AddWarning("Output Feature Layer: %s " \ % parameters[2].valueAsText) arcpy.AddWarning("Row Dimensions: %s " % rowDimensions) result1 = arcpy.MakeNetCDFFeatureLayer_md(parameters[0].valueAsText, variable_list, x_variable, y_variable, parameters[2].value, rowDimensions) # Force the netCDF Feature Layer to be added to the display arcpy.SetParameter(3, result1) else: arcpy.AddError("Unable to automatically determine x and y " + "variables from the netCDF file. Use Make NetCDF " + "Feature Layer tool.") return
41.802281
82
0.571221
794d45cc4bf3d37f04fbd3a8a9331d8734fc5c15
764
py
Python
chapter_1/C.py
staguchi0703/ALDS1
bb3d5865fb0a3f459f7a7331190e146c0e817a33
[ "MIT" ]
null
null
null
chapter_1/C.py
staguchi0703/ALDS1
bb3d5865fb0a3f459f7a7331190e146c0e817a33
[ "MIT" ]
null
null
null
chapter_1/C.py
staguchi0703/ALDS1
bb3d5865fb0a3f459f7a7331190e146c0e817a33
[ "MIT" ]
null
null
null
# %% # VScodeで入力をテキストから読み込んで標準入力に渡す import sys import os f=open(r'.\Chapter-1\C_input.txt', 'r', encoding="utf-8") # inputをフルパスで指定 # win10でファイルを作るとs-jisで保存されるため、読み込みをutf-8へエンコードする必要あり # VScodeでinput file開くとutf8になってるんだけど中身は結局s-jisになっているらしい sys.stdin=f # # 入力スニペット # num = int(input()) # num_list = [int(item) for item in input().split()] # num_list = [input() for _ in range(3)] ################################## # %% # 以下ペースト可 N = int(input()) num_list = [int(input()) for _ in range(N)] import math def prime_chk(n): if 0 < n <= 3: return True else: for i in range(2, int(math.sqrt(n)+1)): if n % i == 0: return False return True res_list = [i for i in num_list if prime_chk(i)] print(len(res_list))
23.151515
57
0.609948
794d4715cad199cc225c5b52c1936ef08435127b
16,354
py
Python
src/ertk/utils/array.py
agkphysics/emotion
36bb9265f9439b10676fb539d5334cce645e49ef
[ "MIT" ]
1
2020-06-14T03:37:46.000Z
2020-06-14T03:37:46.000Z
src/ertk/utils/array.py
agkphysics/emotion
36bb9265f9439b10676fb539d5334cce645e49ef
[ "MIT" ]
null
null
null
src/ertk/utils/array.py
agkphysics/emotion
36bb9265f9439b10676fb539d5334cce645e49ef
[ "MIT" ]
null
null
null
from typing import List, Optional, Sequence, Tuple, Union, overload import numpy as np def make_array_array(x: List[np.ndarray]) -> np.ndarray: """Helper function to make an array of arrays. The returned array has `shape==(len(x),)` and `dtype==object`. """ arr = np.empty(len(x), dtype=object) for i, a in enumerate(x): arr[i] = a return arr def check_3d(arrays: Union[Sequence[np.ndarray], np.ndarray]) -> None: """Checks if an array is 3D or each array in a list is 2D. Raises an exception if this isn't the case. """ if any(len(x.shape) != 2 for x in arrays): raise ValueError("arrays must be 3D (contiguous or vlen).") def is_mono_audio(x: np.ndarray) -> bool: """Checks if an array represents mono audio data. The conditions for this are that x can be squeezed to 1D. Note that this method will also return True for '2D' sequences of length 1, so care should be taken for edge cases. """ return np.squeeze(x).ndim == 1 def frame_array( x: np.ndarray, frame_size: int, frame_shift: int, pad: bool = False, axis: int = 0, copy: bool = True, ) -> np.ndarray: """Frames an array over a given axis with optional padding and copying. Parameters ---------- x: np.ndarray Array to frame. frame_size: int Width of frame. frame_shift: int Amount each window is moved for subsequent frames. pad: bool If the frames don't neatly fit the axis length, this indicated whether to pad or cut the remaining length. Default is `False` which means any remaining length is cut. If `pad=True` then any remaining length is padded with zeros and an additional frame is produced. If frames neatly fit into the axis length then this option has no effect. axis: int Axis to frame over. Default is 0. copy: bool Whether to copy and return a contiguous array (more memory usage). Default is `True` so that a contiguous copy is retured. Note: if `copy=False` it is assumed that the input array is contiguous so that a view with modified strides can be properly created. Returns ------- frames: np.ndarray The resulting frames. If `copy=False` then these are read-only views into x, otherwise a contiguous array is returned. """ idx_slice = tuple(slice(None) for _ in range(axis)) num_frames = (x.shape[axis] - frame_size) // frame_shift + 1 if num_frames <= 0: if not pad: raise ValueError( "The length of the sequence is shorter than frame_size, but pad=False, " "so no frames will be generated." ) num_frames = 0 remainder = (x.shape[axis] - frame_size) % frame_shift if remainder != 0: num_frames += 1 if pad else 0 # Insert new dim before axis with num_frames, and axis is replaced # with the size of each frame. new_shape = x.shape[:axis] + (num_frames, frame_size) + x.shape[axis + 1 :] if copy: out = np.zeros(new_shape, dtype=x.dtype) for i in range(0, x.shape[axis] - frame_size + 1, frame_shift): out[idx_slice + (i // frame_shift,)] = x[ idx_slice + (slice(i, i + frame_size),) ] if remainder != 0 and pad: if num_frames == 1: left = x.shape[axis] else: left = frame_size + remainder - frame_shift # out[..., -1, 0:left, ...] = x[..., -left:, ...] out[idx_slice + (-1, slice(0, left))] = x[idx_slice + (slice(-left, None),)] return out from numpy.lib.stride_tricks import as_strided if remainder != 0: if pad: # Padding creates a new array copy raise ValueError("pad must be False when copy=False") # x[:, :, ..., 0:N-r, :, ...] x = x[idx_slice + (slice(0, x.shape[axis] - remainder),)] # Add a new stride for the num_frames part of shape which moves by # frame_shift for each frame. The stride within each frame is the # same as for the original axis. new_strides = x.strides[:axis] + (frame_shift * x.strides[axis],) + x.strides[axis:] return as_strided(x, new_shape, new_strides, writeable=False) def frame_arrays( arrays: Union[List[np.ndarray], np.ndarray], frame_size: int, frame_shift: int, max_frames: Optional[int] = None, vlen: bool = False, ) -> np.ndarray: """Creates sequences of frames from the given arrays. This is mainly a convenience wrapper around `frame_array()` for instance arrays. Parameters ---------- arrays: list of np.ndarray The arrays to process. frame_size: int Length of each frame. frame_shift: int Amount each frame is shifted. max_frames: int, optional If given, only the first max_frames are returned for each array. vlen: bool Whether to return a contiguous array if `vlen=False` (default) or an array of variable length arrays. Returns ------- framed_arrays: Framed arrays. """ check_3d(arrays) if max_frames is None: max_len = max(len(x) for x in arrays) max_frames = (max_len - frame_size) // frame_shift + 1 # Generator arrs = ( frame_array(x, frame_size, frame_shift, axis=0, pad=True, copy=True)[ :max_frames ] for x in arrays ) if vlen: return make_array_array(list(arrs)) n_features = arrays[0].shape[1] framed_arrays = np.zeros( (len(arrays), max_frames, frame_size, n_features), dtype=arrays[0].dtype ) for i, arr in enumerate(arrs): framed_arrays[i, : len(arr)] = arr return framed_arrays def pad_array( x: np.ndarray, to_multiple: int = None, to_size: int = None, axis: int = 0 ): """Pads an array either to a multiple of `to_multiple` or to the exact length `to_size` along `axis`. Parameters ---------- x: np.ndarray The array to pad. to_multiple: int, optional If given, `x` is padded so that it's length is the nearest larger multiple of `to_multiple`. to_size: int, optional If given, `x` is padded so that it's length is exactly `to_size`. Default is None, and instead `to_multiple` is used. Exactly one of `to_size` and `to_multiple` must be given. axis: int The axis along which to pad. Default is axis 0. Returns ------- padded: np.ndarray The padded array. """ if to_multiple is not None and to_size is not None: raise ValueError("Only one of `to_multiple` and `to_size` should be given.") if to_size is not None: to_pad = to_size - x.shape[axis] if to_pad < 0: raise ValueError("The length of `x` is already greater than `to_size`.") elif to_multiple is not None: to_pad = int(np.ceil(x.shape[axis] / to_multiple)) * to_multiple - x.shape[axis] else: raise ValueError("One of `to_multiple` and `to_size` must be given.") if to_pad == 0: return x.copy() # A copy is expected when padding pad = [(0, 0) for _ in x.shape] pad[axis] = (0, to_pad) return np.pad(x, pad) @overload def pad_arrays(arrays: List[np.ndarray], pad: int) -> List[np.ndarray]: pass @overload def pad_arrays(arrays: np.ndarray, pad: int) -> np.ndarray: pass def pad_arrays(arrays: Union[List[np.ndarray], np.ndarray], pad: int = 32): """Pads each array to the nearest multiple of `pad` greater than the array size. Assumes axis 0 of each sub-array, or axis 1 of x, is the time axis. This is mainly a wrapper around `pad_array()` for instance arrays. """ if isinstance(arrays, np.ndarray) and len(arrays.shape) > 1: # Contiguous 2D/3D return pad_array(arrays, to_multiple=pad, axis=1) new_arrays = [pad_array(x, to_multiple=pad, axis=0) for x in arrays] if isinstance(arrays, np.ndarray): if all(x.shape == new_arrays[0].shape for x in new_arrays): # Contiguous array return np.stack(new_arrays) # Array of variable-length arrays return make_array_array(new_arrays) # List return new_arrays @overload def clip_arrays(arrays: List[np.ndarray], length: int, copy: bool) -> List[np.ndarray]: pass @overload def clip_arrays(arrays: np.ndarray, length: int, copy: bool) -> np.ndarray: pass def clip_arrays( arrays: Union[List[np.ndarray], np.ndarray], length: int, copy: bool = True ): """Clips each array to the specified maximum length.""" if isinstance(arrays, np.ndarray): if len(arrays.shape) > 1: return arrays[:, :length, ...].copy() if copy else arrays[:, :length, ...] new_arrays = [x[:length].copy() if copy else x[:length] for x in arrays] if all(x.shape == new_arrays[0].shape for x in new_arrays): # Return contiguous array return np.stack(new_arrays) return make_array_array(new_arrays) return [x[:length].copy() if copy else x[:length] for x in arrays] @overload def transpose_time(arrays: List[np.ndarray]) -> List[np.ndarray]: pass @overload def transpose_time(arrays: np.ndarray) -> np.ndarray: pass def transpose_time(arrays: Union[List[np.ndarray], np.ndarray]): """Transpose the time and feature axis of each array. Requires each array be 2-D. Note: This function modifies the arrays in-place. """ check_3d(arrays) if isinstance(arrays, np.ndarray) and len(arrays.shape) == 3: arrays = arrays.transpose(0, 2, 1) else: for i in range(len(arrays)): arrays[i] = arrays[i].transpose() assert all(x.shape[0] == arrays[0].shape[0] for x in arrays) return arrays def shuffle_multiple( *arrays: Union[np.ndarray, Sequence], numpy_indexing: bool = True, seed: Optional[int] = None, ): """Shuffles multiple arrays or lists in sync. Useful for shuffling the data and labels in a dataset separately while keeping them synchronised. Parameters ---------- arrays, iterable of array-like The arrays to shuffle. They must all have the same size of first dimension. numpy_indexing: bool, default = True Whether to use NumPy-style indexing or list comprehension. Returns: shuffled_arrays: iterable of array-like The shuffled arrays. """ if any(len(arrays[0]) != len(x) for x in arrays): raise ValueError("Not all arrays have equal first dimension.") perm = np.random.default_rng(seed).permutation(len(arrays[0])) new_arrays = [ array[perm] if numpy_indexing else [array[i] for i in perm] for array in arrays ] return new_arrays def batch_arrays_by_length( arrays_x: Union[np.ndarray, List[np.ndarray]], y: np.ndarray, batch_size: int = 32, shuffle: bool = True, uniform_batch_size: bool = False, ) -> Tuple[np.ndarray, np.ndarray]: """Batches a list of arrays of different sizes, grouping them by size. This is designed for use with variable length sequences. Each batch will have a maximum of batch_size arrays, but may have less if there are fewer arrays of the same length. It is recommended to use the pad_arrays() method of the LabelledDataset instance before using this function, in order to quantise the lengths. Parameters ---------- arrays_x: list of ndarray A list of N-D arrays, possibly of different lengths, to batch. The assumption is that all the arrays have the same rank and only axis 0 differs in length. y: ndarray The labels for each of the arrays in arrays_x. batch_size: int Arrays will be grouped together by size, up to a maximum of batch_size, after which a new batch will be created. Thus each batch produced will have between 1 and batch_size items. shuffle: bool, default = True Whether to shuffle array order in a batch. uniform_batch_size: bool, default = False Whether to keep all batches the same size, batch_size, and pad with zeros if necessary, or have batches of different sizes if there aren't enough sequences to group together. Returns ------- x_list: ndarray, The batched arrays. x_list[i] is the i'th batch, having between 1 and batch_size items, each of length lengths[i]. y_list: ndarray The batched labels corresponding to sequences in x_list. y_list[i] has the same length as x_list[i]. """ if shuffle: arrays_x, y = shuffle_multiple(arrays_x, y, numpy_indexing=False) # TODO: implement batch generator fixed_shape = arrays_x[0].shape[1:] lengths = [x.shape[0] for x in arrays_x] unique_len = np.unique(lengths) x_dtype = arrays_x[0].dtype y_dtype = y.dtype xlist = [] ylist = [] for length in unique_len: idx = np.nonzero(lengths == length)[0] for b in range(0, len(idx), batch_size): batch_idx = idx[b : b + batch_size] size = batch_size if uniform_batch_size else len(batch_idx) _x = np.zeros((size, length) + fixed_shape, dtype=x_dtype) _y = np.zeros(size, dtype=y_dtype) _y[:size] = y[batch_idx] for i, j in enumerate(batch_idx): _x[i, ...] = arrays_x[j] xlist.append(_x) ylist.append(_y) x_batch = make_array_array(xlist) if uniform_batch_size: y_batch = np.array(ylist, dtype=y_dtype) else: y_batch = make_array_array(ylist) return x_batch, y_batch def flat_to_inst(x: np.ndarray, slices: Union[np.ndarray, List[int]]) -> np.ndarray: """Takes a concatenated 2D data array and converts it to either a contiguous 2D/3D array or a variable-length 3D array, with one feature vector/matrix per instance. Parameters ---------- x: numpy.ndarray Contiguous 2D array containing concatenated instance data. slices: list or numpy.ndarray Array containing lengths of sequences in x corresponding to instance data. Returns ------- inst: numpy.ndarray Instance array, which is either contiguous 2D, contiguous 3D or an array of 2D arrays with varying lengths. It should be the case that `inst[i] == x[s : s + slices[i]]`, where `s = sum(slices[:i])`. """ if len(x) == len(slices): # 2D contiguous array return x elif all(x == slices[0] for x in slices): # 3D contiguous array assert len(x) % len(slices) == 0 return x.reshape(len(slices), len(x) // len(slices), x[0].shape[-1]) else: # 3D variable length array start_idx = np.cumsum(slices)[:-1] return make_array_array(np.split(x, start_idx, axis=0)) def inst_to_flat(x: np.ndarray) -> Tuple[np.ndarray, np.ndarray]: """The inverse of flat_to_inst(). Takes an instance 'matrix' and converts to a "flattened" 2D matrix. Parameters ---------- x: numpy.ndarray The instance matrix, which is either a contiguous 2D or 3D array, or an array of 1D or 2D arrays. Returns ------- flat: numpy.ndarray The contiguous 2D matrix containing data for all instances. slices: numpy.ndarray Array of slices (sequence lengths) in `flat`. It should be the case that `x[i] == flat[s : s + slices[i]]`, where `s = sum(slices[:i])`. """ slices = np.ones(len(x), dtype=int) if len(x.shape) != 2: if x[0].ndim == 1: # Array of 1D arrays x = np.stack(x) # type: ignore else: slices = np.array([len(_x) for _x in x]) if len(x.shape) == 3: # Contiguous 3D array x = x.reshape(sum(slices), x.shape[2]) elif x[0].base is not None and all(_x.base is x[0].base for _x in x): # Array of views into contiguous array x = x[0].base else: # Array of separate arrays x = np.concatenate(x) assert sum(slices) == len(x) return x, slices
33.859213
88
0.622722
794d47c4f33f021f52d0631aaedb0d583c411d01
6,598
py
Python
minerl_agent/behaviour_cloning/tfrecrods.py
metataro/minerl_agent
f61a587b778afe5d70d260012ee850013e809a14
[ "MIT" ]
9
2020-11-09T15:27:40.000Z
2021-12-22T01:13:27.000Z
minerl_agent/behaviour_cloning/tfrecrods.py
metataro/minerl_agent
f61a587b778afe5d70d260012ee850013e809a14
[ "MIT" ]
5
2020-06-18T23:08:31.000Z
2022-02-09T23:39:00.000Z
minerl_agent/behaviour_cloning/tfrecrods.py
metataro/minerl_agent
f61a587b778afe5d70d260012ee850013e809a14
[ "MIT" ]
4
2020-07-02T05:30:47.000Z
2021-07-30T04:13:17.000Z
import logging import numpy as np import tensorflow as tf from minerl.env import spaces from utility.utils import flatten_nested_dicts, unflatten_nested_dicts logger = logging.getLogger(__name__) def _bytes_feature(value): return tf.train.Feature(bytes_list=tf.train.BytesList(value=[value])) def _float_feature(value): return tf.train.Feature(float_list=tf.train.FloatList(value=[value])) def _int64_feature(value): return tf.train.Feature(int64_list=tf.train.Int64List(value=[value])) def _bytes_feature_list(values): return tf.train.FeatureList(feature=[_bytes_feature(v) for v in values]) def _float_feature_list(values): return tf.train.FeatureList(feature=[_float_feature(v) for v in values]) def _int64_feature_list(values): return tf.train.FeatureList(feature=[_int64_feature(v) for v in values]) def serialize_sequence_example(sequence) -> tf.train.SequenceExample: def _create_feature(values: np.ndarray): if len(values.shape) > 1: return _bytes_feature_list([value.tobytes() for value in values]) elif values.dtype == np.float32 or values.dtype == np.float64: return _float_feature_list(values.astype(np.float32)) else: return _int64_feature_list(values.astype(np.int64)) observations, actions, rewards, done, step_mul = sequence context = tf.train.Features(feature={ 'sequence_length': _int64_feature(len(rewards)), 'score': _float_feature(sum(rewards)) }) feature_list = { 'observation': tf.nest.map_structure(_create_feature, observations), 'action': tf.nest.map_structure(_create_feature, actions), 'reward': _float_feature_list(rewards), 'done': _int64_feature_list(done), 'step_mul': _int64_feature_list(step_mul), } feature_list = flatten_nested_dicts(feature_list) return tf.train.SequenceExample(context=context, feature_lists=tf.train.FeatureLists(feature_list=feature_list)) def parse_sequence_example(observation_space, action_space, example_proto): def _feature_specs(space: spaces.Dict): if isinstance(space, spaces.Dict): return {k: _feature_specs(v) for k, v in space.spaces.items()} elif space.shape and len(space.shape) > 0: return tf.FixedLenSequenceFeature((), tf.string) else: return tf.FixedLenSequenceFeature((), tf.int64) def _set_type_and_shape(value: tf.Tensor): if value.dtype == tf.int64: return tf.cast(value, tf.int32) elif value.dtype == tf.float32: return value elif value.dtype == tf.string: return tf.decode_raw(value, tf.uint8) else: raise ValueError(f"unexpected type {value.dtype.name}") context_features = { 'sequence_length': tf.FixedLenFeature([], dtype=tf.int64), 'score': tf.FixedLenFeature([], dtype=tf.float32), } sequence_features = { 'observation': tf.nest.map_structure(_feature_specs, observation_space), 'action': tf.nest.map_structure(_feature_specs, action_space), 'reward': tf.FixedLenSequenceFeature((), tf.float32), 'done': tf.FixedLenSequenceFeature((), tf.int64), 'step_mul': tf.FixedLenSequenceFeature((), tf.int64), } sequence_features = flatten_nested_dicts(sequence_features) context, parsed_features = tf.parse_single_sequence_example( serialized=example_proto, context_features=context_features, sequence_features=sequence_features ) parsed_features = unflatten_nested_dicts(parsed_features) context = tf.nest.map_structure(_set_type_and_shape, context) parsed_features = tf.nest.map_structure(_set_type_and_shape, parsed_features) parsed_features['observation']['pov'] = tf.reshape(parsed_features['observation']['pov'], (-1,) + observation_space['pov'].shape) parsed_features['done'] = tf.cast(parsed_features['done'], tf.bool) return context, parsed_features def serialize_single_example(example) -> tf.train.SequenceExample: def _create_feature(values: np.ndarray): if len(values.shape) > 1: return _bytes_feature(values.tobytes()) elif values.dtype == np.float32 or values.dtype == np.float64: return _float_feature(values.astype(np.float32)) else: return _int64_feature(values.astype(np.int64)) observations, actions, rewards, done = example feature_list = { 'observation': tf.nest.map_structure(_create_feature, observations), 'action': tf.nest.map_structure(_create_feature, actions), 'reward': _float_feature(rewards), 'done': _int64_feature(done) } feature_list = flatten_nested_dicts(feature_list) return tf.train.Example(features=tf.train.Features(feature=feature_list)) def parse_single_example(observation_space, action_space, example_proto): def _feature_specs(space: spaces.Dict): if isinstance(space, spaces.Dict): return {k: _feature_specs(v) for k, v in space.spaces.items()} elif space.shape and len(space.shape) > 0: return tf.FixedLenFeature((), tf.string) else: return tf.FixedLenFeature((), tf.int64) def _set_type_and_shape(value: tf.Tensor): if value.dtype == tf.int64: return tf.cast(value, tf.int32) elif value.dtype == tf.float32: return value elif value.dtype == tf.string: return tf.decode_raw(value, tf.uint8) else: raise ValueError(f"unexpected type {value.dtype.name}") features = { 'observation': tf.nest.map_structure(_feature_specs, observation_space), 'action': tf.nest.map_structure(_feature_specs, action_space), 'reward': tf.FixedLenFeature((), tf.float32), 'done': tf.FixedLenFeature((), tf.int64), } features = flatten_nested_dicts(features) context, parsed_features = tf.parse_single_example(example_proto, features) parsed_features = unflatten_nested_dicts(parsed_features) context = tf.nest.map_structure(_set_type_and_shape, context) parsed_features = tf.nest.map_structure(_set_type_and_shape, parsed_features) parsed_features['observation']['pov'] = tf.reshape(parsed_features['observation']['pov'], observation_space['pov'].shape) parsed_features['done'] = tf.cast(parsed_features['done'], tf.bool) return context, parsed_features
37.276836
116
0.683237
794d483b1ce519ea83e2b2065ae0eeeb03a52eb8
6,785
py
Python
mzXML2mat.py
EvgenyZhvansky/R_matrix
66f5da0af89866da533a5714ffaf4a6567d7d5bc
[ "MIT" ]
1
2020-12-28T14:19:12.000Z
2020-12-28T14:19:12.000Z
mzXML2mat.py
EvgenyZhvansky/R_matrix
66f5da0af89866da533a5714ffaf4a6567d7d5bc
[ "MIT" ]
2
2021-03-31T20:19:44.000Z
2021-12-13T20:41:06.000Z
mzXML2mat.py
EvgenyZhvansky/R_matrix
66f5da0af89866da533a5714ffaf4a6567d7d5bc
[ "MIT" ]
null
null
null
# import sys import numpy as np from pyteomics import mzxml as pymz import xml.etree.ElementTree as ET from tkinter import Tk from tkinter.filedialog import askopenfilenames import scipy.io as io def get_scan_count(tree_local): scan_count=0 for (event, elem) in tree_local: if 'msRun' in elem.tag: scan_count = int(elem.attrib['scanCount']) return scan_count def binning(mz, spectr, mzs_net,mz_len): index = np.digitize(mz, mzs_net) return np.bincount(index, weights=spectr,minlength=mz_len+2)[1:-1] def median_filtering(filenames, scans_count, n_median, mzs_net,mz_len): #scans_count=scans_count//n_median spectra_binned=np.zeros((scans_count.sum(),mz_len)) k=0 if n_median==1: for filename in filenames: reader = pymz.read(filename) print(filename) for tree in reader: mz = tree['m/z array'] spectr = tree['intensity array'] spectra_binned[k,:]=binning(mz,spectr,mzs_net,mz_len) k+=1 else: spectra_to_median=np.zeros((n_median,mz_len)) for i, filename in enumerate(filenames): reader = pymz.read(filename) print(filename) for j in range(scans_count[i]): for t in range(n_median): tree=next(reader) mz = tree['m/z array'] spectr = tree['intensity array'] spectra_to_median[t,:]=binning(mz,spectr,mzs_net,mz_len) spectra_binned[k,:]=np.median(spectra_to_median,axis=0) k+=1 return spectra_binned def calc_median(filenames,whole_scans_count, mzmin, mzmax, mz_bin,mem_in_GB): mz_len=int(1+np.round((mzmax-mzmin)/mz_bin)) #estimation of moving median parameter n_median=int(whole_scans_count*mz_len*8/mem_in_GB/2**30) if n_median<1: n_median=1 else: n_median=int(2*np.floor(n_median/2.0)+1) while int(np.sum(np.floor(scans_count/n_median))*mz_len*8/mem_in_GB/2**30)>=1: n_median-=2 n_median+=2 return n_median def calc_scans_count(filenames, mzmin, mzmax, mz_bin, n_median,scans_count,mem_in_GB): mz_len=int(1+np.round((mzmax-mzmin)/mz_bin)) if np.sum(scans_count//n_median)*mz_len*8/mem_in_GB/2**30>1: scans_count=scans_count//n_median over_scans=np.ceil(mem_in_GB*2**30/mz_len/8)-np.sum(scans_count) for i in range(over_scans): scans_count[np.argmax(scans_count)]-=1 scans_count*=n_median else: scans_count = scans_count // n_median return scans_count,n_median def read_and_convert_data(filenames, mzmin=200, mzmax=1000, mz_bin=0.1, n_median=None, GB_max=2): mz_len=int(1+np.round((mzmax-mzmin)/mz_bin)) scans_count=np.zeros(len(filenames), dtype=np.int32) for i,filename in enumerate(filenames): scans_count[i]=get_scan_count(ET.iterparse(filename, events=("start","end"))) whole_scans_count=scans_count.sum() if n_median==None: n_median=calc_median(filenames, whole_scans_count, mzmin, mzmax, mz_bin,GB_max) else: n_median = int(2 * np.floor(n_median / 2.0) + 1) scans_count,n_median=calc_scans_count(filenames, mzmin, mzmax, mz_bin, n_median,scans_count,GB_max) print("Median width is set to ", str(n_median), '\n') whole_median_scans_count=int(np.sum(scans_count//n_median)) mz_array=mzmin+mz_bin*np.array(range(mz_len)) mzs_net=mzmin-mz_bin/2+mz_bin*np.array(range(mz_len+1)) spectra_binned=np.zeros((whole_median_scans_count,mz_len)) return mz_array, median_filtering(filenames, scans_count, n_median, mzs_net, mz_len),scans_count def create_data_var(filenames, mz_array,spectra,scans_count): for filename in filenames: outs = filename.split("/") break out_filename_mat='' for i in range(len(outs)-1): out_filename_mat += outs[i] + "/" out_filename_mat += 'data.mat' data={} data['mz'] = mz_array data['spectra'] = spectra i=0 N=scans_count.sum() arr=np.zeros(N) for id,el in enumerate(scans_count): arr[i:i+el]=id i+=el data['filenames']=np.empty((len(filenames),), dtype=np.object) filenames_only=filenames for i,filename in enumerate(filenames): outs = filename.rsplit("/",maxsplit=1) data['filenames'][i]=outs[1] data['scan_of_file']=arr return out_filename_mat, data if __name__ == '__main__': print('wait for open file diolog ...') root = Tk() root.withdraw() filenames = askopenfilenames(parent=root, filetypes=[("mzXML files", ".mzXML")]) if filenames != '': mzmin = 200 mzmax = 1000 mzbin=0.1 n_median=None GB_max=2 print("Skip (press Enter) any step to use the default value") try: mzmin=float(input("Enter the m/z min > 0 (default is 200): ")) except ValueError: print('Default will be used. Enter only digits. For decimals use " . " ') try: mzmax=float(input("\nEnter the m/z max > m/z min (default is 1000): ")) except ValueError: print('Default will be used. Enter only digits. For decimals use " . " ') try: mzbin=float(input("\nEnter the m/z bin width > 0 (default is 0.1): ")) except ValueError: print('Default will be used. Enter only digits. For decimals use " . " ') try: n_median=int(input("\nEnter the median width value > 0 (enter 1 for no median filtration, skip (press Enter) for automated width chosing): ")) except ValueError: print("Default median parameter will be calculated automatically. Enter only digits") try: GB_max=int(input("\nEnter the RAM space max > 1 (in GB, default is 2GB): ")) if GB_max<1: GB_max=2 print('Too small value. Default will be used (2).') except ValueError: print('Default will be used (2). Enter only digits.') mz_array,spectra,scans_count=read_and_convert_data(filenames, mzmin=mzmin, mzmax=mzmax, mz_bin=mzbin, n_median=n_median, GB_max=GB_max) out_filename_mat, data = create_data_var(filenames, mz_array,spectra,scans_count) io.savemat(out_filename_mat, {'data':data}) print(out_filename_mat) input("Press any key to exit... ")
36.875
154
0.607517
794d48d45a78b2388c058f2480add2c536b6ee20
32,882
py
Python
tensorflow_transform/graph_tools.py
Mikehem/tfx
e803ea6778d8550ec77dcc92bc8172f1a3a90f38
[ "Apache-2.0" ]
null
null
null
tensorflow_transform/graph_tools.py
Mikehem/tfx
e803ea6778d8550ec77dcc92bc8172f1a3a90f38
[ "Apache-2.0" ]
null
null
null
tensorflow_transform/graph_tools.py
Mikehem/tfx
e803ea6778d8550ec77dcc92bc8172f1a3a90f38
[ "Apache-2.0" ]
null
null
null
# Copyright 2017 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tools for analyzing a TensorFlow graph. This module exports the function determine_ready_tensors_and_table_initializers which analyzes a TensorFlow graph to determine which tensors and table initializers are "ready". The concept of readiness arises as tf.Transform works by building a single TF graph containing placeholders for the outputs of analyzers. These placeholders are progressively replaced by constants in a number of phases, where in each phase we run some analyzers and replace their outputs with constants. We analyze the structure of the graph to determine which analyzers to run in each phase. """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import collections import itertools import uuid # GOOGLE-INITIALIZATION import six import tensorflow as tf from tensorflow_transform import analyzer_nodes from tensorflow_transform import tf_utils # TODO(https://issues.apache.org/jira/browse/SPARK-22674): Switch to # `collections.namedtuple` or `typing.NamedTuple` once the Spark issue is # resolved. from tfx_bsl.types import tfx_namedtuple # pylint: disable=g-direct-tensorflow-import from tensorflow.python.framework import composite_tensor from tensorflow.python.framework import function_def_to_graph from tensorflow.python.util import object_identity # pylint: enable=g-direct-tensorflow-import _INITIALIZABLE_TABLE_OP_TYPES = [ 'CuckooTable', 'CuckooTableV2', 'HashTable', 'HashTableV2', 'IndexTable', 'IndexTableV2', ] _TABLE_INIT_OP_TYPES = [ 'InitializeTable', 'InitializeTableV2', 'InitializeTableFromTextFile', 'InitializeTableFromTextFileV2', 'InitializeTableFromDataset', 'LookupTableImport', 'LookupTableImportV2', # If a TF 2 SavedModel/Hub module with tables is loaded inside the # pre-processing fn, a StatefulPartitionedCall is added to the # TABLE_INITIALIZERS collection. 'StatefulPartitionedCall', ] def _decompose_tensor_or_op(tensor_or_op): """Yields the raw components of a `tf.CompositeTensor`. If tensor_or_op is a `tf.Operation`, or `tf.Tensor`, then _decompose_tensor_or_op will act as a pass through. Args: tensor_or_op: `tf.Tensor`, `tf.CompositeTensor`, or `tf.Operation`. Yields: A tf.Tensor or tf.Operation, depending on what tensor_or_op is. """ if isinstance(tensor_or_op, composite_tensor.CompositeTensor): for component in tf.nest.flatten(tensor_or_op, expand_composites=True): yield component else: yield tensor_or_op def retrieve_sources(sinks, ignore_control_dependencies=False): """Captures subgraph between sources and sinks. Walk a Graph backwards from `sinks` and return any sources encountered in the subgraph. This util is refactored from `_map_subgraph` in tensorflow/.../ops/op_selector.py. Arguments: sinks: An iterable of Operations where the subgraph terminates. ignore_control_dependencies: (Optional) If `True`, ignore any `control_inputs` for all ops while walking the graph. Returns: The set of placeholders upon which `sinks` depend. This could also contain placeholders representing `captures` in the graph. """ stop_at_tensors = object_identity.ObjectIdentitySet() ops_to_visit = object_identity.ObjectIdentitySet(sinks) visited_ops = object_identity.ObjectIdentitySet() potential_extra_sources = object_identity.ObjectIdentitySet() while ops_to_visit: op = ops_to_visit.pop() visited_ops.add(op) if op.type == 'Placeholder': potential_extra_sources.update(op.outputs) input_ops = [t.op for t in op.inputs if t not in stop_at_tensors] if not ignore_control_dependencies: input_ops = itertools.chain(input_ops, op.control_inputs) for input_op in input_ops: if input_op not in visited_ops: ops_to_visit.add(input_op) return potential_extra_sources def get_func_graph_for_name(graph, func_name): """Returns the FuncGraph associated to the given func_name if possible.""" outer_graph = graph while graph is not None: func = graph._get_function(str(func_name)) # pylint: disable=protected-access if func is not None: if hasattr(func, 'graph'): return func.graph # `outer_graph` may not be the same as `ops.get_default_graph()` e.g. # in the case of nested if ops or when the gradient is being computed # from inside a Defun. We build the `func_graph` with `outer_graph` as its # outer graph. with outer_graph.as_default(): # This is a _DefinedFunction. func_graph = ( function_def_to_graph.function_def_to_graph(func.definition)) if func_graph is not None: return func_graph if hasattr(graph, 'outer_graph'): graph = graph.outer_graph else: raise ValueError( 'Function {} does not exist in the graph.'.format(func_name)) class _UnexpectedPlaceholderError(Exception): def __init__(self, op, func_graph_name): tensor = op.outputs[0] msg = 'An unexpected placeholder was encountered ({})'.format(tensor) super(_UnexpectedPlaceholderError, self).__init__(msg) self.tensor = tensor self.func_graph_name = func_graph_name class _UnexpectedTableError(Exception): def __init__(self, op, func_graph_name): msg = 'An unexpected initializable table was encountered ({})'.format(op) super(_UnexpectedTableError, self).__init__(msg) self.op = op self.func_graph_name = func_graph_name def _reraise_unexpected_error(func): """A decorator that reraises certain exceptions with modified msg and type.""" def wrapper(self, tensor_or_op): """Wrapper when calling func to re-raise exceptions.""" try: return func(self, tensor_or_op) except _UnexpectedPlaceholderError as e: if e.func_graph_name: raise ValueError( 'The tensor_or_op {} depended on a placeholder ({}) that is part ' 'of a tf.function graph ({}), this is not supported. This may be a ' 'result of calling a tf.Transform analyzer in a tf.function' ''.format(tensor_or_op, e.tensor, e.func_graph_name)) else: raise ValueError( 'The tensor_or_op {} depended on a placeholder ({}) that was not ' 'in the input_signature. This may have be caused by manually ' 'adding a placeholder to the graph'.format(tensor_or_op, e.tensor)) except _UnexpectedTableError as e: if e.func_graph_name: raise ValueError( 'The tensor_or_op {} depended on an initializable table ({}) that ' 'is part of a tf.function graph ({}), this is not supported. This' ' may be a result of initializing a table in a tf.function' ''.format(tensor_or_op, e.op, e.func_graph_name)) else: raise ValueError( 'The tensor_or_op {} depended on an initializable table ({}) that ' 'was not tracked by the graph analysis. This may be caused by ' 'adding an initializable table without adding its initializer to ' 'the collection tf.GraphKeys.TABLE_INITIALIZERS'.format( tensor_or_op, e.op)) return wrapper class _AnalysisResult( tfx_namedtuple.namedtuple('_AnalysisResult', ['is_ready_to_run', 'path', 'dependent_sources']) ): pass class _SourceInfo( tfx_namedtuple.namedtuple('_SourceInfo', ['is_ready_to_run', 'name'])): pass class _GraphAnalyzer(object): """Class that analyzes a graph to determine readiness of tensors.""" def __init__(self, source_info_dict, translate_path_fn, graph): """Init method for _GraphAnalyzer. Args: source_info_dict: A dict from `Tensor Reference` or `Operation` to `_SourceInfo`. translate_path_fn: A function with the signature: (identifier, parents) -> Any which will be used to construct a unique path for a given `Tensor`. graph: A `tf.Graph` which the given tensors belong to. """ self._memoized_analyze_tensor_result = {} self._source_info_dict = source_info_dict self._translate_path_fn = translate_path_fn self._graph = graph def _get_parents(self, tensor_or_op): """Get the parents of the given `tensor_or_op`.""" if tf_utils.hashable_tensor_or_op(tensor_or_op) in self._source_info_dict: return [] # func_graph_name is not None only if the graph is a FuncGraph. func_graph_name = getattr(self._graph, 'name', None) if isinstance(tensor_or_op, tf.Operation): if tensor_or_op.type in _INITIALIZABLE_TABLE_OP_TYPES: raise _UnexpectedTableError(tensor_or_op, func_graph_name) if tensor_or_op.type == 'Placeholder': # If we're not in the context of a tf.function, this is an error. if func_graph_name is None: raise _UnexpectedPlaceholderError(tensor_or_op, func_graph_name) # If we're in the context of a tf.function and this op is part of its # inputs, that's expected. if tensor_or_op not in [x.op for x in self._graph.inputs]: raise _UnexpectedPlaceholderError(tensor_or_op, func_graph_name) parents = list( itertools.chain(tensor_or_op.inputs, tensor_or_op.control_inputs)) elif isinstance(tensor_or_op, tf.Tensor): parents = [tensor_or_op.op] else: raise TypeError('Expected Tensor or Operation, got {} of type {}'.format( tensor_or_op, type(tensor_or_op))) return parents def _compute_analysis_results_for_func_attributes(self, tensor_or_op, parent_analysis_results): """Analyzes `FuncGraph`s if tensor_or_op has them as attributes. This functionality is added to support `Operation`s such as PartitionedCall (tf.function call) and control flow ops which use `func` attributes. These func attributes are references to `FuncGraph`s which can also be analyzed, and the result of their analysis can be used as additional information for the current node (`tensor_or_op`). Since `FuncGraph`s are completely different graphs than the one that this _GraphAnalyzer is analyzing, their analysis wouldn't be taken into account when analysing the current graph even though they will affect the runtime results of running it. This is why we have to manually analyze those sub-graphs as well as the main graph when computing graph information such as dependent_inputs, unique_path, etc. Args: tensor_or_op: A `Tensor` or `Operation` object. parent_analysis_results: A list of `_AnalysisResult`s, results of analysis of the parents of tensor_or_op. Returns: A list of `_AnalysisResult`s, the results of analysis of `tensor_or_op`'s func attributes. All `Tensor`s in dependent_sources belong to self._graph. """ if not isinstance(tensor_or_op, tf.Operation): return [] func_attributes = [ attr.name for attr in tensor_or_op.op_def.attr if attr.type == 'func' ] func_names = [tensor_or_op.get_attr(str(n)).name for n in func_attributes] func_graphs = [get_func_graph_for_name(self._graph, n) for n in func_names] result = [] for func_graph in func_graphs: if not hasattr(func_graph, 'inputs'): # Since the body of the graph is not visible we insert a random string # to the path in order to reflect that we don't know its full contents. result.append( _AnalysisResult( is_ready_to_run=True, path=self._translate_path_fn(uuid.uuid4().hex), dependent_sources={})) continue op_inputs = list( itertools.chain(tensor_or_op.inputs, tensor_or_op.control_inputs)) assert len(op_inputs) == len(parent_analysis_results), ( op_inputs, parent_analysis_results) func_graph_inputs_ready = [ (next_input, r.is_ready_to_run) for (next_input, r) in zip(func_graph.inputs, parent_analysis_results) ] infos = { tf_utils.hashable_tensor_or_op(t): _SourceInfo(ready, 'FuncGraphInput[{}]'.format(idx)) for idx, (t, ready) in enumerate(func_graph_inputs_ready) } func_graph_analyzer = _GraphAnalyzer(infos, self._translate_path_fn, func_graph) analyzed_list = [ func_graph_analyzer.analyze_tensor(t) for t in func_graph.outputs ] if len(tensor_or_op.inputs) == len(func_graph.inputs): tensor_pairs = zip(tensor_or_op.inputs, func_graph.inputs) else: # Control flow ops such as while store this information in captures. tensor_pairs = func_graph.captures tensor_map = { tf_utils.hashable_tensor_or_op(b): a for a, b in tensor_pairs } # Make sure that the dependent sources Tensors are translated from the # FuncGraph to the outer graph in order to align with the rest of the # traversal. for analysis in analyzed_list: translated_dependent_sources = { tf_utils.hashable_tensor_or_op(tensor_map[s]) for s in analysis.dependent_sources if s in tensor_map } result.append( analysis._replace(dependent_sources=translated_dependent_sources)) return result def _compute_analysis_result(self, tensor_or_op, parent_analysis_results): """Compute analysis result for a tensor or op with its parent results.""" hashable = tf_utils.hashable_tensor_or_op(tensor_or_op) if hashable in self._source_info_dict: source_info = self._source_info_dict[hashable] # source_info.name may be None but that just means that it relies on an # output of a previous analyzer, so that's ok. return _AnalysisResult( is_ready_to_run=source_info.is_ready_to_run, path=self._translate_path_fn(source_info.name), dependent_sources={hashable}) func_graphs_analysis_results = ( self._compute_analysis_results_for_func_attributes( tensor_or_op, parent_analysis_results)) result = _AnalysisResult( is_ready_to_run=all( analysis_result.is_ready_to_run for analysis_result in (parent_analysis_results + func_graphs_analysis_results)), path=self._translate_path_fn( tensor_or_op, parents=[ parent_analysis_result.path for parent_analysis_result in parent_analysis_results ] + [func_result.path for func_result in func_graphs_analysis_results]), dependent_sources=set()) for parent_analysis_result in parent_analysis_results: result.dependent_sources.update(parent_analysis_result.dependent_sources) for func_result in func_graphs_analysis_results: result.dependent_sources.update(func_result.dependent_sources) return result def analyze_tensor(self, tensor_or_op): """Analyzes the `tensor_or_op` for its dependencies and readiness. Computes the transitive dependencies of a tensor or operation and decides whether it is ready to run using iterative DFS. `source_info_dict` are used as terminal nodes. An error is thrown if a table or placeholder is reached: they must be set using source_info_dict. This function is memoized using the _memoized_analyze_tensor_result cache. Cycles are ignored (so a cycle is considered ready to run). Args: tensor_or_op: A `Tensor` or `Operation`. Returns: An _AnalysisResult which includes whether this op or tensor is ready to run, a path from it to its sources and its dependent sources from `source_info_dict`. Raises: _UnexpectedTableError: If an initializable table op is encountered. _UnexpectedPlaceholderError: If a placeholder is encountered. """ stack = collections.deque() # Note that because tensors are no longer hashable, we need to convert to # their reference in order to use them in sets or dicts. stack.append(tf_utils.hashable_tensor_or_op(tensor_or_op)) # Contains the nodes of the path starting from tensor_or_op to current # visiting node, used for loop detection. We assume that any loop is a # valid while loop and so it will be able to run as long as all the other # parents are ready. path = set() while stack: current = stack[-1] if current in self._memoized_analyze_tensor_result: stack.pop() continue path.add(current) parents = self._get_parents(tf_utils.deref_tensor_or_op(current)) parents = [parent for parent in map(tf_utils.hashable_tensor_or_op, parents) if parent not in path] if all( parent in self._memoized_analyze_tensor_result for parent in parents): parent_results = [ self._memoized_analyze_tensor_result[parent] for parent in parents ] current_result = self._compute_analysis_result( tf_utils.deref_tensor_or_op(current), parent_results) self._memoized_analyze_tensor_result[current] = current_result path.discard(stack.pop()) else: stack.extend(parents) return self._memoized_analyze_tensor_result[tf_utils.hashable_tensor_or_op( tensor_or_op)] def ready_to_run(self, tensor_or_op): """Determine if a given tensor or op is ready to run. A tensor is ready to run if every tensor in all its transitive dependencies are set to `True` in `known_ready`. Note that if a placeholder is encountered, this will result in an error as it is assumed that all placeholders are keys in `known_ready`. This is to avoid unexpected behavior when the user creates placeholders (as opposed to placeholders created by the tf.Transform framework). Similarly encountering a Table op is an error because a table should be a key in `known_ready` (in the case of analyzing the main session run) or should not be encountered (in the case of analyzing the graph init run). Args: tensor_or_op: A `Tensor`, `SparseTensor`, `RaggedTensor` or `Operation` Returns: A bool indicating whether then tensor is ready to run. Raises: ValueError: If a placeholder or table is encountered. _UnexpectedTableError: If an initializable table op is encountered. _UnexpectedPlaceholderError: If a placeholder is encountered. """ if not isinstance( tensor_or_op, (tf.Tensor, tf.SparseTensor, tf.RaggedTensor, tf.Operation)): raise TypeError( 'Expected Tensor, SparseTensor, RaggedTensor, or Operation got {} of type {}' .format(tensor_or_op, type(tensor_or_op))) return all( self.analyze_tensor(component).is_ready_to_run for component in _decompose_tensor_or_op(tensor_or_op)) def get_unique_path(self, tensor): """Gets the analyzed path from the tensor to its root(s). This path is defined recursively as: Path(root) := translate_path_fn(root) Path(x) := translate_path_fn( x, [translate_path_fn(p) for p in parents(x)]) When root is defined as a tensor that has no parents. Args: tensor: A `Tensor` for which a path should be computed. Returns: The result of translate_path_fn on the computed path as described above. Raises: TypeError: if the given tensor is not of type `Tensor` _UnexpectedTableError: If an initializable table op is encountered. _UnexpectedPlaceholderError: If a placeholder is encountered. """ if not isinstance(tensor, tf.Tensor): raise TypeError('Expected Tensor got {} of type {}'.format( tensor, type(tensor))) return self.analyze_tensor(tensor).path def _set_unique_value_in_dict(input_dict, key, value): assert value not in input_dict.values(), value input_dict[tf_utils.hashable_tensor_or_op(key)] = value class InitializableGraphAnalyzer(object): """Determines which tensors will be ready when running the graph. Determines which tensors from `fetches` are ready to run, using following algorithm. 1. Determine which table initializers are ready to run. A table initializer is an element of the TABLE_INITIALIZERS collection and it is ready to run if all the tensors it depends on are set to ready in `replaced_tensors_ready`. 2. Determine which of `fetches` are ready to run. A fetch is ready to run if it only depends on tensors in `feeds` and tensors that are set to ready in `replaced_tensors_ready`. """ def __init__(self, graph, input_signature, replaced_tensors_ready, translate_path_fn=None): """Init method for InitializableGraphAnalyzer. Args: graph: a `Graph`. input_signature: A dict whose keys are strings and values are `Tensor`s, `SparseTensor`s, or `RaggedTensor`s. replaced_tensors_ready: a list of `Tensor`, `SparseTensor`s, or `RaggedTensor`s, bool pairs indicating whether the `Tensor`, `SparseTensor`s, or `RaggedTensor`s is ready in this phase. translate_path_fn: (Optional) A function with the signature: (identifier, optional(parents)) -> Any which will be used to construct a unique path for a given `Tensor`. Raises: ValueError: If unexpected placeholders or tables are encountered, or table initializers do not have the expected structure in the graph. """ if translate_path_fn is None: translate_path_fn = lambda x, parents=None: None self._ready_table_initializers = [] self._input_signature = input_signature replaced_tensors_ready = {tf_utils.hashable_tensor_or_op(t): ready for t, ready in replaced_tensors_ready} initial_source_infos_dict = self._make_source_infos_dict( {}, replaced_tensors_ready) # Determine which table initializers are ready, based on the replaced # tensors. Since no input tensors are fed during table initialization, we do # not set the value of any tensors in `input_signature`. graph_analyzer_for_table_init = _GraphAnalyzer(initial_source_infos_dict, translate_path_fn, graph) complete_source_info_dict = self._make_source_infos_dict( input_signature, replaced_tensors_ready) for table_init_op_or_tensor in graph.get_collection( tf.compat.v1.GraphKeys.TABLE_INITIALIZERS): table_init_op, table_input_ops = ( self._get_table_init_op_and_inputs(table_init_op_or_tensor)) source_info = self._get_table_init_op_source_info( table_init_op, graph_analyzer_for_table_init, translate_path_fn) for key in table_input_ops: complete_source_info_dict[tf_utils.hashable_tensor_or_op( key)] = source_info if source_info.is_ready_to_run: self._ready_table_initializers.append(table_init_op_or_tensor) # Now determine which tensors are ready to run once the table has been # initialized. self._graph_analyzer = _GraphAnalyzer(complete_source_info_dict, translate_path_fn, graph) def _get_table_init_op_and_inputs(self, table_init_op_or_tensor): """Get a tuple of table init op and keys for its input ops.""" # If a TF2 exported SavedModel with a table is loaded inside the # preprocessing_fn, the TABLE_INITIALIZERS collection of the outer graph # contains a Tensor whose parent op is of type StatefulPartitionedCall. # The nested func graph for this StatefulPartitionedCall contains the # table initializer. if (isinstance(table_init_op_or_tensor, tf.Tensor) and table_init_op_or_tensor.op.type == 'StatefulPartitionedCall'): result = (table_init_op_or_tensor.op, [input_t.op for input_t in table_init_op_or_tensor.op.inputs]) else: assert isinstance(table_init_op_or_tensor, tf.Operation) # We are using the table init op information and the table op information, # since that is a unique description of the table op. table_ops = [] for input_t in table_init_op_or_tensor.inputs: # One of the inputs to the initializer op should be the table op. If # no table op is found, (as in the case of a StatefulPartitionedCall) # all inputs are added to the source dict. if input_t.dtype == tf.resource: table_ops.append(input_t.op) assert len(table_ops) == 1 result = (table_init_op_or_tensor, [table_ops[0]]) return result def _make_source_infos_dict(self, input_signature, replaced_tensors_ready): """Builds a dictionary from source tensors to _SourceInfos. This dictionary stores information about the sources of the graph. Each tensor in replaced_tensors_ready is a source whose readiness is known and has no name. Each tensor (or component of a tensor) in input_signature is ready to run and has a name determined by the signature. Args: input_signature: A dict whose keys are strings and values are `Tensor`s, `SparseTensor`s, or `RaggedTensor`s. replaced_tensors_ready: a dict from `Tensor`, `SparseTensor`s, or `RaggedTensor`s to bool indicating whether the tensor is ready in this phase. Returns: a dictionary from source tensors to _SourceInfos. """ result = {} for tensor_or_op, is_ready in six.iteritems(replaced_tensors_ready): for component in _decompose_tensor_or_op( tf_utils.deref_tensor_or_op(tensor_or_op)): result[tf_utils.hashable_tensor_or_op(component)] = _SourceInfo( is_ready, None) for name, tensor in six.iteritems(input_signature): if isinstance(tensor, tf.Tensor): _set_unique_value_in_dict(result, tensor, _SourceInfo(True, '{}$tensor'.format(name))) elif isinstance(tensor, composite_tensor.CompositeTensor): for idx, tensor_component in enumerate(_decompose_tensor_or_op(tensor)): _set_unique_value_in_dict( result, tensor_component, _SourceInfo(True, '{}$composite_tensor_{}'.format(name, idx))) else: raise TypeError( 'Expected Tensor, or CompositeTensor, got {} of type {}'.format( tensor, type(tensor))) return result def _get_table_init_op_source_info(self, table_init_op, graph_analyzer, translate_path_fn): """Gets a _SourceInfo for a given table init op.""" if table_init_op.type not in _TABLE_INIT_OP_TYPES: raise ValueError( 'Table initializer {} did not have expected op type'.format( table_init_op)) if not table_init_op.inputs: raise ValueError( 'Table initializer {} did not have expected number if inputs ' '(expected >= 1 inputs, got 0)'.format(table_init_op)) table_op = table_init_op.inputs[0].op table_init_inputs = table_init_op.inputs[1:] try: ready = all(map(graph_analyzer.ready_to_run, table_init_inputs)) path = translate_path_fn( table_op, parents=list(map(graph_analyzer.get_unique_path, table_init_inputs))) except _UnexpectedPlaceholderError as e: if e.func_graph_name: raise e raise ValueError( 'The table initializer {} depended on a placeholder ({}). Note ' 'placeholders will not be fed during table initialization'.format( table_init_op, e.tensor)) except _UnexpectedTableError as e: if e.func_graph_name: raise e raise ValueError( 'The table initializer {} depended on an initializable table ({}). ' 'Note tables are initialized in one pass so a table initializer ' 'cannot depend on the output of an initializeable table'.format( table_init_op, e.op)) return _SourceInfo(ready, path) @property def ready_table_initializers(self): return self._ready_table_initializers @_reraise_unexpected_error def ready_to_run(self, tensor_or_op): """Determine if a given tensor or op is ready to run.""" return self._graph_analyzer.ready_to_run(tensor_or_op) @_reraise_unexpected_error def get_unique_path(self, tensor): """Gets the analyzed path from the tensor to its root(s). This path is defined recursively as: Path(root) := translate_path_fn(root) Path(x) := translate_path_fn( x, [translate_path_fn(p) for p in parents(x)]) When root is defined as a tensor that has no parents. Args: tensor: A `Tensor` for which a path should be computed. Returns: The result of translate_path_fn on the computed path as described above. """ return self._graph_analyzer.get_unique_path(tensor) @_reraise_unexpected_error def get_dependent_inputs(self, tensor_or_op): """Gets the inputs that the given `tensor_or_op` transitively depends on. Args: tensor_or_op: A `Tensor`, `SparseTensor`, `RaggedTensor` or `Operation`. Returns: A dict of name to `Tensor`, `SparseTensor`, or `RaggedTensor` (sub-dict of `input_signature`) that the given `tensor_or_op` depends on. Raises: TypeError: If `tensor_or_op` is of an unsupported type. """ if not isinstance( tensor_or_op, (tf.Tensor, tf.SparseTensor, tf.RaggedTensor, tf.Operation)): raise TypeError( 'Expected Tensor, SparseTensor, RaggedTensor or Operation got {} of ' 'type {}'.format(tensor_or_op, type(tensor_or_op))) dependents = set() for component in _decompose_tensor_or_op(tensor_or_op): dependents.update( self._graph_analyzer.analyze_tensor(component).dependent_sources) result = {} for name, tensor in six.iteritems(self._input_signature): if any( tf_utils.hashable_tensor_or_op(component) in dependents for component in _decompose_tensor_or_op(tensor)): result[name] = tensor return result def get_dependent_inputs(graph, input_tensors, output_tensors): """Returns tensors in input_tensors that (transitively) produce output_tensors. Args: graph: A `tf.Graph`. It could be the (intermediate) output tf graph in any transform phase (including phase 0 where no tensor replacement has yet happened). input_tensors: A dict of logical name to `tf.Tensor`, `tf.SparseTensor`, or `tf.RaggedTensor`. Logical name doesn't have any implications in this method and can be anything. In some cases it is the feature name corresponding to the input tensor. output_tensors: A dict of logical name to `tf.Tensor`, `tf.SparseTensor`, or `tf.RaggedTensor`, or a list of `tf.Tensor`, `tf.SparseTensor`, or `tf.RaggedTensor`. Returns: A dict of logical name to `tf.Tensor`, `tf.SparseTensor`, or `tf.RaggedTensor` that are filtered from input_tensors (transitively) producing output_tensors """ if isinstance(output_tensors, list): output_iterator = output_tensors else: output_iterator = six.itervalues(output_tensors) # Since this method may be called before all tensor replacements are ready, to # fulfill the precondition of InitializableGraphAnalyzer, we fake the # readiness of tensor replacements. Note that the readiness of replacement # tensors doesn't affect the correctness of dependencies tracing. tensor_sinks = graph.get_collection(analyzer_nodes.TENSOR_REPLACEMENTS) sink_tensors_ready = [(sink.tensor, False) for sink in tensor_sinks] graph_analyzer = InitializableGraphAnalyzer(graph, input_tensors, sink_tensors_ready) dependent_inputs = {} for output_tensor in output_iterator: dependent_inputs.update(graph_analyzer.get_dependent_inputs(output_tensor)) return { name: tensor for name, tensor in six.iteritems(input_tensors) if name in dependent_inputs }
41.309045
87
0.701904
794d4b246589362869921f1b0f9a4939b010b1a0
12,985
py
Python
tests/test_scheduler.py
YiSang665/redbeat
ee4a43c8ac9e8135b27b7fef766eda7519738244
[ "Apache-2.0" ]
746
2016-02-18T22:40:49.000Z
2022-03-30T22:20:25.000Z
tests/test_scheduler.py
YiSang665/redbeat
ee4a43c8ac9e8135b27b7fef766eda7519738244
[ "Apache-2.0" ]
178
2015-08-21T17:00:21.000Z
2022-03-22T00:33:09.000Z
tests/test_scheduler.py
YiSang665/redbeat
ee4a43c8ac9e8135b27b7fef766eda7519738244
[ "Apache-2.0" ]
118
2016-05-16T15:56:08.000Z
2022-03-15T00:33:39.000Z
import ssl import unittest from copy import deepcopy from datetime import datetime, timedelta from unittest.mock import Mock, patch import pytz from celery.beat import DEFAULT_MAX_INTERVAL from celery.schedules import schedstate, schedule from celery.utils.time import maybe_timedelta from redbeat import RedBeatScheduler from redbeat.schedulers import get_redis from tests.basecase import AppCase, RedBeatCase class mocked_schedule(schedule): def __init__(self, remaining): self._remaining = maybe_timedelta(remaining) self.run_every = timedelta(seconds=1) self.nowfun = datetime.utcnow def remaining_estimate(self, last_run_at): return self._remaining class mocked_expired_schedule(schedule): def __init__(self): self.nowfun = datetime.utcnow self.run_every = Mock() self.run_every.total_seconds.return_value = 1 def remaining_estimate(self, last_run_at): return None def is_due(self, last_run_at): return schedstate(False, None) due_now = mocked_schedule(0) due_next = mocked_schedule(1) class RedBeatSchedulerTestBase(RedBeatCase): def setUp(self): super().setUp() self.s = RedBeatScheduler(app=self.app) self.due_later = mocked_schedule(self.s.max_interval * 10) self.send_task = patch.object(self.s, 'send_task') self.send_task.start() def tearDown(self): self.send_task.stop() class test_RedBeatScheduler_schedule(RedBeatSchedulerTestBase): def test_empty_schedule(self): self.assertEqual(self.s.schedule, {}) def test_schedule_includes_current_and_next(self): due = self.create_entry(name='due', s=due_now).save() up_next = self.create_entry(name='up_next', s=due_next).save() up_next2 = self.create_entry(name='up_next2', s=due_next).save() later = self.create_entry(name='later', s=self.due_later).save() schedule = self.s.schedule self.assertEqual(len(schedule), 2) self.assertIn(due.name, schedule) self.assertEqual(due.key, schedule[due.name].key) self.assertIn(up_next.name, schedule) self.assertEqual(up_next.key, schedule[up_next.name].key) self.assertNotIn(up_next2.name, schedule) self.assertNotIn(later.name, schedule) class test_RedBeatScheduler_tick(RedBeatSchedulerTestBase): def test_empty(self): with patch.object(self.s, 'send_task') as send_task: sleep = self.s.tick() self.assertFalse(send_task.called) self.assertEqual(sleep, self.s.max_interval) def test_due_next_never_run(self): self.create_entry(name='next', s=due_next).save() with patch.object(self.s, 'send_task') as send_task: sleep = self.s.tick() self.assertFalse(send_task.called) self.assertLess(sleep, 1.0) def test_due_next_just_ran(self): e = self.create_entry(name='next', s=due_next) e.save().reschedule() with patch.object(self.s, 'send_task') as send_task: sleep = self.s.tick() self.assertFalse(send_task.called) self.assertLess(0.8, sleep) self.assertLess(sleep, 1.0) @unittest.skip('test is breaking and unsure what the correct behaviour should be') def test_due_next_never_run_tz_positive(self): self.app.timezone = pytz.timezone('Europe/Moscow') e = self.create_entry(name='next', s=due_next).save() with patch.object(self.s, 'send_task') as send_task: sleep = self.s.tick() send_task.assert_called_with(e.task, e.args, e.kwargs, **self.s._maybe_due_kwargs) # would be more correct to # self.assertFalse(send_task.called) self.assertEqual(sleep, 1.0) self.app.timezone = pytz.utc @unittest.skip('test is breaking and unsure what the correct behaviour should be') def test_due_next_never_run_tz_negative(self): self.app.timezone = pytz.timezone('America/Chicago') e = self.create_entry(name='next', s=due_next).save() with patch.object(self.s, 'send_task') as send_task: sleep = self.s.tick() send_task.assert_called_with(e.task, e.args, e.kwargs, **self.s._maybe_due_kwargs) # would be more correct to # self.assertFalse(send_task.called) self.assertEqual(sleep, 1.0) self.app.timezone = pytz.utc def test_due_later_task_never_run(self): self.create_entry(s=self.due_later).save() with patch.object(self.s, 'send_task') as send_task: sleep = self.s.tick() self.assertFalse(send_task.called) self.assertEqual(sleep, self.s.max_interval) def test_due_now(self): e = self.create_entry( name='now', args=[], s=due_now, last_run_at=datetime.utcnow() - timedelta(seconds=1), ).save() with patch.object(self.s, 'send_task') as send_task: sleep = self.s.tick() send_task.assert_called_with(e.task, e.args, e.kwargs, **self.s._maybe_due_kwargs) self.assertEqual(sleep, 1.0) def test_old_static_entries_are_removed(self): redis = self.app.redbeat_redis schedule = {'test': {'task': 'test', 'schedule': mocked_schedule(42)}} self.app.redbeat_conf.schedule = schedule self.s.setup_schedule() self.assertIn('test', self.s.schedule) self.assertIn('test', redis.smembers(self.app.redbeat_conf.statics_key)) self.app.redbeat_conf.schedule = {} self.s.setup_schedule() self.assertNotIn('test', self.s.schedule) self.assertNotIn('test', redis.smembers(self.app.redbeat_conf.statics_key)) def test_lock_timeout(self): self.assertEqual(self.s.lock_timeout, self.s.max_interval * 5) class NotSentinelRedBeatCase(AppCase): config_dict = { 'BROKER_URL': 'redis://', } def setup(self): self.app.conf.update(self.config_dict) def test_sentinel_scheduler(self): redis_client = get_redis(app=self.app) assert 'Sentinel' not in str(redis_client.connection_pool) class SentinelRedBeatCase(AppCase): config_dict = { 'REDBEAT_KEY_PREFIX': 'rb-tests:', 'redbeat_key_prefix': 'rb-tests:', 'BROKER_URL': 'redis-sentinel://redis-sentinel:26379/0', 'CELERY_RESULT_BACKEND': 'redis-sentinel://redis-sentinel:26379/1', } BROKER_TRANSPORT_OPTIONS = { 'sentinels': [('192.168.1.1', 26379), ('192.168.1.2', 26379), ('192.168.1.3', 26379)], 'service_name': 'master', 'socket_timeout': 0.1, } def setup(self): # celery3 self.app.conf.add_defaults(deepcopy(self.config_dict)) def test_sentinel_scheduler(self): self.app.conf.update({'BROKER_TRANSPORT_OPTIONS': self.BROKER_TRANSPORT_OPTIONS}) redis_client = get_redis(app=self.app) assert 'Sentinel' in str(redis_client.connection_pool) def test_sentinel_scheduler_options(self): for options in [ "BROKER_TRANSPORT_OPTIONS", "redbeat_redis_options", "REDBEAT_REDIS_OPTIONS", ]: config = deepcopy(self.config_dict) config[options] = self.BROKER_TRANSPORT_OPTIONS self.app.conf.clear() self.app.conf.update(config) redis_client = get_redis(app=self.app) assert 'Sentinel' in str(redis_client.connection_pool) class SeparateOptionsForSchedulerCase(AppCase): config_dict = { 'REDBEAT_KEY_PREFIX': 'rb-tests:', 'REDBEAT_REDIS_URL': 'redis-sentinel://redis-sentinel:26379/0', 'REDBEAT_REDIS_OPTIONS': { 'sentinels': [('192.168.1.1', 26379), ('192.168.1.2', 26379), ('192.168.1.3', 26379)], 'password': '123', 'service_name': 'master', 'socket_timeout': 0.1, }, 'CELERY_RESULT_BACKEND': 'redis-sentinel://redis-sentinel:26379/1', } def setup(self): # celery3 self.app.conf.add_defaults(deepcopy(self.config_dict)) def test_sentinel_scheduler(self): redis_client = get_redis(app=self.app) assert 'Sentinel' in str(redis_client.connection_pool) class SSLConnectionToRedis(AppCase): config_dict = { 'REDBEAT_KEY_PREFIX': 'rb-tests:', 'REDBEAT_REDIS_URL': 'rediss://redishost:26379/0', 'REDBEAT_REDIS_OPTIONS': { 'password': '123', 'service_name': 'master', 'socket_timeout': 0.1, }, 'CELERY_RESULT_BACKEND': 'redis-sentinel://redis-sentinel:26379/1', 'REDBEAT_REDIS_USE_SSL': { 'ssl_cert_reqs': ssl.CERT_REQUIRED, 'ssl_keyfile': '/path/to/file.key', 'ssl_certfile': '/path/to/cert.pem', 'ssl_ca_certs': '/path/to/ca.pem', }, } def setup(self): # celery3 self.app.conf.add_defaults(deepcopy(self.config_dict)) def test_ssl_connection_scheduler(self): redis_client = get_redis(app=self.app) assert 'SSLConnection' in str(redis_client.connection_pool) assert redis_client.connection_pool.connection_kwargs['ssl_cert_reqs'] == ssl.CERT_REQUIRED assert redis_client.connection_pool.connection_kwargs['ssl_keyfile'] == '/path/to/file.key' assert redis_client.connection_pool.connection_kwargs['ssl_certfile'] == '/path/to/cert.pem' assert redis_client.connection_pool.connection_kwargs['ssl_ca_certs'] == '/path/to/ca.pem' class SSLConnectionToRedisDefaultBrokerSSL(AppCase): config_dict = { 'REDBEAT_KEY_PREFIX': 'rb-tests:', 'REDBEAT_REDIS_URL': 'rediss://redishost:26379/0', 'REDBEAT_REDIS_OPTIONS': { 'password': '123', 'service_name': 'master', 'socket_timeout': 0.1, }, 'CELERY_RESULT_BACKEND': 'redis-sentinel://redis-sentinel:26379/1', 'BROKER_USE_SSL': { 'ssl_cert_reqs': ssl.CERT_REQUIRED, 'ssl_keyfile': '/path/to/file.key', 'ssl_certfile': '/path/to/cert.pem', 'ssl_ca_certs': '/path/to/ca.pem', }, } def setup(self): # celery3 self.app.conf.add_defaults(deepcopy(self.config_dict)) def test_ssl_connection_scheduler(self): redis_client = get_redis(app=self.app) assert 'SSLConnection' in str(redis_client.connection_pool) assert redis_client.connection_pool.connection_kwargs['ssl_cert_reqs'] == ssl.CERT_REQUIRED assert redis_client.connection_pool.connection_kwargs['ssl_keyfile'] == '/path/to/file.key' assert redis_client.connection_pool.connection_kwargs['ssl_certfile'] == '/path/to/cert.pem' assert redis_client.connection_pool.connection_kwargs['ssl_ca_certs'] == '/path/to/ca.pem' class SSLConnectionToRedisNoCerts(AppCase): config_dict = { 'REDBEAT_KEY_PREFIX': 'rb-tests:', 'REDBEAT_REDIS_URL': 'rediss://redishost:26379/0', 'REDBEAT_REDIS_OPTIONS': { 'password': '123', 'service_name': 'master', 'socket_timeout': 0.1, }, 'CELERY_RESULT_BACKEND': 'redis-sentinel://redis-sentinel:26379/1', 'REDBEAT_REDIS_USE_SSL': True, } def setup(self): # celery3 self.app.conf.add_defaults(deepcopy(self.config_dict)) def test_ssl_connection_scheduler(self): redis_client = get_redis(app=self.app) assert 'SSLConnection' in str(redis_client.connection_pool) assert redis_client.connection_pool.connection_kwargs['ssl_cert_reqs'] == ssl.CERT_REQUIRED assert 'ssl_keyfile' not in redis_client.connection_pool.connection_kwargs assert 'ssl_certfile' not in redis_client.connection_pool.connection_kwargs assert 'ssl_ca_certs' not in redis_client.connection_pool.connection_kwargs class RedBeatLockTimeoutDefaultValues(RedBeatCase): def test_no_values(self): scheduler = RedBeatScheduler(app=self.app) assert DEFAULT_MAX_INTERVAL * 5 == scheduler.lock_timeout assert DEFAULT_MAX_INTERVAL == scheduler.max_interval class RedBeatLockTimeoutCustomMaxInterval(RedBeatCase): config_dict = { 'beat_max_loop_interval': 5, } def test_no_lock_timeout(self): scheduler = RedBeatScheduler(app=self.app) assert self.config_dict['beat_max_loop_interval'] * 5 == scheduler.lock_timeout assert self.config_dict['beat_max_loop_interval'] == scheduler.max_interval class RedBeatLockTimeoutCustomAll(RedBeatCase): config_dict = { 'beat_max_loop_interval': 7, 'redbeat_lock_timeout': 13, } def test_custom_lock_timeout(self): scheduler = RedBeatScheduler(app=self.app) assert self.config_dict['beat_max_loop_interval'] == scheduler.max_interval assert self.config_dict['redbeat_lock_timeout'] == scheduler.lock_timeout
35.381471
100
0.662149
794d4b99e4b7c216595006a211b0fd2b25f6ef13
650
py
Python
src/devices/__init__.py
BMW-Group-Quantum/QUARK
739a55b10bd8d841410dc437bc10ca289bfd3005
[ "Apache-2.0" ]
12
2022-03-01T13:26:10.000Z
2022-03-28T07:59:25.000Z
src/devices/__init__.py
BMW-Group-Quantum/QUARK
739a55b10bd8d841410dc437bc10ca289bfd3005
[ "Apache-2.0" ]
null
null
null
src/devices/__init__.py
BMW-Group-Quantum/QUARK
739a55b10bd8d841410dc437bc10ca289bfd3005
[ "Apache-2.0" ]
null
null
null
# Copyright 2022 The QUARK Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Module containing all Devices"""
40.625
75
0.749231
794d4f0c4ca2b1a25432e86cc0334eae1b098170
2,838
py
Python
tests/test_reduce_max.py
bendecoste/tf-encrypted
00c9b80435411299a518f777de644c61d64cceea
[ "Apache-2.0" ]
null
null
null
tests/test_reduce_max.py
bendecoste/tf-encrypted
00c9b80435411299a518f777de644c61d64cceea
[ "Apache-2.0" ]
null
null
null
tests/test_reduce_max.py
bendecoste/tf-encrypted
00c9b80435411299a518f777de644c61d64cceea
[ "Apache-2.0" ]
null
null
null
import unittest import numpy as np import tensorflow as tf import tf_encrypted as tfe class TestReduceMax(unittest.TestCase): def setUp(self): tf.reset_default_graph() def test_reduce_max_1d(self): t = np.array([1, 2, 3, 4, 5, 6, 7, 8]).astype(float) with tf.Session() as sess: out_tf = tf.reduce_max(t) expected = sess.run(out_tf) with tfe.protocol.SecureNN() as prot: b = prot.define_private_variable(tf.constant(t)) out_tfe = prot.reduce_max(b) with tfe.Session() as sess: sess.run(tf.global_variables_initializer()) for _ in range(2): actual = sess.run(out_tfe.reveal(), tag='test_1d') np.testing.assert_array_equal(actual, expected) def test_reduce_max_2d_axis0(self): t = np.array([1, 2, 3, 4, 5, 6, 7, 8]).reshape(2, 4).astype(float) with tf.Session() as sess: out_tf = tf.reduce_max(t, axis=0) expected = sess.run(out_tf) with tfe.protocol.SecureNN() as prot: b = prot.define_private_variable(tf.constant(t)) out_tfe = prot.reduce_max(b, axis=0) with tfe.Session() as sess: sess.run(tf.global_variables_initializer()) for _ in range(2): actual = sess.run(out_tfe.reveal(), tag='test_2d_axis0') np.testing.assert_array_equal(actual, expected) def test_reduce_max_2d_axis1(self): t = np.array([1, 2, 3, 4, 5, 6, 7, 8]).reshape(2, 4).astype(float) with tf.Session() as sess: out_tf = tf.reduce_max(t, axis=1) expected = sess.run(out_tf) with tfe.protocol.SecureNN() as prot: b = prot.define_private_variable(tf.constant(t)) out_tfe = prot.reduce_max(b, axis=1) with tfe.Session() as sess: sess.run(tf.global_variables_initializer()) for _ in range(2): actual = sess.run(out_tfe.reveal(), tag='test_2d_axis1') np.testing.assert_array_equal(actual, expected) def test_reduce_max_3d_axis0(self): t = np.array([1, 2, 3, 4, 5, 6, 7, 8]).reshape(2, 2, 2) with tf.Session() as sess: out = tf.reduce_max(t, axis=0) expected = sess.run(out) with tfe.protocol.SecureNN() as prot: b = prot.define_private_variable(tf.constant(t)) out_tfe = prot.reduce_max(b, axis=0) with tfe.Session() as sess: sess.run(tf.global_variables_initializer()) for _ in range(2): actual = sess.run(out_tfe.reveal(), tag='test_3d_axis0') np.testing.assert_array_equal(actual, expected) if __name__ == '__main__': unittest.main()
30.847826
76
0.575758
794d4f25e6aafbb98f1345b9ee3cfa1aeb1a5fff
5,501
py
Python
modules/tools/fuel_proxy/control_calibration_sanity_check.py
Gavin-Tao/apollo
0b12edf51f20bec189b8cdd5eaf26f8ffe305186
[ "Apache-2.0" ]
4
2019-12-27T05:37:20.000Z
2020-03-24T07:21:54.000Z
modules/tools/fuel_proxy/control_calibration_sanity_check.py
Gavin-Tao/apollo
0b12edf51f20bec189b8cdd5eaf26f8ffe305186
[ "Apache-2.0" ]
2
2020-06-11T12:48:33.000Z
2020-08-28T01:07:43.000Z
modules/tools/fuel_proxy/control_calibration_sanity_check.py
Gavin-Tao/apollo
0b12edf51f20bec189b8cdd5eaf26f8ffe305186
[ "Apache-2.0" ]
57
2019-11-22T05:56:22.000Z
2019-12-28T08:51:10.000Z
#!/usr/bin/env python3 ############################################################################### # Copyright 2019 The Apollo Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. ############################################################################### """ Sanity Check for calibration table data files. """ import collections import os import math import sys import colored_glog as glog import google.protobuf.text_format as text_format from cyber_py3.record import RecordReader import modules.common.configs.proto.vehicle_config_pb2 as vehicle_config_pb2 import fueling.common.email_utils as email_utils import fueling.common.proto_utils as proto_utils import fueling.common.record_utils as record_utils import fueling.control.common.multi_vehicle_utils as multi_vehicle_utils # could be a list ConfFile = 'vehicle_param.pb.txt' CHANNELS = {record_utils.CHASSIS_CHANNEL, record_utils.LOCALIZATION_CHANNEL} def list_records(path): glog.info("in list_records:%s" % path) records = [] for (dirpath, _, filenames) in os.walk(path): glog.info('filenames: %s' % filenames) glog.info('dirpath %s' % dirpath) for filename in filenames: end_file = os.path.join(dirpath, filename) glog.info("end_files: %s" % end_file) if record_utils.is_record_file(end_file): records.append(end_file) return records def missing_file(path): vehicles = multi_vehicle_utils.get_vehicle(path) glog.info("vehicles %s" % vehicles) for vehicle in vehicles: # config file conf = os.path.join(path, vehicle, ConfFile) glog.info("vehicles conf %s" % conf) if os.path.exists(conf) is False: glog.error('Missing configuration file in %s' % vehicle) return True # record file glog.info("list of records:" % list_records(os.path.join(path, vehicle))) if len(list_records(os.path.join(path, vehicle))) == 0: glog.error('No record files in %s' % vehicle) return True return False def parse_error(path): vehicles = multi_vehicle_utils.get_vehicle(path) pb_value = vehicle_config_pb2.VehicleConfig() for vehicle in vehicles: conf = os.path.join(path, vehicle, ConfFile) try: proto_utils.get_pb_from_text_file(conf, pb_value) return False except text_format.ParseError: glog.error('Error: Cannot parse %s as binary or text proto' % conf) return True def check_vehicle_id(conf): vehicle_id = conf.vehicle_param.vehicle_id if vehicle_id.HasField('vin') or vehicle_id.HasField('plate') \ or vehicle_id.HasField('other_unique_id'): return True glog.error("Error: No vehicle ID") return False def missing_field(path): vehicles = multi_vehicle_utils.get_vehicle(path) glog.info("vehicles in missing field: %s" % vehicles) for vehicle in vehicles: conf_file = os.path.join(path, vehicle, ConfFile) glog.info("conf_file: %s" % conf_file) # reset for each vehicle to avoid overwrite pb_value = vehicle_config_pb2.VehicleConfig() conf = proto_utils.get_pb_from_text_file(conf_file, pb_value) glog.info("vehicles conf %s" % conf) if not check_vehicle_id(conf): return True # required field fields = [conf.vehicle_param.brake_deadzone, conf.vehicle_param.throttle_deadzone, conf.vehicle_param.max_acceleration, conf.vehicle_param.max_deceleration] for field in fields: if math.isnan(field): return True return False def missing_message_data(path, channels=CHANNELS): for record in list_records(path): glog.info("reading records %s" % record) reader = RecordReader(record) for channel in channels: glog.info("has %d messages" % reader.get_messagenumber(channel)) if reader.get_messagenumber(channel) == 0: return True return False def sanity_check(input_folder, email_receivers=None): err_msg = None if missing_file(input_folder): err_msg = "One or more files are missing in %s" % input_folder elif parse_error(input_folder): err_msg = "Confige file cannot be parsed in %s" % input_folder elif missing_field(input_folder): err_msg = "One or more field is missing in Confige file %s" % input_folder elif missing_message_data(input_folder): err_msg = "Messages are missing in records of %s" % input_folder else: glog.info("%s Passed sanity check." % input_folder) return True if email_receivers: title = 'Error occurred during data sanity check' email_utils.send_email_error(title, {'Error': err_msg}, email_receivers) glog.error(err_msg) return False if __name__ == '__main__': sanity_check(sys.argv[1])
35.720779
82
0.662425
794d4f69985c29dd774bc40886c9c6f02e9e7065
2,970
py
Python
generated/nidmm/nidmm/__init__.py
sukruthG/nimi-python
4f98addb04aea5748c8cac9f635065072e31a867
[ "MIT" ]
null
null
null
generated/nidmm/nidmm/__init__.py
sukruthG/nimi-python
4f98addb04aea5748c8cac9f635065072e31a867
[ "MIT" ]
null
null
null
generated/nidmm/nidmm/__init__.py
sukruthG/nimi-python
4f98addb04aea5748c8cac9f635065072e31a867
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- # This file was generated __version__ = '1.4.1.dev0' from nidmm.enums import * # noqa: F403,F401,H303 from nidmm.errors import DriverWarning # noqa: F401 from nidmm.errors import Error # noqa: F401 from nidmm.session import Session # noqa: F401 def get_diagnostic_information(): '''Get diagnostic information about the system state that is suitable for printing or logging returns: dict note: Python bitness may be incorrect when running in a virtual environment ''' import os import pkg_resources import platform import struct import sys def is_python_64bit(): return (struct.calcsize("P") == 8) def is_os_64bit(): return platform.machine().endswith('64') def is_venv(): return 'VIRTUAL_ENV' in os.environ info = {} info['os'] = {} info['python'] = {} info['driver'] = {} info['module'] = {} if platform.system() == 'Windows': try: import winreg as winreg except ImportError: import _winreg as winreg os_name = 'Windows' try: driver_version_key = winreg.OpenKey(winreg.HKEY_LOCAL_MACHINE, r"SOFTWARE\National Instruments\NI-DMM\CurrentVersion") driver_version = winreg.QueryValueEx(driver_version_key, "Version")[0] except WindowsError: driver_version = 'Unknown' elif platform.system() == 'Linux': os_name = 'Linux' driver_version = 'Unknown' else: raise SystemError('Unsupported platform: {}'.format(platform.system())) installed_packages = pkg_resources.working_set installed_packages_list = [{'name': i.key, 'version': i.version, } for i in installed_packages] info['os']['name'] = os_name info['os']['version'] = platform.version() info['os']['bits'] = '64' if is_os_64bit() else '32' info['driver']['name'] = "NI-DMM" info['driver']['version'] = driver_version info['module']['name'] = 'nidmm' info['module']['version'] = "1.4.1.dev0" info['python']['version'] = sys.version info['python']['bits'] = '64' if is_python_64bit() else '32' info['python']['is_venv'] = is_venv() info['python']['packages'] = installed_packages_list return info def print_diagnostic_information(): '''Print diagnostic information in a format suitable for issue report note: Python bitness may be incorrect when running in a virtual environment ''' info = get_diagnostic_information() row_format = ' {:<10} {}' for type in ['OS', 'Driver', 'Module', 'Python']: typename = type.lower() print(type + ':') for item in info[typename]: if item != 'packages': print(row_format.format(item.title() + ':', info[typename][item])) print(' Installed Packages:') for p in info['python']['packages']: print((' ' * 8) + p['name'] + '==' + p['version']) return info
30.618557
130
0.620539
794d50b7f7c975490689806c72a22c61f196fdab
539
py
Python
src/dbus/busctl_tree.py
RedHatQE/rhsm-scenarion
0a22ff11a092d8a6d796267bee7424e93b38fad2
[ "Apache-2.0" ]
null
null
null
src/dbus/busctl_tree.py
RedHatQE/rhsm-scenarion
0a22ff11a092d8a6d796267bee7424e93b38fad2
[ "Apache-2.0" ]
null
null
null
src/dbus/busctl_tree.py
RedHatQE/rhsm-scenarion
0a22ff11a092d8a6d796267bee7424e93b38fad2
[ "Apache-2.0" ]
null
null
null
import os.path import tempfile from time import time from src.shell import run_cmd from src.testware.message import send_message from src.types import (StartOfScenario, EndOfScenario) timestamp=int(time()) scenario=os.path.basename(__file__) workdir=tempfile.mkdtemp(prefix="{0}-{1}.".format(scenario,timestamp)) print("working directory: {0}".format(workdir)) send_message(StartOfScenario(scenario,timestamp)) result = run_cmd("busctl",["tree","--no-pager"],workdir) send_message(result) send_message(EndOfScenario(scenario,timestamp))
31.705882
70
0.797774
794d520542a3908317588169bcdc01af809e7212
3,660
py
Python
bigbench/api/util.py
colinzhaoust/BIG-bench
34cf4e737879f2fed7642c8df7648fef47cc463b
[ "Apache-2.0" ]
null
null
null
bigbench/api/util.py
colinzhaoust/BIG-bench
34cf4e737879f2fed7642c8df7648fef47cc463b
[ "Apache-2.0" ]
null
null
null
bigbench/api/util.py
colinzhaoust/BIG-bench
34cf4e737879f2fed7642c8df7648fef47cc463b
[ "Apache-2.0" ]
null
null
null
# Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Utility functions for BIG-bench. """ import re import abc import importlib import bigbench.api.task as task import bigbench.api.json_task as json_task def postprocess_output(text, max_length, stop_string, output_regex): """ Modify model output to satisfy stop_string and output_regex keywords. Args: text: A string or list of strings containing model outputs. max_length: Model output will be truncated to be at most this length. stop_string: Model output will be truncated to the shortest string which includes stop_string. If None, no truncation will be performed. e.g. if stop_string='.', the model output "The apple is on the ground. What is" will be truncated to "The apple is on the ground." before being returned. output_regex: Rather than returning the full output, return the first match to the python regular expression output_regex in the output string. If there is no match, an empty string will be returned. If output_regex=None, no regular expression matching will be performed. e.g. if output_regex=r"\\d+", and the model output is "= 42. 7*7 = 49. 7*8 =", then "42" will be returned. Returns: A string or list of strings, all processed as described for each argument. """ if isinstance(text, list): return [ postprocess_output(mo, max_length, stop_string, output_regex) for mo in text ] # Ensure it is a string (will convert from bytes, ... as needed) if not isinstance(text, str): text = str(text, "utf-8") # truncate at max_length if max_length: text = text[:max_length] # Remove all text after any stop_string if stop_string: index = text.find(stop_string) if index > 0: text = text[: index + len(stop_string)] # extract substring matching regex (empty string for no match) if output_regex: text = next(iter(re.findall(output_regex, text)), "") return text def load_json_task(json_path): this_task = json_task.JsonTask(json_path, shot_list=[2], verbose=True) return this_task def load_programmatic_task(task_module_name: str) -> task.Task: """Gets the bigbench.api.Task instance from the given module. Args: task_module_name: the module name to search, like bigbench.benchmark_tasks.simple_arithmetic Returns: Task instance """ this_task_class = importlib.import_module(task_module_name) task_class_index = -1 for i, k in enumerate(dir(this_task_class)): this_method = getattr(this_task_class, k) if isinstance(this_method, abc.ABCMeta): this_method_object = this_method() if issubclass(this_method_object.__class__, task.Task): task_class_index = i assert ( task_class_index != -1 ), f"Task class definition that extends bigbench.api. Task not found in module {task_module_name}" return getattr(this_task_class, dir(this_task_class)[task_class_index])()
35.882353
102
0.69235
794d5415eff95ca2978bcc4fd8a4cb4877716e84
5,460
py
Python
sdk/python/kfp/v2/compiler/main.py
algs/pipelines
ab63956f3a61d4d11b27ac26f097e1784588fed9
[ "Apache-2.0" ]
1
2020-12-20T11:26:00.000Z
2020-12-20T11:26:00.000Z
sdk/python/kfp/v2/compiler/main.py
algs/pipelines
ab63956f3a61d4d11b27ac26f097e1784588fed9
[ "Apache-2.0" ]
null
null
null
sdk/python/kfp/v2/compiler/main.py
algs/pipelines
ab63956f3a61d4d11b27ac26f097e1784588fed9
[ "Apache-2.0" ]
null
null
null
# Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """KFP SDK compiler CLI tool.""" import argparse import json import os import sys from typing import Any, Callable, List, Mapping, Optional import kfp.dsl as dsl import kfp.v2.compiler as compiler def parse_arguments() -> argparse.Namespace: """Parse command line arguments.""" parser = argparse.ArgumentParser() parser.add_argument( '--py', type=str, required=True, help='local absolute path to a py file.') parser.add_argument( '--function', type=str, help='The name of the function to compile if there are multiple.') parser.add_argument( '--pipeline-root', type=str, help='The root of the pipeline outputs.') parser.add_argument( '--pipeline-parameters', type=json.loads, help='The pipeline parameters in JSON dict format.') parser.add_argument( '--namespace', type=str, help='The namespace for the pipeline function') parser.add_argument( '--output', type=str, required=True, help='local path to the output PipelineJob json file.') parser.add_argument( '--disable-type-check', action='store_true', help='disable the type check, default is enabled.') args = parser.parse_args() return args def _compile_pipeline_function(pipeline_funcs: List[Callable], function_name: Optional[str], pipeline_root: str, pipeline_parameters: Optional[Mapping[str, Any]], package_path: str, type_check: bool) -> None: """Compiles a pipeline function. Args: pipeline_funcs: A list of pipeline_functions. function_name: The name of the pipeline function to compile if there were multiple. pipeline_root: The root output directory for pipeline runtime. pipeline_parameters: The pipeline parameters as a dict of {name: value}. package_path: The output path of the compiled result. type_check: Whether to enable the type checking. """ if len(pipeline_funcs) == 0: raise ValueError( 'A function with @dsl.pipeline decorator is required in the py file.') if len(pipeline_funcs) > 1 and not function_name: func_names = [x.__name__ for x in pipeline_funcs] raise ValueError( 'There are multiple pipelines: %s. Please specify --function.' % func_names) if function_name: pipeline_func = next( (x for x in pipeline_funcs if x.__name__ == function_name), None) if not pipeline_func: raise ValueError('The function "%s" does not exist. ' 'Did you forget @dsl.pipeline decoration?' % function_name) else: pipeline_func = pipeline_funcs[0] compiler.Compiler().compile( pipeline_func=pipeline_func, pipeline_root=pipeline_root, pipeline_parameters=pipeline_parameters, package_path=package_path, type_check=type_check) class PipelineCollectorContext(): def __enter__(self): pipeline_funcs = [] def add_pipeline(func: Callable) -> Callable: pipeline_funcs.append(func) return func self.old_handler = dsl._pipeline._pipeline_decorator_handler dsl._pipeline._pipeline_decorator_handler = add_pipeline return pipeline_funcs def __exit__(self, *args): dsl._pipeline._pipeline_decorator_handler = self.old_handler def compile_pyfile(pyfile: str, function_name: Optional[str], pipeline_root: str, pipeline_parameters: Optional[Mapping[str, Any]], package_path: str, type_check: bool) -> None: """Compiles a pipeline written in a .py file. Args: pyfile: The path to the .py file that contains the pipeline definition. function_name: The name of the pipeline function. pipeline_root: The root output directory for pipeline runtime. pipeline_parameters: The pipeline parameters as a dict of {name: value}. package_path: The output path of the compiled result. type_check: Whether to enable the type checking. """ sys.path.insert(0, os.path.dirname(pyfile)) try: filename = os.path.basename(pyfile) with PipelineCollectorContext() as pipeline_funcs: __import__(os.path.splitext(filename)[0]) _compile_pipeline_function( pipeline_funcs=pipeline_funcs, function_name=function_name, pipeline_root=pipeline_root, pipeline_parameters=pipeline_parameters, package_path=package_path, type_check=type_check) finally: del sys.path[0] def main(): args = parse_arguments() if args.py is None: raise ValueError('The --py option must be specified.') compile_pyfile( pyfile=args.py, function_name=args.function, pipeline_root=args.pipeline_root, pipeline_parameters=args.pipeline_parameters, package_path=args.output, type_check=not args.disable_type_check, )
33.703704
80
0.692857
794d556512448efca66313cff0ad19af477ee0f0
2,319
py
Python
src/ui/mwx/gauge_dlg.py
xxao/papyrus
dec48453676d7c7dbc94a2fd7076bfc3630d8cf9
[ "MIT" ]
2
2019-07-27T13:04:16.000Z
2019-08-06T17:44:48.000Z
src/ui/mwx/gauge_dlg.py
xxao/papyrus
dec48453676d7c7dbc94a2fd7076bfc3630d8cf9
[ "MIT" ]
null
null
null
src/ui/mwx/gauge_dlg.py
xxao/papyrus
dec48453676d7c7dbc94a2fd7076bfc3630d8cf9
[ "MIT" ]
2
2019-11-13T09:22:36.000Z
2021-09-28T14:27:52.000Z
# Created byMartin.cz # Copyright (c) Martin Strohalm. All rights reserved. # import modules import time import wx from .constants import * class GaugeDlg(wx.Dialog): """Creates generic gauge dialog.""" def __init__(self, parent, id=-1, message="", title="", cancellation=None, style=wx.CAPTION): """Initializes a new instance of GaugeDlg.""" wx.Dialog.__init__(self, parent, id, title, style=style) self._cancellation = cancellation # init sizer self.Sizer = wx.BoxSizer(wx.VERTICAL) # add message self._message = wx.StaticText(self, -1, message) self._message.SetFont(wx.SMALL_FONT) self.Sizer.Add(self._message, 0, wx.ALIGN_LEFT | wx.TOP | wx.LEFT | wx.RIGHT, PANEL_SPACE_MAIN) # add gauge self._gauge = wx.Gauge(self, -1, 100, size=(300, GAUGE_HEIGHT)) self.Sizer.Add(self._gauge, 0, wx.ALL, PANEL_SPACE_MAIN) # add cancel button if self._cancellation is not None: cancel_butt = wx.Button(self, -1, "Cancel") cancel_butt.Bind(wx.EVT_BUTTON, self.OnCancel) self.Sizer.Add(cancel_butt, 0, wx.ALIGN_RIGHT | wx.RIGHT | wx.BOTTOM, PANEL_SPACE_MAIN) # display dialog self.Layout() self.Sizer.Fit(self) self.Centre() wx.Yield() def OnCancel(self, evt=None): """Handles cancel button.""" if self._cancellation is not None: self._cancellation.cancel() def SetMessage(self, message): """Sets message text.""" if self._message.GetLabel() != message: self._message.SetLabel(message) def SetRange(self, value): """Sets gauge range.""" if self._gauge.GetRange() != value: self._gauge.SetRange(value) def SetValue(self, value): """Sets gauge value.""" value = int(value) if self._gauge.GetValue() != value: self._gauge.SetValue(value) def Pulse(self): """Pulses gauge.""" self._gauge.Pulse() def Close(self): """Closes panel""" # self.MakeModal(False) self.Destroy()
27.282353
103
0.559293
794d5629fa8cd102dcef2d6037eb60215cd12a99
1,124
py
Python
Python/Google_News_Scrapper/app.py
tanishq-arya/Rotten-Scripts
93f62eb1213f739c13103610d3f502e1dc2d3790
[ "MIT" ]
4
2021-04-27T19:54:06.000Z
2021-09-28T15:34:12.000Z
Python/Google_News_Scrapper/app.py
tanishq-arya/Rotten-Scripts
93f62eb1213f739c13103610d3f502e1dc2d3790
[ "MIT" ]
null
null
null
Python/Google_News_Scrapper/app.py
tanishq-arya/Rotten-Scripts
93f62eb1213f739c13103610d3f502e1dc2d3790
[ "MIT" ]
null
null
null
import requests from xml.dom.minidom import parseString import pandas as pd def get_google_news_result(term, count): results = [] obj = parseString(requests.get( 'http://news.google.com/news?q=%s&output=rss' % term).text) items = obj.getElementsByTagName('item') # Storing the Titles and Links titles = list() links = list() for item in items[:count]: title, link = '', '' for node in item.childNodes: if node.nodeName == 'title': title = node.childNodes[0].data elif node.nodeName == 'link': link = node.childNodes[0].data titles.append(title) links.append(link) return titles, links if __name__ == '__main__': titleName = input("Enter the news title keyword: ") articleCount = int(input('Enter the number of article count: ')) titles, links = get_google_news_result(titleName, articleCount) news = {'title': titles, 'links': links } df = pd.DataFrame(news, columns=['title', 'links']) df.to_excel('{}_news_scrapper.xlsx'.format(titleName))
30.378378
68
0.614769
794d5737799db61159ce3a028686f8a2c419033f
736
py
Python
ros/src/f16/models/rocket/meshes/subdivide.py
HenryTingle/aae497-f19
29df124e31fcc1a1a8462211b1390e5a74f40cd9
[ "BSD-3-Clause" ]
4
2019-08-27T10:13:17.000Z
2020-02-12T13:05:35.000Z
ros/src/f16/models/rocket/meshes/subdivide.py
HenryTingle/aae497-f19
29df124e31fcc1a1a8462211b1390e5a74f40cd9
[ "BSD-3-Clause" ]
4
2019-08-31T21:57:41.000Z
2019-10-31T04:41:34.000Z
ros/src/f16/models/rocket/meshes/subdivide.py
HenryTingle/aae497-f19
29df124e31fcc1a1a8462211b1390e5a74f40cd9
[ "BSD-3-Clause" ]
17
2019-08-21T20:32:18.000Z
2019-10-28T00:12:27.000Z
import bpy for part in ['nose_cone', 'body', 'fin', 'plume']: bpy.ops.object.mode_set(mode='OBJECT') bpy.ops.object.select_by_type(type='MESH') bpy.ops.object.delete(use_global=False) print('importing', part) bpy.ops.import_mesh.stl(filepath="./stl/{:s}.stl".format(part)) print('subdividing') bpy.ops.object.select_all(action='DESELECT') print(list(bpy.data.objects)) bpy.ops.object.mode_set(mode='OBJECT') bpy.ops.object.select_by_type(type='MESH') bpy.ops.object.editmode_toggle() bpy.ops.mesh.subdivide() bpy.ops.mesh.subdivide() print('exporting') bpy.ops.export_mesh.stl(filepath="./stl/{:s}.stl".format(part)) print('exporting complete') bpy.ops.wm.quit_blender()
29.44
67
0.684783
794d575375ba3c5142a9767b21acf3bf35d20866
9,897
py
Python
Lib/fileinput.py
M-Spencer-94/configNOW
56828587253202089e77cfdfcf5329f2a7f09b3f
[ "PSF-2.0", "Apache-2.0", "MIT" ]
3
2019-07-09T20:02:48.000Z
2021-11-21T20:00:37.000Z
Lib/fileinput.py
M-Spencer-94/configNOW
56828587253202089e77cfdfcf5329f2a7f09b3f
[ "PSF-2.0", "Apache-2.0", "MIT" ]
null
null
null
Lib/fileinput.py
M-Spencer-94/configNOW
56828587253202089e77cfdfcf5329f2a7f09b3f
[ "PSF-2.0", "Apache-2.0", "MIT" ]
1
2019-04-11T11:27:01.000Z
2019-04-11T11:27:01.000Z
"""Helper class to quickly write a loop over all standard input files. Typical use is: import fileinput for line in fileinput.input(): process(line) This iterates over the lines of all files listed in sys.argv[1:], defaulting to sys.stdin if the list is empty. If a filename is '-' it is also replaced by sys.stdin. To specify an alternative list of filenames, pass it as the argument to input(). A single file name is also allowed. Functions filename(), lineno() return the filename and cumulative line number of the line that has just been read; filelineno() returns its line number in the current file; isfirstline() returns true iff the line just read is the first line of its file; isstdin() returns true iff the line was read from sys.stdin. Function nextfile() closes the current file so that the next iteration will read the first line from the next file (if any); lines not read from the file will not count towards the cumulative line count; the filename is not changed until after the first line of the next file has been read. Function close() closes the sequence. Before any lines have been read, filename() returns None and both line numbers are zero; nextfile() has no effect. After all lines have been read, filename() and the line number functions return the values pertaining to the last line read; nextfile() has no effect. All files are opened in text mode. If an I/O error occurs during opening or reading a file, the IOError exception is raised. If sys.stdin is used more than once, the second and further use will return no lines, except perhaps for interactive use, or if it has been explicitly reset (e.g. using sys.stdin.seek(0)). Empty files are opened and immediately closed; the only time their presence in the list of filenames is noticeable at all is when the last file opened is empty. It is possible that the last line of a file doesn't end in a newline character; otherwise lines are returned including the trailing newline. Class FileInput is the implementation; its methods filename(), lineno(), fileline(), isfirstline(), isstdin(), nextfile() and close() correspond to the functions in the module. In addition it has a readline() method which returns the next input line, and a __getitem__() method which implements the sequence behavior. The sequence must be accessed in strictly sequential order; sequence access and readline() cannot be mixed. Optional in-place filtering: if the keyword argument inplace=1 is passed to input() or to the FileInput constructor, the file is moved to a backup file and standard output is directed to the input file. This makes it possible to write a filter that rewrites its input file in place. If the keyword argument backup=".<some extension>" is also given, it specifies the extension for the backup file, and the backup file remains around; by default, the extension is ".bak" and it is deleted when the output file is closed. In-place filtering is disabled when standard input is read. XXX The current implementation does not work for MS-DOS 8+3 filesystems. Performance: this module is unfortunately one of the slower ways of processing large numbers of input lines. Nevertheless, a significant speed-up has been obtained by using readlines(bufsize) instead of readline(). A new keyword argument, bufsize=N, is present on the input() function and the FileInput() class to override the default buffer size. XXX Possible additions: - optional getopt argument processing - specify open mode ('r' or 'rb') - fileno() - isatty() - read(), read(size), even readlines() """ import sys, os, stat __all__ = ["input","close","nextfile","filename","lineno","filelineno", "isfirstline","isstdin","FileInput"] _state = None DEFAULT_BUFSIZE = 8*1024 def input(files=None, inplace=0, backup="", bufsize=0): global _state if _state and _state._file: raise RuntimeError, "input() already active" _state = FileInput(files, inplace, backup, bufsize) return _state def close(): global _state state = _state _state = None if state: state.close() def nextfile(): if not _state: raise RuntimeError, "no active input()" return _state.nextfile() def filename(): if not _state: raise RuntimeError, "no active input()" return _state.filename() def lineno(): if not _state: raise RuntimeError, "no active input()" return _state.lineno() def filelineno(): if not _state: raise RuntimeError, "no active input()" return _state.filelineno() def isfirstline(): if not _state: raise RuntimeError, "no active input()" return _state.isfirstline() def isstdin(): if not _state: raise RuntimeError, "no active input()" return _state.isstdin() class FileInput: def __init__(self, files=None, inplace=0, backup="", bufsize=0): if type(files) == type(''): files = (files,) else: if files is None: files = sys.argv[1:] if not files: files = ('-',) else: files = tuple(files) self._files = files self._inplace = inplace self._backup = backup self._bufsize = bufsize or DEFAULT_BUFSIZE self._savestdout = None self._output = None self._filename = None self._lineno = 0 self._filelineno = 0 self._file = None self._isstdin = 0 self._backupfilename = None self._buffer = [] self._bufindex = 0 def __del__(self): self.close() def close(self): self.nextfile() self._files = () def __getitem__(self, i): try: line = self._buffer[self._bufindex] except IndexError: pass else: self._bufindex += 1 self._lineno += 1 self._filelineno += 1 return line if i != self._lineno: raise RuntimeError, "accessing lines out of order" line = self.readline() if not line: raise IndexError, "end of input reached" return line def nextfile(self): savestdout = self._savestdout self._savestdout = 0 if savestdout: sys.stdout = savestdout output = self._output self._output = 0 if output: output.close() file = self._file self._file = 0 if file and not self._isstdin: file.close() backupfilename = self._backupfilename self._backupfilename = 0 if backupfilename and not self._backup: try: os.unlink(backupfilename) except: pass self._isstdin = 0 self._buffer = [] self._bufindex = 0 def readline(self): try: line = self._buffer[self._bufindex] except IndexError: pass else: self._bufindex += 1 self._lineno += 1 self._filelineno += 1 return line if not self._file: if not self._files: return "" self._filename = self._files[0] self._files = self._files[1:] self._filelineno = 0 self._file = None self._isstdin = 0 self._backupfilename = 0 if self._filename == '-': self._filename = '<stdin>' self._file = sys.stdin self._isstdin = 1 else: if self._inplace: self._backupfilename = ( self._filename + (self._backup or ".bak")) try: os.unlink(self._backupfilename) except os.error: pass # The next few lines may raise IOError os.rename(self._filename, self._backupfilename) self._file = open(self._backupfilename, "r") try: perm = os.fstat(self._file.fileno())[stat.ST_MODE] except: self._output = open(self._filename, "w") else: fd = os.open(self._filename, os.O_CREAT | os.O_WRONLY | os.O_TRUNC, perm) self._output = os.fdopen(fd, "w") try: os.chmod(self._filename, perm) except: pass self._savestdout = sys.stdout sys.stdout = self._output else: # This may raise IOError self._file = open(self._filename, "r") self._buffer = self._file.readlines(self._bufsize) self._bufindex = 0 if not self._buffer: self.nextfile() # Recursive call return self.readline() def filename(self): return self._filename def lineno(self): return self._lineno def filelineno(self): return self._filelineno def isfirstline(self): return self._filelineno == 1 def isstdin(self): return self._isstdin def _test(): import getopt inplace = 0 backup = 0 opts, args = getopt.getopt(sys.argv[1:], "ib:") for o, a in opts: if o == '-i': inplace = 1 if o == '-b': backup = a for line in input(args, inplace=inplace, backup=backup): if line[-1:] == '\n': line = line[:-1] if line[-1:] == '\r': line = line[:-1] print "%d: %s[%d]%s %s" % (lineno(), filename(), filelineno(), isfirstline() and "*" or "", line) print "%d: %s[%d]" % (lineno(), filename(), filelineno()) if __name__ == '__main__': _test()
32.880399
75
0.605941
794d579ee5f6f21a34f3ae9332f4ad2f71431629
4,453
py
Python
mmedit/models/common/model_utils.py
rivergold/mmediting
fd972635c48bb065db29d1b5090592a87c7263d2
[ "Apache-2.0" ]
1
2021-04-30T23:08:16.000Z
2021-04-30T23:08:16.000Z
mmedit/models/common/model_utils.py
rivergold/mmediting
fd972635c48bb065db29d1b5090592a87c7263d2
[ "Apache-2.0" ]
null
null
null
mmedit/models/common/model_utils.py
rivergold/mmediting
fd972635c48bb065db29d1b5090592a87c7263d2
[ "Apache-2.0" ]
2
2021-09-07T05:21:18.000Z
2021-09-17T22:34:54.000Z
import numpy as np import torch def set_requires_grad(nets, requires_grad=False): """Set requies_grad for all the networks. Args: nets (nn.Module | list[nn.Module]): A list of networks or a single network. requires_grad (bool): Whether the networks require gradients or not """ if not isinstance(nets, list): nets = [nets] for net in nets: if net is not None: for param in net.parameters(): param.requires_grad = requires_grad def extract_bbox_patch(bbox, img, channel_first=True): """Extract patch from a given bbox Args: bbox (torch.Tensor | numpy.array): Bbox with (top, left, h, w). If `img` has batch dimension, the `bbox` must be stacked at first dimension. The shape should be (4,) or (n, 4). img (torch.Tensor | numpy.array): Image data to be extracted. If organized in batch dimension, the batch dimension must be the first order like (n, h, w, c) or (n, c, h, w). channel_first (bool): If True, the channel dimension of img is before height and width, e.g. (c, h, w). Otherwise, the img shape (samples in the batch) is like (h, w, c). Returns: (torch.Tensor | numpy.array): Extracted patches. The dimension of the \ output should be the same as `img`. """ def _extract(bbox, img): assert len(bbox) == 4 t, l, h, w = bbox if channel_first: img_patch = img[..., t:t + h, l:l + w] else: img_patch = img[t:t + h, l:l + w, ...] return img_patch input_size = img.shape assert len(input_size) == 3 or len(input_size) == 4 bbox_size = bbox.shape assert bbox_size == (4, ) or (len(bbox_size) == 2 and bbox_size[0] == input_size[0]) # images with batch dimension if len(input_size) == 4: output_list = [] for i in range(input_size[0]): img_patch_ = _extract(bbox[i], img[i:i + 1, ...]) output_list.append(img_patch_) if isinstance(img, torch.Tensor): img_patch = torch.cat(output_list, dim=0) else: img_patch = np.concatenate(output_list, axis=0) # standardize image else: img_patch = _extract(bbox, img) return img_patch def scale_bbox(bbox, target_size): """Modify bbox to target size. The original bbox will be enlarged to the target size with the original bbox in the center of the new bbox. Args: bbox (np.ndarray | torch.Tensor): Bboxes to be modified. Bbox can be in batch or not. The shape should be (4,) or (n, 4). target_size (tuple[int]): Target size of final bbox. Returns: (np.ndarray | torch.Tensor): Modified bboxes. """ def _mod(bbox, target_size): top_ori, left_ori, h_ori, w_ori = bbox h, w = target_size assert h >= h_ori and w >= w_ori top = int(max(0, top_ori - (h - h_ori) // 2)) left = int(max(0, left_ori - (w - w_ori) // 2)) if isinstance(bbox, torch.Tensor): bbox_new = torch.Tensor([top, left, h, w]).type_as(bbox) else: bbox_new = np.asarray([top, left, h, w]) return bbox_new if isinstance(bbox, torch.Tensor): bbox_new = torch.zeros_like(bbox) elif isinstance(bbox, np.ndarray): bbox_new = np.zeros_like(bbox) else: raise TypeError('bbox mush be torch.Tensor or numpy.ndarray' f'but got type {type(bbox)}') bbox_shape = list(bbox.shape) if len(bbox_shape) == 2: for i in range(bbox_shape[0]): bbox_new[i, :] = _mod(bbox[i], target_size) else: bbox_new = _mod(bbox, target_size) return bbox_new def extract_around_bbox(img, bbox, target_size, channel_first=True): """Extract patches around the given bbox. Args: bbox (np.ndarray | torch.Tensor): Bboxes to be modified. Bbox can be in batch or not. target_size (List(int)): Target size of final bbox. Returns: (torch.Tensor | numpy.array): Extracted patches. The dimension of the \ output should be the same as `img`. """ bbox_new = scale_bbox(bbox, target_size) img_patch = extract_bbox_patch(bbox_new, img, channel_first=channel_first) return img_patch, bbox_new
32.742647
79
0.593083
794d57b67d86c0c46e16343780489dc76383b8c2
4,625
py
Python
help.py
Nugetzrul3/MBC-Tip
cb3046ab451fc4fab44c22bdb3c7f428c7344202
[ "MIT" ]
null
null
null
help.py
Nugetzrul3/MBC-Tip
cb3046ab451fc4fab44c22bdb3c7f428c7344202
[ "MIT" ]
null
null
null
help.py
Nugetzrul3/MBC-Tip
cb3046ab451fc4fab44c22bdb3c7f428c7344202
[ "MIT" ]
null
null
null
import discord from discord.ext import commands import user_db import config class Help(commands.Cog): def __init__(self, bot): self.bot = bot @commands.command() async def help(self, ctx): user_id = str(ctx.author.id) user_name = ctx.author.name if ctx.channel.id == 723493757860708714: if not user_db.check_user(user_id): user_db.add_user(user_id, user_name) embed = discord.Embed( title="**MBC Bot Command List:**", color=0x7152b6) embed.set_author( name=ctx.author.display_name, icon_url=ctx.author.avatar_url_as(format='png', size=256)) embed.add_field( name="**//help**", value="This is the command you have just used :wink:", inline=True) embed.add_field( name="**//info**", value="Show MBC Core wallet/blockchain info.", inline=True) embed.add_field( name="**//balance**", value="Show your MBC balances.", inline=True) embed.add_field( name="**//deposit**", value="Show your MBC deposit address.", inline=True) embed.add_field( name="**//tip**", value="Tip specified user. [//tip @Nugetzrul3 1] to tip 1 MBC", inline=True) embed.add_field( name="**//withdraw**", value="Withdraw MBC from your wallet. [//withdraw ADDRESS AMOUNT]", inline=True) embed.add_field( name="**//withdrawall**", value="Withdraw all MBC from your wallet. [//withdrawall ADDRESS]", inline=True) embed.set_thumbnail(url=self.bot.user.avatar_url_as(format='png', size=1024)) embed.set_footer(text="TipBot v{0}".format(config.VERSION), icon_url=self.bot.user.avatar_url_as(format='png', size=256)) await ctx.channel.send(embed=embed) else: pass embed = discord.Embed( title="**MBC Bot Command List:**", color=0x7152b6) embed.set_author( name=ctx.author.display_name, icon_url=ctx.author.avatar_url_as(format='png', size=256)) embed.add_field( name="**//help**", value="This is the command you have just used.", inline=True) embed.add_field( name="**//info**", value="Show MBC Core wallet/blockchain info.", inline=True) embed.add_field( name="**//balance**", value="Show your MBC balances.", inline=True) embed.add_field( name="**//deposit**", value="Show your MBC deposit address.", inline=True) embed.add_field( name="**//tip**", value="Tip specified user. [//tip @Nugetzrul3 1] to tip 1 MBC", inline=True) embed.add_field( name="**//withdraw**", value="Withdraw MBC from your wallet. [//withdraw ADDRESS AMOUNT]", inline=True) embed.add_field( name="**//withdrawall**", value="Withdraw all MBC from your wallet. [//withdrawall ADDRESS]", inline=True) embed.set_thumbnail(url=self.bot.user.avatar_url_as(format='png', size=1024)) embed.set_footer(text="TipBot v{0}".format(config.VERSION), icon_url=self.bot.user.avatar_url_as(format='png', size=256)) await ctx.channel.send(embed=embed) else: embed = discord.Embed(title="Oops", color=0x7152b6) embed.set_author(name=ctx.author.display_name, icon_url=ctx.author.avatar_url_as(format='png', size=256)) embed.add_field( name="Wrong Channel", value="Please use #tipbot to use this tipbot", ) await ctx.channel.send(embed=embed) def setup(bot): bot.remove_command('help') bot.add_cog(Help(bot))
40.929204
137
0.483676
794d57e03268412c5a8227fee9692c07eda7e151
308
py
Python
student.py
rowrodney/Automation
b6aa8bc802fa369a974868764c941c80f9edd2d2
[ "MIT" ]
null
null
null
student.py
rowrodney/Automation
b6aa8bc802fa369a974868764c941c80f9edd2d2
[ "MIT" ]
null
null
null
student.py
rowrodney/Automation
b6aa8bc802fa369a974868764c941c80f9edd2d2
[ "MIT" ]
null
null
null
from flask import Flask from flask_restful import Resource, Api app = Flask(__name__) api = Api(app) class Student(Resource): def get(self, name): return {'student': name} api.add_resource(Student, '/student/<string:name>') # http://127.0.0.1:5000/student/Rolf app.run(port=5000)
22
92
0.675325
794d57e790fa2d7486e63dd007e0109a9a672c3a
3,873
py
Python
maistra/vendor/com_googlesource_chromium_v8/wee8/build/android/gyp/create_java_binary_script.py
knm3000/proxy
f2bb57b7294aea2cb344824785be42849d7d63c9
[ "Apache-2.0" ]
3
2020-11-30T15:35:37.000Z
2022-01-06T14:17:18.000Z
maistra/vendor/com_googlesource_chromium_v8/wee8/build/android/gyp/create_java_binary_script.py
knm3000/proxy
f2bb57b7294aea2cb344824785be42849d7d63c9
[ "Apache-2.0" ]
54
2020-06-23T17:34:04.000Z
2022-03-31T02:04:06.000Z
maistra/vendor/com_googlesource_chromium_v8/wee8/build/android/gyp/create_java_binary_script.py
knm3000/proxy
f2bb57b7294aea2cb344824785be42849d7d63c9
[ "Apache-2.0" ]
12
2020-07-14T23:59:57.000Z
2022-03-22T09:59:18.000Z
#!/usr/bin/env python # # Copyright 2014 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Creates a simple script to run a java "binary". This creates a script that sets up the java command line for running a java jar. This includes correctly setting the classpath and the main class. """ import optparse import os import sys from util import build_utils # The java command must be executed in the current directory because there may # be user-supplied paths in the args. The script receives the classpath relative # to the directory that the script is written in and then, when run, must # recalculate the paths relative to the current directory. script_template = """\ #!/usr/bin/env python # # This file was generated by build/android/gyp/create_java_binary_script.py import argparse import os import sys self_dir = os.path.dirname(__file__) classpath = [{classpath}] extra_program_args = {extra_program_args} java_path = {java_path} if os.getcwd() != self_dir: offset = os.path.relpath(self_dir, os.getcwd()) fix_path = lambda p: os.path.normpath(os.path.join(offset, p)) classpath = [fix_path(p) for p in classpath] java_path = fix_path(java_path) java_cmd = [java_path] # This is a simple argparser for jvm, jar, and classpath arguments. parser = argparse.ArgumentParser() parser.add_argument('--jar-args') parser.add_argument('--jvm-args') parser.add_argument('--classpath') # Test_runner parses the classpath for sharding junit tests. parser.add_argument('--print-classpath', action='store_true', help='Prints the classpass. Used by test_runner.') known_args, unknown_args = parser.parse_known_args(sys.argv[1:]) if known_args.print_classpath: sys.stdout.write(':'.join(classpath)) sys.exit(0) if known_args.jvm_args: jvm_arguments = known_args.jvm_args.strip('"').split() java_cmd.extend(jvm_arguments) if known_args.jar_args: jar_arguments = known_args.jar_args.strip('"').split() if unknown_args: raise Exception('There are unknown arguments') else: jar_arguments = unknown_args if known_args.classpath: classpath += [known_args.classpath] {noverify_flag} java_cmd.extend( ['-classpath', ':'.join(classpath), '-enableassertions', \"{main_class}\"]) java_cmd.extend(extra_program_args) java_cmd.extend(jar_arguments) os.execvp(java_cmd[0], java_cmd) """ def main(argv): argv = build_utils.ExpandFileArgs(argv) parser = optparse.OptionParser() parser.add_option('--output', help='Output path for executable script.') parser.add_option('--main-class', help='Name of the java class with the "main" entry point.') parser.add_option('--classpath', action='append', default=[], help='Classpath for running the jar.') parser.add_option('--noverify', action='store_true', help='JVM flag: noverify.') options, extra_program_args = parser.parse_args(argv) if (options.noverify): noverify_flag = 'java_cmd.append("-noverify")' else: noverify_flag = '' classpath = [] for cp_arg in options.classpath: classpath += build_utils.ParseGnList(cp_arg) run_dir = os.path.dirname(options.output) classpath = [os.path.relpath(p, run_dir) for p in classpath] java_path = os.path.relpath( os.path.join(build_utils.JAVA_HOME, 'bin', 'java'), run_dir) with build_utils.AtomicOutput(options.output) as script: script.write( script_template.format(classpath=('"%s"' % '", "'.join(classpath)), java_path=repr(java_path), main_class=options.main_class, extra_program_args=repr(extra_program_args), noverify_flag=noverify_flag)) os.chmod(options.output, 0750) if __name__ == '__main__': sys.exit(main(sys.argv[1:]))
33.102564
80
0.714175
794d584ac49eaf4afd3f864c9da97c81ebc9cf60
1,973
py
Python
exercises/kindergarten-garden/kindergarten_garden_test.py
Alexhans/python
41287dda364e11edbaeda7cc50994d951007e835
[ "MIT" ]
2
2019-07-25T04:40:24.000Z
2020-12-18T21:29:02.000Z
exercises/kindergarten-garden/kindergarten_garden_test.py
Alexhans/python
41287dda364e11edbaeda7cc50994d951007e835
[ "MIT" ]
null
null
null
exercises/kindergarten-garden/kindergarten_garden_test.py
Alexhans/python
41287dda364e11edbaeda7cc50994d951007e835
[ "MIT" ]
null
null
null
import unittest from kindergarten_garden import Garden # Tests adapted from `problem-specifications//canonical-data.json` @ v1.1.0 class KindergartenGardenTest(unittest.TestCase): def test_garden_with_single_student(self): self.assertEqual( Garden("RC\nGG").plants("Alice"), "Radishes Clover Grass Grass".split()) def test_different_garden_with_single_student(self): self.assertEqual( Garden("VC\nRC").plants("Alice"), "Violets Clover Radishes Clover".split()) def test_garden_with_two_students(self): garden = Garden("VVCG\nVVRC") self.assertEqual( garden.plants("Bob"), "Clover Grass Radishes Clover".split()) def test_multiple_students_for_the_same_garden_with_three_students(self): garden = Garden("VVCCGG\nVVCCGG") self.assertEqual(garden.plants("Bob"), ["Clover"] * 4) self.assertEqual(garden.plants("Charlie"), ["Grass"] * 4) def test_full_garden(self): garden = Garden("VRCGVVRVCGGCCGVRGCVCGCGV\nVRCCCGCRRGVCGCRVVCVGCGCV") self.assertEqual( garden.plants("Alice"), "Violets Radishes Violets Radishes".split()) self.assertEqual( garden.plants("Bob"), "Clover Grass Clover Clover".split()) self.assertEqual( garden.plants("Kincaid"), "Grass Clover Clover Grass".split()) self.assertEqual( garden.plants("Larry"), "Grass Violets Clover Violets".split()) # Additional tests for this track def test_disordered_test(self): garden = Garden( "VCRRGVRG\nRVGCCGCV", students="Samantha Patricia Xander Roger".split()) self.assertEqual( garden.plants("Patricia"), "Violets Clover Radishes Violets".split()) self.assertEqual( garden.plants("Xander"), "Radishes Grass Clover Violets".split()) if __name__ == '__main__': unittest.main()
35.872727
77
0.646731
794d592077571ae0f63c9044f18bd94731dd3a01
607
py
Python
setup.py
Chinacolt/pip-cintay
72187be8b393a60fb3316d4ad40277eeb2d05d47
[ "MIT" ]
1
2019-07-10T22:24:17.000Z
2019-07-10T22:24:17.000Z
setup.py
Chinacolt/pip-cintay
72187be8b393a60fb3316d4ad40277eeb2d05d47
[ "MIT" ]
1
2021-11-15T17:48:55.000Z
2021-11-15T17:48:55.000Z
setup.py
Chinacolt/pip-cintay
72187be8b393a60fb3316d4ad40277eeb2d05d47
[ "MIT" ]
null
null
null
import setuptools with open("README.md", "r") as fh: long_description = fh.read() setuptools.setup( name='cintay', version='0.1', author="Emre Cintay", author_email="emre@cintay.com", description="Print CINTAY", long_description=long_description, long_description_content_type="text/markdown", url="https://github.com/Chinacolt/cintay-pip", packages=setuptools.find_packages(), classifiers=[ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", ], )
25.291667
51
0.639209
794d592e701d667e9e90c72d76a085f838490335
305
py
Python
2016/09/north-korea-coal-20160914/graphic_config.py
nprapps/graphics-archive
97b0ef326b46a959df930f5522d325e537f7a655
[ "FSFAP" ]
14
2015-05-08T13:41:51.000Z
2021-02-24T12:34:55.000Z
2016/09/north-korea-coal-20160914/graphic_config.py
nprapps/graphics-archive
97b0ef326b46a959df930f5522d325e537f7a655
[ "FSFAP" ]
null
null
null
2016/09/north-korea-coal-20160914/graphic_config.py
nprapps/graphics-archive
97b0ef326b46a959df930f5522d325e537f7a655
[ "FSFAP" ]
7
2015-04-04T04:45:54.000Z
2021-02-18T11:12:48.000Z
#!/usr/bin/env python import base_filters COPY_GOOGLE_DOC_KEY = '1vnsg2zLO--g6ppYdGZsyH0tb9vrP5VIjQS24xeormpA' USE_ASSETS = False # Use these variables to override the default cache timeouts for this graphic # DEFAULT_MAX_AGE = 20 # ASSETS_MAX_AGE = 300 JINJA_FILTER_FUNCTIONS = base_filters.FILTERS
21.785714
77
0.813115
794d5a5551b647332cd8b5ae289763e8971950ae
3,037
py
Python
sdk/python/pulumi_azure_native/logic/v20160601/list_workflow_run_action_repetition_expression_traces.py
pulumi-bot/pulumi-azure-native
f7b9490b5211544318e455e5cceafe47b628e12c
[ "Apache-2.0" ]
null
null
null
sdk/python/pulumi_azure_native/logic/v20160601/list_workflow_run_action_repetition_expression_traces.py
pulumi-bot/pulumi-azure-native
f7b9490b5211544318e455e5cceafe47b628e12c
[ "Apache-2.0" ]
null
null
null
sdk/python/pulumi_azure_native/logic/v20160601/list_workflow_run_action_repetition_expression_traces.py
pulumi-bot/pulumi-azure-native
f7b9490b5211544318e455e5cceafe47b628e12c
[ "Apache-2.0" ]
null
null
null
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi SDK Generator. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union from ... import _utilities, _tables from . import outputs __all__ = [ 'ListWorkflowRunActionRepetitionExpressionTracesResult', 'AwaitableListWorkflowRunActionRepetitionExpressionTracesResult', 'list_workflow_run_action_repetition_expression_traces', ] @pulumi.output_type class ListWorkflowRunActionRepetitionExpressionTracesResult: def __init__(__self__, inputs=None): if inputs and not isinstance(inputs, list): raise TypeError("Expected argument 'inputs' to be a list") pulumi.set(__self__, "inputs", inputs) @property @pulumi.getter def inputs(self) -> Optional[Sequence['outputs.ExpressionRootResponseResult']]: return pulumi.get(self, "inputs") class AwaitableListWorkflowRunActionRepetitionExpressionTracesResult(ListWorkflowRunActionRepetitionExpressionTracesResult): # pylint: disable=using-constant-test def __await__(self): if False: yield self return ListWorkflowRunActionRepetitionExpressionTracesResult( inputs=self.inputs) def list_workflow_run_action_repetition_expression_traces(action_name: Optional[str] = None, repetition_name: Optional[str] = None, resource_group_name: Optional[str] = None, run_name: Optional[str] = None, workflow_name: Optional[str] = None, opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableListWorkflowRunActionRepetitionExpressionTracesResult: """ Use this data source to access information about an existing resource. :param str action_name: The workflow action name. :param str repetition_name: The workflow repetition. :param str resource_group_name: The resource group name. :param str run_name: The workflow run name. :param str workflow_name: The workflow name. """ __args__ = dict() __args__['actionName'] = action_name __args__['repetitionName'] = repetition_name __args__['resourceGroupName'] = resource_group_name __args__['runName'] = run_name __args__['workflowName'] = workflow_name if opts is None: opts = pulumi.InvokeOptions() if opts.version is None: opts.version = _utilities.get_version() __ret__ = pulumi.runtime.invoke('azure-native:logic/v20160601:listWorkflowRunActionRepetitionExpressionTraces', __args__, opts=opts, typ=ListWorkflowRunActionRepetitionExpressionTracesResult).value return AwaitableListWorkflowRunActionRepetitionExpressionTracesResult( inputs=__ret__.inputs)
44.014493
201
0.686203
794d5a95e038d92c7da9ab527394617f809ccb35
705
py
Python
src/tp_user_1.0/sawtooth_user/exceptions.py
jlee7x2/SParts
fd39e491d0433a7429c5fdb037d5826d7f34b291
[ "Apache-2.0" ]
12
2018-08-29T23:19:48.000Z
2021-11-18T19:09:21.000Z
src/tp_user_1.0/sawtooth_user/exceptions.py
jlee7x2/SParts
fd39e491d0433a7429c5fdb037d5826d7f34b291
[ "Apache-2.0" ]
18
2018-07-13T22:37:44.000Z
2020-02-20T19:22:14.000Z
src/tp_user_1.0/sawtooth_user/exceptions.py
jlee7x2/SParts
fd39e491d0433a7429c5fdb037d5826d7f34b291
[ "Apache-2.0" ]
16
2018-03-12T21:52:02.000Z
2021-07-13T20:46:10.000Z
# Copyright 2016 Intel Corporation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ------------------------------------------------------------------------------ class UserException(Exception): pass
37.105263
80
0.675177
794d5bda2c79ab00330e7114c6a37ef9b5885c4e
265
py
Python
tests/artificial/transf_Anscombe/trend_LinearTrend/cycle_30/ar_12/test_artificial_32_Anscombe_LinearTrend_30_12_0.py
shaido987/pyaf
b9afd089557bed6b90b246d3712c481ae26a1957
[ "BSD-3-Clause" ]
377
2016-10-13T20:52:44.000Z
2022-03-29T18:04:14.000Z
tests/artificial/transf_Anscombe/trend_LinearTrend/cycle_30/ar_12/test_artificial_32_Anscombe_LinearTrend_30_12_0.py
ysdede/pyaf
b5541b8249d5a1cfdc01f27fdfd99b6580ed680b
[ "BSD-3-Clause" ]
160
2016-10-13T16:11:53.000Z
2022-03-28T04:21:34.000Z
tests/artificial/transf_Anscombe/trend_LinearTrend/cycle_30/ar_12/test_artificial_32_Anscombe_LinearTrend_30_12_0.py
ysdede/pyaf
b5541b8249d5a1cfdc01f27fdfd99b6580ed680b
[ "BSD-3-Clause" ]
63
2017-03-09T14:51:18.000Z
2022-03-27T20:52:57.000Z
import pyaf.Bench.TS_datasets as tsds import tests.artificial.process_artificial_dataset as art art.process_dataset(N = 32 , FREQ = 'D', seed = 0, trendtype = "LinearTrend", cycle_length = 30, transform = "Anscombe", sigma = 0.0, exog_count = 0, ar_order = 12);
37.857143
165
0.732075
794d5c0b4f0b2473663fe6db3537edf561e9352a
1,905
py
Python
DM900-USB-ArduinoNano/DM900-Enigmalight-Service/home/elight-addons/wifilight/philips_hue/enigmalight_hue_LR.py
NeedfulThings/hyperionconfigs
1428d62b0d0d59bd062309870481d1b360e625e1
[ "MIT" ]
null
null
null
DM900-USB-ArduinoNano/DM900-Enigmalight-Service/home/elight-addons/wifilight/philips_hue/enigmalight_hue_LR.py
NeedfulThings/hyperionconfigs
1428d62b0d0d59bd062309870481d1b360e625e1
[ "MIT" ]
null
null
null
DM900-USB-ArduinoNano/DM900-Enigmalight-Service/home/elight-addons/wifilight/philips_hue/enigmalight_hue_LR.py
NeedfulThings/hyperionconfigs
1428d62b0d0d59bd062309870481d1b360e625e1
[ "MIT" ]
2
2019-10-12T17:14:13.000Z
2019-10-20T14:59:41.000Z
import sys import os import time import json import httplib from rgb_xy import Converter from rgb_xy import GamutC # or GamutA, GamutB (you must look for the type of your lamps in rgb_xy.py from line 42) counter = 12 def popen(): converter = Converter(GamutC) spidev = file( os.getcwd()+'/aufruf.log', "wb") key = "HIER DEN KEY DER BRIDGE EINTRAGEN" ip = "xxx.xxx.xxx.xxx" url = '/api/' + key + '/lights/' lurl = url + '10/state' rurl = url + '11/state' MINIMAL_VALUE=0.000000000 while True: eingabe = sys.stdin.readline() if len(eingabe)>0: global counter counter += 1 try: lr,lg,lb,rr,rg,rb,x = eingabe.split(' ') except ValueError: spidev.write("Not enough input parameter, do you have the same amount of lights (channels) in your enigmalight config?") spidev.flush() raise lr = (float(lr))*255 lg = (float(lg))*255 lb = (float(lb))*255 rr = (float(rr))*255 rg = (float(rg))*255 rb = (float(rb))*255 lll = calcLuminance(lr,lg,lb) llr = calcLuminance(rr,rg,rb) if (counter>=13): connection = httplib.HTTPConnection(ip, timeout=10) lparams = {'xy': converter.rgb_to_xy(lr,lg,lb), 'colormode': 'xy', 'bri': int(lll), 'on': True} connection.request('PUT', lurl, json.dumps(lparams)) response = connection.getresponse() rparams = {'xy': converter.rgb_to_xy(rr,rg,rb), 'colormode': 'xy', 'bri': int(llr), 'on': True} connection.request('PUT', rurl, json.dumps(rparams)) response = connection.getresponse() connection.close() counter=0 else: os.system("curl -d '{\"on\":false}' -X PUT 192.168.xxx.xxx/api/HIER DEN KEY DER BRIDGE EINTRAGEN/groups/0/action") break def calcLuminance(r,g,b): LUM_VALUE=20 luminance=1 if (r + g + b > 2): luminance= r + g + b + LUM_VALUE if (luminance>=255): luminance=254 return luminance import time time.sleep(1) popen()
24.74026
124
0.648294
794d5d1fa213dbac57eeb85a91dc785f95783a14
704
py
Python
ex21.py
KT12/lpthw
1a97c33cc65db9b2bad9222e6c6739d8e6bf5f03
[ "MIT" ]
null
null
null
ex21.py
KT12/lpthw
1a97c33cc65db9b2bad9222e6c6739d8e6bf5f03
[ "MIT" ]
null
null
null
ex21.py
KT12/lpthw
1a97c33cc65db9b2bad9222e6c6739d8e6bf5f03
[ "MIT" ]
null
null
null
def add(a, b): print('ADDING %d + %d' % (a, b)) return a + b def subtract(a, b): print('SUBTRACTNG %d - %d' % (a, b)) return a - b def multiply(a, b): print('MULTIPLYING %d * %d' % (a, b)) return a * b def divide(a, b): print('DIVIDING %d / %d' % (a,b)) return a / b print("Let's do some math with just functions!") age = add(30, 5) height = subtract(78, 4) weight = multiply(90, 2) iq = divide(100, 2) print('Age: %d, Height: %d, Weight: %d, IQ: %d' % (age, height, weight, iq)) print('Here is a puzzle.') what = add(age, subtract(height, multiply(weight, divide(iq, 2)))) print('That becomes: ', what, '. Can you do it by hand?')
22.709677
77
0.544034
794d5e6b791f8e3b5941522bd826abab9772f5b7
395
py
Python
django/django_fundamentals/django_intro/ninjagold/ninjagold/wsgi.py
fatimaalheeh/python_stack
9ba84e6dc030a65494f105152a97f0a38aa2e4f3
[ "MIT" ]
null
null
null
django/django_fundamentals/django_intro/ninjagold/ninjagold/wsgi.py
fatimaalheeh/python_stack
9ba84e6dc030a65494f105152a97f0a38aa2e4f3
[ "MIT" ]
null
null
null
django/django_fundamentals/django_intro/ninjagold/ninjagold/wsgi.py
fatimaalheeh/python_stack
9ba84e6dc030a65494f105152a97f0a38aa2e4f3
[ "MIT" ]
null
null
null
""" WSGI config for ninjagold project. It exposes the WSGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/2.2/howto/deployment/wsgi/ """ import os from django.core.wsgi import get_wsgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'ninjagold.settings') application = get_wsgi_application()
23.235294
78
0.787342
794d5e77f7f2d8a2e2cbe73ec7adb1a8e46eaa1a
4,657
py
Python
database_sanitizer/utils/mysql.py
sharescape/python-database-sanitizer
560bf402e3896285980abb21a74d5be8d2da1698
[ "MIT" ]
37
2018-05-07T13:07:25.000Z
2022-02-07T18:58:10.000Z
database_sanitizer/utils/mysql.py
sharescape/python-database-sanitizer
560bf402e3896285980abb21a74d5be8d2da1698
[ "MIT" ]
31
2018-04-27T13:16:28.000Z
2021-12-10T10:08:00.000Z
database_sanitizer/utils/mysql.py
sharescape/python-database-sanitizer
560bf402e3896285980abb21a74d5be8d2da1698
[ "MIT" ]
15
2018-05-04T12:28:12.000Z
2022-02-17T09:27:58.000Z
# -*- coding: utf-8 -*- from __future__ import unicode_literals import re import pymysql import six def get_mysqldump_args_and_env_from_url(url): """ Constructs list of command line arguments and dictionary of environment variables that can be given to `mysqldump` executable to obtain database dump of the database described in given URL. :param url: Parsed database URL. :type url: urllib.urlparse.ParseResult :return: List of command line arguments as well as dictionary of environment variables that can be used to launch the MySQL dump process to obtain dump of the database. :rtype: tuple[list[str],dict[str,str]] """ args = [ # Without this, `INSERT INTO` statements will exclude column names from # the output, which are required for sanitation. "--complete-insert", # This enables use for "exteded inserts" where multiple rows of a table # are included in a single `INSERT INTO` statement (contents of the # entire table even, if it's within limits). We use it to increase the # performance of the sanitation and to decrease the dump size. "--extended-insert", # This makes the `mysqldump` to attempt to limit size of a single line # into 10 megabytes. We use it to reduce memory consumption. "--net_buffer_length=10240", # Hostname of the database to connect into, should be always present in # the parsed database URL. "-h", url.hostname, ] env = {} if url.port is not None: args.extend(("-P", six.text_type(url.port))) if url.username: args.extend(("-u", url.username)) if url.password: env["MYSQL_PWD"] = url.password if len(url.path) < 2 or not url.path.startswith("/"): raise ValueError("Name of the database is missing from the URL") args.append(url.path[1:]) return args, env MYSQL_NULL_PATTERN = re.compile(r"^NULL$", re.IGNORECASE) MYSQL_BOOLEAN_PATTERN = re.compile(r"^(TRUE|FALSE)$", re.IGNORECASE) MYSQL_FLOAT_PATTERN = re.compile(r"^[+-]?\d*\.\d+([eE][+-]?\d+)?$") MYSQL_INT_PATTERN = re.compile(r"^\d+$") MYSQL_STRING_PATTERN = re.compile(r"'(?:[^']|''|\\')*(?<![\\])'") def decode_mysql_literal(text): """ Attempts to decode given MySQL literal into Python value. :param text: Value to be decoded, as MySQL literal. :type text: str :return: Python version of the given MySQL literal. :rtype: any """ if MYSQL_NULL_PATTERN.match(text): return None if MYSQL_BOOLEAN_PATTERN.match(text): return text.lower() == "true" if MYSQL_FLOAT_PATTERN.match(text): return float(text) if MYSQL_INT_PATTERN.match(text): return int(text) if MYSQL_STRING_PATTERN.match(text): return decode_mysql_string_literal(text) raise ValueError("Unable to decode given value: %r" % (text,)) MYSQL_STRING_ESCAPE_SEQUENCE_PATTERN = re.compile(r"\\(.)") MYSQL_STRING_ESCAPE_SEQUENCE_MAPPING = { "\\0": "\000", "\\b": "\b", "\\n": "\n", "\\r": "\r", "\\t": "\t", "\\Z": "\032", } def decode_mysql_string_literal(text): """ Removes quotes and decodes escape sequences from given MySQL string literal returning the result. :param text: MySQL string literal, with the quotes still included. :type text: str :return: Given string literal with quotes removed and escape sequences decoded. :rtype: str """ assert text.startswith("'") assert text.endswith("'") # Ditch quotes from the string literal. text = text[1:-1] return MYSQL_STRING_ESCAPE_SEQUENCE_PATTERN.sub( unescape_single_character, text, ) def unescape_single_character(match): """ Unescape a single escape sequence found from a MySQL string literal, according to the rules defined at: https://dev.mysql.com/doc/refman/5.6/en/string-literals.html#character-escape-sequences :param match: Regular expression match object. :return: Unescaped version of given escape sequence. :rtype: str """ value = match.group(0) assert value.startswith("\\") return MYSQL_STRING_ESCAPE_SEQUENCE_MAPPING.get(value) or value[1:] def encode_mysql_literal(value): """ Converts given Python value into MySQL literal, suitable to be used inside `INSERT INTO` statement. :param value: Value to convert into MySQL literal. :type value: any :return: Given value encoded into MySQL literal. :rtype: str """ return pymysql.converters.escape_item(value, "utf-8")
28.746914
91
0.657934
794d5f322a62128147eea66034bb79576951624c
1,721
py
Python
jina/clients/python/request.py
phamtrancsek12/jina
efa339fd7e19760d9033b51e54a1b5eca60f1c5f
[ "Apache-2.0" ]
2
2021-06-18T11:55:15.000Z
2021-08-30T20:15:46.000Z
jina/clients/python/request.py
phamtrancsek12/jina
efa339fd7e19760d9033b51e54a1b5eca60f1c5f
[ "Apache-2.0" ]
null
null
null
jina/clients/python/request.py
phamtrancsek12/jina
efa339fd7e19760d9033b51e54a1b5eca60f1c5f
[ "Apache-2.0" ]
1
2020-08-28T09:08:29.000Z
2020-08-28T09:08:29.000Z
__copyright__ = "Copyright (c) 2020 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import ctypes import random from typing import Iterator, Union from ...helper import batch_iterator from ...proto import jina_pb2 def _generate(data: Union[Iterator[bytes], Iterator['jina_pb2.Document']], batch_size: int = 0, first_doc_id: int = 0, first_request_id: int = 0, random_doc_id: bool = False, mode: str = 'index', top_k: int = 50, in_proto: bool = False, *args, **kwargs) -> Iterator['jina_pb2.Message']: for pi in batch_iterator(data, batch_size): req = jina_pb2.Request() req.request_id = first_request_id if mode == 'search': if top_k <= 0: raise ValueError('"top_k: %d" is not a valid number' % top_k) else: req.search.top_k = top_k for raw_bytes in pi: d = getattr(req, mode).docs.add() if in_proto: d.CopyFrom(raw_bytes) else: d.raw_bytes = raw_bytes d.doc_id = first_doc_id if not random_doc_id else random.randint(0, ctypes.c_uint(-1).value) d.weight = 1.0 first_doc_id += 1 yield req first_request_id += 1 def index(*args, **kwargs): """Generate indexing request""" yield from _generate(*args, **kwargs) def train(*args, **kwargs): """Generate training request """ yield from _generate(*args, **kwargs) req = jina_pb2.Request() req.request_id = 1 req.train.flush = True yield req def search(*args, **kwargs): """Generate search request """ yield from _generate(*args, **kwargs)
30.192982
104
0.595003
794d5f5ef8759d60c397d0983bc0c0a02bf3196b
136
py
Python
counterfactualms/experiments/classifier/__init__.py
jcreinhold/counterfactualms
9be5919c8885354fe1ac91c852d196969cfe16be
[ "Apache-2.0" ]
14
2021-03-08T11:51:34.000Z
2022-03-21T03:30:21.000Z
counterfactualms/experiments/classifier/__init__.py
jcreinhold/counterfactualms
9be5919c8885354fe1ac91c852d196969cfe16be
[ "Apache-2.0" ]
1
2021-10-03T15:20:30.000Z
2021-10-03T15:20:30.000Z
counterfactualms/experiments/classifier/__init__.py
jcreinhold/counterfactualms
9be5919c8885354fe1ac91c852d196969cfe16be
[ "Apache-2.0" ]
4
2021-04-03T15:23:24.000Z
2021-09-30T04:14:28.000Z
from counterfactualms.experiments.classifier.classifier_experiment import ClassifierExperiment __all__ = [ 'ClassifierExperiment' ]
27.2
94
0.845588
794d5fa1cba4ac5812edf456bcabc465f77998bf
16,122
py
Python
cellfinder/extract/extract_cubes.py
brainglobe/cellfinder
0466e94f5c5ad58c853a73257d80944a1607ac81
[ "BSD-3-Clause" ]
53
2020-10-22T10:01:12.000Z
2022-03-14T04:03:43.000Z
cellfinder/extract/extract_cubes.py
brainglobe/cellfinder
0466e94f5c5ad58c853a73257d80944a1607ac81
[ "BSD-3-Clause" ]
59
2020-10-08T16:23:46.000Z
2022-03-14T09:47:53.000Z
cellfinder/extract/extract_cubes.py
brainglobe/cellfinder
0466e94f5c5ad58c853a73257d80944a1607ac81
[ "BSD-3-Clause" ]
9
2020-11-20T18:48:08.000Z
2022-03-06T13:03:22.000Z
""" Cube extraction for CNN-based classification. Based on, and mostly copied from, https://github.com/SainsburyWellcomeCentre/cell_count_analysis by Charly Rousseau (https://github.com/crousseau). """ import os from collections import deque from concurrent.futures import ProcessPoolExecutor import logging from datetime import datetime import numpy as np from skimage import transform from numpy.linalg.linalg import LinAlgError from math import floor from tifffile import tifffile from tqdm import tqdm from imlib.general.system import get_num_processes from imlib.cells.cells import group_cells_by_z from imlib.general.numerical import is_even from cellfinder.tools import image_processing as img_tools from cellfinder.tools import system class StackSizeError(Exception): pass class Cube(object): class CubeBoundaryError(Exception): pass def __init__( self, cell, channel, stacks, x_pix_um=1, y_pix_um=1, x_pix_um_network=1, y_pix_um_network=1, final_depth=20, width=50, height=50, depth=20, interpolation_order=0, ): self.cell = cell self.channel = channel self.stack = stacks[self.channel] self.scale_cubes = False self.scale_cubes_in_z = False self.interpolation_order = interpolation_order self.rescaling_factor_x = 1 self.rescaling_factor_y = 1 if x_pix_um != x_pix_um_network: self.rescaling_factor_x = float(x_pix_um_network) / float(x_pix_um) self.scale_cubes = True if y_pix_um != y_pix_um_network: self.rescaling_factor_y = float(y_pix_um_network) / float(y_pix_um) self.scale_cubes = True # This should be decided earlier if depth != final_depth: self.rescaling_factor_z = final_depth / depth self.scale_cubes_in_z = True self.width = width self.height = height self.depth = depth self.final_depth = depth self.empty = False self.data = self.extract_cube(self.stack) def __str__(self): # TODO: 0 pad (when changing classification) prefix = "pCell" return "{}z{}y{}x{}Ch{}.tif".format( prefix, self.z, self.y, self.x, self.channel ) @property def x(self): return self.cell.x @property def y(self): return self.cell.y @property def z(self): return self.cell.z @property def stack_x_max(self): return self.stack.shape[1] @property def stack_y_max(self): return self.stack.shape[0] @property def stack_z_max(self): return self.stack.shape[2] def extract_cube(self, stack): """ :param np.array stack: :return: """ try: self.check_boundaries() except Cube.CubeBoundaryError: return self.make_blank_data() half_width, half_height, x0, x1, y0, y1 = self.get_boundaries() data = stack[y0:y1, x0:x1, :] if data.ndim < 3: logging.error( "Error occurred during extraction of cube {}, " "original shape: {}, dtype: {}".format( self, data.shape, data.dtype ) ) return self.make_blank_data() out = [] # TODO: combine these scalings for i in range(self.depth): if self.scale_cubes: try: # To match classification 1um rescaling inv_rescale_x = 1 / self.rescaling_factor_x inv_rescale_y = 1 / self.rescaling_factor_y plane = transform.rescale( data[:, :, i], scale=(inv_rescale_y, inv_rescale_x), order=self.interpolation_order, preserve_range=True, multichannel=False, mode="constant", anti_aliasing=False, ) except LinAlgError as err: logging.error( f"Error occurred during rescale of cube {self}, " f"original shape: {data[:, :, i].shape}, dtype: " f"{data[:, :, i].dtype}; {err}." ) return self.make_blank_data() else: plane = data[:, :, i] plane = self.adjust_size(plane) out.append(plane) out = np.array(out, dtype=np.uint16) if self.scale_cubes_in_z: try: # To match classification cube scaling in z # TODO: presumably this will have rounding errors out = transform.rescale( out, scale=(self.rescaling_factor_z, 1, 1), order=self.interpolation_order, preserve_range=True, multichannel=False, mode="constant", anti_aliasing=False, ) out = out.astype(np.uint16) except LinAlgError as err: logging.error( f"Error occurred during z-rescale of cube {self}, " f"from depth: {self.depth} to final depth " f"{self.final_depth} ; {err}" ) print.error( f"Error occurred during z-rescale of cube {self}, " f"from depth: {self.depth} to final depth " f"{self.final_depth} ; {err}" ) return self.make_blank_data() return out def adjust_size(self, plane): plane = self.crop_to_correct_size(plane) plane = self.pad_to_correct_size(plane) return plane def pad_to_correct_size(self, plane): return img_tools.pad_center_2d( plane, x_size=self.width, y_size=self.height ) def crop_to_correct_size(self, plane): return img_tools.crop_center_2d( plane, crop_x=self.width, crop_y=self.height ) def get_boundaries(self): if self.scale_cubes: half_width = int(round(self.width / 2 * self.rescaling_factor_x)) half_height = int(round(self.height / 2 * self.rescaling_factor_y)) else: half_width = int(round(self.width / 2)) half_height = int(round(self.height / 2)) x0 = int(self.x - half_width) x1 = int(self.x + half_width) y0 = int(self.y - half_height) y1 = int(self.y + half_height) return half_width, half_height, x0, x1, y0, y1 def make_blank_data(self): self.empty = True return np.zeros((self.depth, self.width, self.height), dtype=np.uint16) def check_boundaries(self): out_bounds_message = ( "Cell candidate (x:{}, y:{}, z:{}) too close to the edge of the" " image for cube extraction." ).format(self.x, self.y, self.z) if self.stack_z_max < self.depth: error_msg = "Stack has {} planes but cube needs {}".format( self.stack_z_max, self.depth ) logging.error(error_msg) raise Cube.CubeBoundaryError(error_msg) boundaries = self.get_boundaries() half_height, half_width, x0, x1, y0, y1 = boundaries if x0 < 0 or y0 < 0: logging.warning(out_bounds_message) raise Cube.CubeBoundaryError(out_bounds_message) # WARNING: dimensions inverted (numpy y=dim0, x=dim1) elif x1 > self.stack_x_max or y1 > self.stack_y_max: logging.warning(out_bounds_message) raise Cube.CubeBoundaryError(out_bounds_message) def save_cubes( cells, planes_paths, planes_to_read, planes_shape, voxel_sizes, network_voxel_sizes, num_planes_for_cube=20, cube_width=50, cube_height=50, cube_depth=20, thread_id=0, output_dir="", save_empty_cubes=False, ): """ :param cells: :param planes_paths: :param planes_to_read: :param planes_shape: :param x_pix_um: :param y_pix_um: :param x_pix_um_network: :param y_pix_um_network: :param num_planes_for_cube: :param cube_width: :param cube_height: :param cube_depth: :param thread_id: :param output_dir: :param save_empty_cubes: :return: """ channels = list(planes_paths.keys()) stack_shape = planes_shape + (num_planes_for_cube,) stacks = {} planes_queues = {} for ch in channels: stacks[ch] = np.zeros(stack_shape, dtype=np.uint16) planes_queues[ch] = deque(maxlen=num_planes_for_cube) for plane_idx in tqdm(planes_to_read, desc="Thread: {}".format(thread_id)): for ch in channels: plane_path = planes_paths[ch][plane_idx] planes_queues[ch].append(tifffile.imread(plane_path)) if len(planes_queues[ch]) == num_planes_for_cube: if is_even(num_planes_for_cube): cell_z = int(plane_idx - num_planes_for_cube / 2 + 1) else: cell_z = int( plane_idx - floor(num_planes_for_cube) / 2 + 1 ) for j, plane in enumerate(planes_queues[ch]): stacks[ch][:, :, j] = plane # ensures no cube_depth planes at the end planes_queues[ch].popleft() # required since we provide all cells # TODO: if len(planes_queues[ch]) # < num_planes_for_cube -1: break for cell in cells[cell_z]: cube = Cube( cell, ch, stacks, x_pix_um=voxel_sizes[2], y_pix_um=voxel_sizes[1], x_pix_um_network=network_voxel_sizes[2], y_pix_um_network=network_voxel_sizes[1], final_depth=cube_depth, width=cube_width, height=cube_height, depth=num_planes_for_cube, ) if not cube.empty or (cube.empty and save_empty_cubes): tifffile.imsave( os.path.join(output_dir, str(cube)), cube.data ) def get_ram_requirement_per_process( single_image_path, cube_depth, num_channels=2, copies=2 ): """ Calculates how much RAM is needed per CPU core for cube extraction. Used later to set the number of CPU cores to be used. RAM requirement is currently: image_size x n_cores x n_channels x cube_depth x copies :param single_image_path: A single 2D image, used to find the size. :param num_channels: Number of channels to be extracted :param cube_depth: Depth of the cube to be extracted (z-planes) to use. :param int copies: How many copies of the data are loaded at any time :return int: memory requirement for one CPU core """ logging.debug( "Determining how much RAM is needed for each CPU for cube extraction" ) file_size = os.path.getsize(single_image_path) total_mem_need = file_size * num_channels * cube_depth * copies logging.debug( "File size: {:.2f} GB, " "number of channels: {}, " "cube depth: {}, " "Therefore theoretically {:.2f} GB is needed per CPU core." "".format( (file_size / (1024 ** 3)), num_channels, cube_depth, (total_mem_need / (1024 ** 3)), ) ) return total_mem_need def main( cells, cubes_output_dir, planes_paths, cube_depth, cube_width, cube_height, voxel_sizes, network_voxel_sizes, max_ram, n_free_cpus=4, save_empty_cubes=False, ): start_time = datetime.now() if voxel_sizes[0] != network_voxel_sizes[0]: plane_scaling_factor = float(network_voxel_sizes[0]) / float( voxel_sizes[0] ) num_planes_needed_for_cube = round(cube_depth * plane_scaling_factor) else: num_planes_needed_for_cube = cube_depth if num_planes_needed_for_cube > len(planes_paths[0]): raise StackSizeError( "The number of planes provided is not sufficient " "for any cubes to be extracted. Please check the " "input data" ) first_plane = tifffile.imread(list(planes_paths.values())[0][0]) planes_shape = first_plane.shape brain_depth = len(list(planes_paths.values())[0]) # TODO: use to assert all centre planes processed center_planes = sorted(list(set([cell.z for cell in cells]))) # REFACTOR: rename (clashes with different meaning of planes_to_read below) planes_to_read = np.zeros(brain_depth, dtype=np.bool) if is_even(num_planes_needed_for_cube): half_nz = num_planes_needed_for_cube // 2 # WARNING: not centered because even for p in center_planes: planes_to_read[p - half_nz : p + half_nz] = 1 else: half_nz = num_planes_needed_for_cube // 2 # centered for p in center_planes: planes_to_read[p - half_nz : p + half_nz + 1] = 1 planes_to_read = np.where(planes_to_read)[0] if not planes_to_read.size: logging.error( f"No planes found, you need at the very least " f"{num_planes_needed_for_cube} " f"planes to proceed (i.e. cube z size)" f"Brain z dimension is {brain_depth}.", stack_info=True, ) raise ValueError( f"No planes found, you need at the very least " f"{num_planes_needed_for_cube} " f"planes to proceed (i.e. cube z size)" f"Brain z dimension is {brain_depth}." ) # TODO: check if needs to flip args.cube_width and args.cube_height cells_groups = group_cells_by_z(cells) # copies=2 is set because at all times there is a plane queue (deque) # and an array passed to `Cube` ram_per_process = get_ram_requirement_per_process( planes_paths[0][0], num_planes_needed_for_cube, copies=2, ) n_processes = get_num_processes( min_free_cpu_cores=n_free_cpus, ram_needed_per_process=ram_per_process, n_max_processes=len(planes_to_read), fraction_free_ram=0.2, max_ram_usage=system.memory_in_bytes(max_ram, "GB"), ) # TODO: don't need to extract cubes from all channels if # n_signal_channels>1 with ProcessPoolExecutor(max_workers=n_processes) as executor: n_planes_per_chunk = len(planes_to_read) // n_processes for i in range(n_processes): start_idx = i * n_planes_per_chunk end_idx = ( start_idx + n_planes_per_chunk + num_planes_needed_for_cube - 1 ) if end_idx > planes_to_read[-1]: end_idx = None sub_planes_to_read = planes_to_read[start_idx:end_idx] executor.submit( save_cubes, cells_groups, planes_paths, sub_planes_to_read, planes_shape, voxel_sizes, network_voxel_sizes, num_planes_for_cube=num_planes_needed_for_cube, cube_width=cube_width, cube_height=cube_height, cube_depth=cube_depth, thread_id=i, output_dir=cubes_output_dir, save_empty_cubes=save_empty_cubes, ) total_cubes = system.get_number_of_files_in_dir(cubes_output_dir) time_taken = datetime.now() - start_time logging.info( "All cubes ({}) extracted in: {}".format(total_cubes, time_taken) )
32.635628
79
0.579395
794d613e2788e0fc47bd62ab625127271573926d
53,180
py
Python
src/Analysis/Ast/Impl/scrape_module.py
jamesralstin/python-language-server
53eb5886776c9e75590bf2f5a787ba4015879c4d
[ "Apache-2.0" ]
null
null
null
src/Analysis/Ast/Impl/scrape_module.py
jamesralstin/python-language-server
53eb5886776c9e75590bf2f5a787ba4015879c4d
[ "Apache-2.0" ]
null
null
null
src/Analysis/Ast/Impl/scrape_module.py
jamesralstin/python-language-server
53eb5886776c9e75590bf2f5a787ba4015879c4d
[ "Apache-2.0" ]
null
null
null
# Python Tools for Visual Studio # Copyright(c) Microsoft Corporation # All rights reserved. # # Licensed under the Apache License, Version 2.0 (the License); you may not use # this file except in compliance with the License. You may obtain a copy of the # License at http://www.apache.org/licenses/LICENSE-2.0 # # THIS CODE IS PROVIDED ON AN *AS IS* BASIS, WITHOUT WARRANTIES OR CONDITIONS # OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING WITHOUT LIMITATION ANY # IMPLIED WARRANTIES OR CONDITIONS OF TITLE, FITNESS FOR A PARTICULAR PURPOSE, # MERCHANTABLITY OR NON-INFRINGEMENT. # # See the Apache Version 2.0 License for specific language governing # permissions and limitations under the License. from __future__ import print_function __author__ = "Microsoft Corporation <ptvshelp@microsoft.com>" __version__ = "15.8" import ast import keyword import inspect import io import re import sys import tokenize import warnings if sys.version_info >= (3, 0): try: import builtins except ImportError: import __builtin__ as builtins else: import __builtin__ as builtins try: bytes except NameError: bytes = str try: unicode except NameError: unicode = str class InspectWarning(UserWarning): pass def _triple_quote(s): if "'" not in s: return "'''" + s + "'''" if '"' not in s: return '"""' + s + '"""' if not s.startswith("'"): return "'''" + s.replace("'''", "\\'\\'\\'") + " '''" if not s.startswith('"'): return '"""' + s.replace('"""', '\\"\\"\\"') + ' """' return "''' " + s.replace("'''", "\\'\\'\\'") + " '''" try: EXACT_TOKEN_TYPES = tokenize.EXACT_TOKEN_TYPES except AttributeError: # Bare minimum that we need here EXACT_TOKEN_TYPES = { '(': tokenize.LPAR, ')': tokenize.RPAR, '[': tokenize.LSQB, ']': tokenize.RSQB, '{': tokenize.LBRACE, '}': tokenize.RBRACE, ',': tokenize.COMMA, ':': tokenize.COLON, '*': tokenize.STAR, '**': tokenize.DOUBLESTAR, '=': tokenize.EQUAL, } SKIP_TYPENAME_FOR_TYPES = bool, str, bytes, int, float STATICMETHOD_TYPES = () CLASSMETHOD_TYPES = type(float.fromhex), PROPERTY_TYPES = type(int.real), type(property.fget) INVALID_ARGNAMES = set(keyword.kwlist) | set(('None', 'True', 'False')) # These full names are known to be lies. When we encounter # them while scraping a module, assume that we need to write # out the full type rather than including them by reference. LIES_ABOUT_MODULE = frozenset([ builtins.__name__ + ".weakcallableproxy", builtins.__name__ + ".weakproxy", builtins.__name__ + ".weakref", "ctypes.ArgumentError", "os.stat_result", "os.statvfs_result", "xml.parsers.expat.ExpatError", "numpy.broadcast", "numpy.busdaycalendar", "numpy.dtype", "numpy.flagsobj", "numpy.flatiter", "numpy.ndarray", "numpy.nditer", # These modules contain multiple members that lie about their # module. Always write out all members of these in full. "_asyncio.*", "_bsddb.*", "_decimal.*", "_elementtree.*", "_socket.*", "_sqlite3.*", "_ssl.*", "_testmultiphase.*", ]) # These type names cause conflicts with their values, so # we need to forcibly rename them. SYS_INFO_TYPES = frozenset(( "float_info", "hash_info", "int_info", "thread_info", "version_info", )) VALUE_REPR_FIX = { float('inf'): "float('inf')", float('-inf'): "float('-inf')", } IMPLICIT_CLASSMETHOD = ( '__new__', ) if sys.version_info[0] < 3: SKIP_TYPENAME_FOR_TYPES += unicode, long def safe_callable(v): try: return hasattr(v, '__call__') except Exception: return False def safe_module_name(n): if n: return '_mod_' + n.replace('.', '_') return n def do_not_inspect(v): # https://github.com/Microsoft/python-language-server/issues/740 # https://github.com/cython/cython/issues/1470 if type(v).__name__ != "fused_cython_function": return False # If a fused function has __defaults__, then attempting to access # __kwdefaults__ will fail if generated before cython 0.29.6. return bool(getattr(v, "__defaults__", False)) class Signature(object): # These two dictionaries start with Python 3 values. # There is an update below for Python 2 differences. # They will be used as fallbacks for known protocols KNOWN_RESTYPES = { "__abs__": "__T__()", "__add__": "__T__()", "__and__": "__T__()", "__annotations__": "{}", "__base__": "type", "__bases__": "(type,)", "__bool__": "False", "__call__": "Any", "__ceil__": "__T__()", "__code__": "object()", "__contains__": "False", "__del__": "None", "__delattr__": "None", "__delitem__": "None", "__dict__": "{'': Any}", "__dir__": "['']", "__divmod__": "(0, 0)", "__eq__": "False", "__format__": "''", "__float__": "0.0", "__floor__": "__T__()", "__floordiv__": "0", "__ge__": "False", "__get__": "__T__()", "__getattribute__": "Any", "__getitem__": "Any", "__getnewargs__": "()", "__getnewargs_ex__": "((), {})", "__getslice__": "__T__()", "__globals__": "{}", "__gt__": "False", "__hash__": "0", "__iadd__": "None", "__iand__": "None", "__imul__": "None", "__index__": "0", "__init__": "", "__init_subclass__": "None", "__int__": "0", "__invert__": "__T__()", "__ior__": "None", "__isub__": "None", "__iter__": "__T__()", "__ixor__": "None", "__le__": "False", "__len__": "0", "__length_hint__": "0", "__lshift__": "__T__()", "__lt__": "False", "__mod__": "__T__()", "__mul__": "__T__()", "__ne__": "False", "__neg__": "__T__()", "__next__": "Any", "__pos__": "__T__()", "__pow__": "__T__()", "__or__": "__T__()", "__radd__": "__T__()", "__rand__": "__T__()", "__rdivmod__": "(0, 0)", "__rfloordiv__": "__T__()", "__rlshift__": "__T__()", "__rmod__": "__T__()", "__rmul__": "__T__()", "__ror__": "__T__()", "__round__": "__T__()", "__rpow__": "__T__()", "__rrshift__": "__T__()", "__rshift__": "__T__()", "__rsub__": "__T__()", "__rtruediv__": "__T__()", "__rxor__": "__T__()", "__reduce__": ["''", "()"], "__reduce_ex__": ["''", "()"], "__repr__": "''", "__set__": "None", "__setattr__": "None", "__setitem__": "None", "__setstate__": "None", "__sizeof__": "0", "__str__": "''", "__sub__": "__T__()", "__truediv__": "0.0", "__trunc__": "__T__()", "__xor__": "__T__()", "__subclasscheck__": "False", "__subclasshook__": "False", } KNOWN_ARGSPECS = { "__contains__": "(self, value)", "__del__": "(self)", "__dir__": "(self)", "__floor__": "(self)", "__format__": "(self, format_spec)", "__getitem__": "(self, index)", "__getnewargs__": "(self)", "__getnewargs_ex__": "(self)", "__init_subclass__": "(cls)", "__instancecheck__": "(self, instance)", "__length_hint__": "(self)", "__prepare__": "(cls, name, bases, **kwds)", "__round__": "(self, ndigits=0)", "__reduce__": "(self)", "__reduce_ex__": "(self, protocol)", "__reversed__": "(self)", "__setitem__": "(self, index, value)", "__setstate__": "(self, state)", "__sizeof__": "(self)", "__subclasses__": "(cls)", "__subclasscheck__": "(cls, subclass)", "__subclasshook__": "(cls, subclass)", "__trunc__": "(self)", } def __init__(self, name, callable, scope=None, defaults=None, scope_alias=None, decorators=None, module_doc=None, ): self.callable = callable self.name = name self.scope = scope self.decorators = decorators or () self._signature = None self._defaults = defaults or () if scope and '@staticmethod' not in self.decorators: def_arg = 'cls' if ('@classmethod' in self.decorators or name in IMPLICIT_CLASSMETHOD) else 'self' if len(self._defaults) == 0 or self._defaults[0] != def_arg: self._defaults = (def_arg,) + self._defaults self.fullsig = None if self.name in ('__init__', '__new__') and module_doc: self.fullsig = self._init_argspec_fromdocstring(self._defaults, module_doc, override_name=self.name) elif not hasattr(self.callable, '__call__') and hasattr(self.callable, '__get__'): # We have a property self.decorators = '@property', self.fullsig = self.name + "(" + ", ".join(self._defaults) + ")" self.fullsig = ( self.fullsig or # Disable fromsignature() because it doesn't work as well as argspec #self._init_argspec_fromsignature(self._defaults) or self._init_argspec_fromargspec(self._defaults) or self._init_argspec_fromknown(self._defaults, scope_alias) or self._init_argspec_fromdocstring(self._defaults) or (self.name + "(" + ", ".join(self._defaults) + ")") ) self.restype = ( self._init_restype_fromsignature() or self._init_restype_fromknown(scope_alias) or self._init_restype_fromdocstring() or 'pass' ) if scope: self.restype = self.restype.replace('__T__', scope) if self.restype in ('return Any', 'return Unknown'): self.restype = 'pass' #Special case for 'with' statement and built-ins like open() or memoryview if name == '__enter__' and self.restype == 'pass': self.restype = 'return self' def __str__(self): return self.fullsig def _init_argspec_fromsignature(self, defaults): if do_not_inspect(self.callable): return try: sig = inspect.signature(self.callable) except Exception: return new_args = [] for arg in sig.parameters: p = sig.parameters[arg] if p.default != inspect.Signature.empty: try: ast.literal_eval(repr(p.default)) except Exception: p = p.replace(default=None) if p.kind == inspect.Parameter.POSITIONAL_ONLY: p = p.replace(kind=inspect.Parameter.POSITIONAL_OR_KEYWORD) new_args.append(p) sig = sig.replace(parameters=new_args) return self.name + str(sig) def _init_restype_fromsignature(self): if do_not_inspect(self.callable): return try: sig = inspect.signature(self.callable) except Exception: return # If signature has a return annotation, it's in the # full signature and we don't need it from here. if not sig or sig.return_annotation == inspect._empty: return ann = inspect.formatannotation(sig.return_annotation) if not ann or not self._can_eval(ann): return return 'return ' + ann def _init_argspec_fromargspec(self, defaults): if do_not_inspect(self.callable): return try: args = (getattr(inspect, 'getfullargspec', None) or inspect.getargspec)(self.callable) except Exception: return argn = [] seen_names = set(INVALID_ARGNAMES) defaults = list(defaults) default_set = set(defaults) for a in args.args: if a in default_set: default_set.discard(a) argn.append(self._make_unique_name(a, seen_names)) if default_set: argn[:0] = [a for a in defaults if a in default_set] if getattr(args, 'varargs', None): argn.append('*' + args.varargs) if getattr(args, 'varkw', None): argn.append('**' + args.varkw) if argn and argn[-1] in ('*', '**'): argn[-1] += self._make_unique_name('_', seen_names) return self.name + '(' + ', '.join(argn) + ')' def _init_argspec_fromknown(self, defaults, scope_alias): spec = None if scope_alias and not spec: spec = self.KNOWN_ARGSPECS.get(scope_alias + '.' + self.name) if self.scope and not spec: spec = self.KNOWN_ARGSPECS.get(self.scope + '.' + self.name) if not spec: spec = self.KNOWN_ARGSPECS.get(self.name) if not spec: return return self.name + spec def _init_restype_fromknown(self, scope_alias): restype = None if scope_alias and not restype: restype = self.KNOWN_RESTYPES.get(scope_alias + '.' + self.name) if self.scope and not restype: restype = self.KNOWN_RESTYPES.get(self.scope + '.' + self.name) if not restype: restype = self.KNOWN_RESTYPES.get(self.name) if not restype: return if isinstance(restype, list): return "return " + "; return ".join(restype) return "return " + restype def _init_restype_fromdocstring(self): doc = getattr(self.callable, '__doc__', None) if not isinstance(doc, str): return first_line = doc.partition('\n')[0].strip() if not '->' in first_line: return index = first_line.index('->') typeName = first_line[index + 2:].strip() if typeName.startswith('str'): return "return ''" if typeName.startswith('float'): return "return 1.0" if typeName.startswith('int'): return "return 1" if typeName.startswith('long'): if sys.version_info[0] < 3: return "return 1L" else: return "return 1" if typeName.startswith('list'): return "return list()" if typeName.startswith('dict'): return "return dict()" if typeName.startswith('('): return "return tuple()" if typeName.startswith('bool'): return "return True" if 'Return a string' in first_line: return "return ''" return def _init_argspec_fromdocstring(self, defaults, doc=None, override_name=None): if not doc: doc = getattr(self.callable, '__doc__', None) if not isinstance(doc, str): return doc = self._get_first_function_call(doc) if not doc: return if(override_name): allow_name_mismatch = override_name not in doc else: allow_name_mismatch = False return self._parse_funcdef(doc, allow_name_mismatch, defaults, override_name) def _make_unique_name(self, name, seen_names): if name not in seen_names: seen_names.add(name) return name n = name + '_' if n not in seen_names: seen_names.add(n) return n i = 0 while True: i += 1 n = name + '_' + str(i) if n not in seen_names: seen_names.add(n) return n raise RuntimeError("Too many arguments in definition") def _tokenize(self, expr): if sys.version_info[0] == 3 and sys.version_info[1] <= 2: expr = '# coding: utf-8\n' + expr buf = io.BytesIO(expr.strip().encode('utf-8')) if sys.version_info[0] == 3: tokens = tokenize.tokenize(buf.readline) else: tokens = tokenize.generate_tokens(buf.readline) return [(EXACT_TOKEN_TYPES.get(s, tt) if tt == tokenize.OP else tt, s) for tt, s, _, _, _ in tokens] _PAREN_TOKEN_MAP = { tokenize.LPAR: tokenize.RPAR, tokenize.LBRACE: tokenize.RBRACE, tokenize.LSQB: tokenize.RSQB, } def _parse_take_expr(self, tokens, *stop_at): nesting = [] expr = [] while tokens: tt, s = tokens[0] if tt == tokenize.LSQB and len(tokens) > 2 and tokens[1][0] in stop_at: return expr if tt in self._PAREN_TOKEN_MAP: expr.append((tt, s)) nesting.append(self._PAREN_TOKEN_MAP[tt]) elif nesting and tt == nesting[-1]: expr.append((tt, s)) nesting.pop() elif tt in (tokenize.RPAR, tokenize.RBRACE, tokenize.RSQB): return expr elif not nesting and tt in stop_at: return expr else: expr.append((tt, s)) tokens.pop(0) return expr def _can_eval(self, s): if not s: return False try: ast.parse(s, mode='eval') except SyntaxError: return False else: return True def _parse_format_arg(self, name, args, defaults): defaults = list(defaults) default_set = set(defaults) seen_names = set(INVALID_ARGNAMES) parts = [name or '<function>', '('] arg_parts = [] any_default = False for a_names, a_ann, a_def, a_opt in args: if not a_names: continue a_name = ''.join(a_names) if a_name in default_set: default_set.discard(a_name) arg_parts.append(self._make_unique_name(a_name, seen_names)) if self._can_eval(''.join(a_ann)): arg_parts.append(': ') arg_parts.extend(a_ann) if self._can_eval(''.join(a_def)): arg_parts.append('=') arg_parts.extend(a_def) any_default = True elif a_opt[0] or (any_default and '*' not in a_name and '**' not in a_name): arg_parts.append('=None') any_default = True if a_name.startswith('*'): any_default = True arg_parts.append(', ') if default_set: for a in defaults: if a in default_set: parts.append(a) parts.append(', ') parts.extend(arg_parts) if parts[-1] == ', ': parts.pop() if parts and parts[-1] in ('*', '**'): parts[-1] += self._make_unique_name('_', seen_names) parts.append(')') return ''.join(parts) def _parse_funcdef(self, expr, allow_name_mismatch, defaults, override_name=None): '''Takes a call expression that was part of a docstring and parses the AST as if it were a definition. If the parsed AST matches the callable we are wrapping, returns the node. ''' try: tokens = self._tokenize(expr) except (TypeError, tokenize.TokenError): warnings.warn('failed to tokenize ' + expr, InspectWarning) return name = None seen_open_paren = False args = [([], [], [], [False])] optional = False while tokens: tt, s = tokens.pop(0) if tt == tokenize.NAME: if name is None: name = s elif seen_open_paren: args[-1][0].append(s) args[-1][3][0] = optional elif tt in (tokenize.STAR, tokenize.DOUBLESTAR): args[-1][0].append(s) elif tt == tokenize.COLON: e = self._parse_take_expr(tokens, tokenize.EQUAL, tokenize.COMMA) args[-1][1].append(''.join(i[1] for i in e)) elif tt == tokenize.EQUAL: e = self._parse_take_expr(tokens, tokenize.COMMA) args[-1][2].append(''.join(i[1] for i in e)) elif tt == tokenize.COMMA: args.append(([], [], [], [False])) elif tt == tokenize.LSQB: optional = True elif tt == tokenize.RSQB: optional = False elif tt == tokenize.LPAR: seen_open_paren = True elif tt == tokenize.RPAR: break elif s in ('->', '...'): return if name and (allow_name_mismatch or name == self.name): return self._parse_format_arg(override_name or name, args, defaults) def _get_first_function_call(self, expr): '''Scans the string for the first closing parenthesis, handling nesting, which is the best heuristic we have for an example call at the start of the docstring.''' if not expr or ')' not in expr: return found = [] n = 0 for line in expr.splitlines(): found_one = False line = line.strip('\r\n\t ') if not line: break for i, c in enumerate(line): if c == ')': n -= 1 if n == 0: found.append(line[:i + 1]) found_one = True elif c == '(': n += 1 if not found_one: break if found: found.sort(key=len) return found[-1] return class MemberInfo(object): NO_VALUE = object() def __init__(self, name, value, literal=None, scope=None, module=None, alias=None, module_doc=None, scope_alias=None): self.name = name self.value = value self.literal = literal self.members = [] self.values = [] self.need_imports = () self.type_name = None self.scope_name = None self.bases = () self.signature = None self.documentation = getattr(value, '__doc__', None) self.alias = alias if not isinstance(self.documentation, str): self.documentation = None # Special case for __init__ that refers to class docs if self.name == '__init__' and ( not self.documentation or 'See help(type(self))' in self.documentation): self.documentation = module_doc if self.name: self.name = self.name.replace('-', '_') value_type = type(value) if issubclass(value_type, type): self.need_imports, type_name = self._get_typename(value, module) if '.' in type_name: m, s, n = type_name.rpartition('.') self.literal = safe_module_name(m) + s + n else: self.scope_name = self.type_name = type_name self._collect_bases(value, module, self.type_name) elif safe_callable(value): dec = () if scope: if value_type in STATICMETHOD_TYPES: dec += '@staticmethod', elif value_type in CLASSMETHOD_TYPES: dec += '@classmethod', self.signature = Signature(name, value, scope, scope_alias=scope_alias, decorators=dec, module_doc=module_doc) elif value is not None: if value_type in PROPERTY_TYPES: self.signature = Signature(name, value, scope, scope_alias=scope_alias) if value_type not in SKIP_TYPENAME_FOR_TYPES: self.need_imports, self.type_name = self._get_typename(value_type, module) self._collect_bases(value_type, module, self.type_name) if isinstance(value, float) and repr(value) == 'nan': self.literal = "float('nan')" try: self.literal = VALUE_REPR_FIX[value] except Exception: pass elif not self.literal: self.literal = 'None' def _collect_bases(self, value_type, module, type_name): try: bases = getattr(value_type, '__bases__', ()) except Exception: pass else: self.bases = [] self.need_imports = list(self.need_imports) for ni, t in (self._get_typename(b, module) for b in bases): if not t: continue if t == type_name and module in ni: continue self.bases.append(t) self.need_imports.extend(ni) @classmethod def _get_typename(cls, value_type, in_module): try: type_name = value_type.__name__.replace('-', '_') module = getattr(value_type, '__module__', None) # Special workaround for Python 2 exceptions lying about their module if sys.version_info[0] == 2 and module == 'exceptions' and in_module == builtins.__name__: module = builtins.__name__ # Special workaround for IronPython types that include their module name if in_module and type_name.startswith(in_module + '.'): type_name = type_name[len(in_module) + 1:] module = in_module if module and module != '<unknown>': if module == in_module: return (module,), type_name fullname = module + '.' + type_name if in_module and (fullname in LIES_ABOUT_MODULE or (in_module + '.*') in LIES_ABOUT_MODULE): # Treat the type as if it came from the current module return (in_module,), type_name return (module,), fullname return (), type_name except Exception: warnings.warn('could not get type of ' + repr(value_type), InspectWarning) raise return (), None def _str_from_literal(self, lit): return self.name + ' = ' + lit def _str_from_typename(self, type_name): mod_name, sep, name = type_name.rpartition('.') return self.name + ' = ' + safe_module_name(mod_name) + sep + name + '()' def _str_from_value(self, v): return self.name + ' = ' + repr(v) def _lines_with_members(self): if self.bases: split_bases = [n.rpartition('.') for n in self.bases] bases = ','.join((safe_module_name(n[0]) + n[1] + n[2]) for n in split_bases) yield 'class ' + self.name + '(' + bases + '):' else: yield 'class ' + self.name + ':' if self.documentation: yield ' ' + repr(self.documentation) if self.members: for mi in self.members: if hasattr(mi, 'documentation') and mi.documentation != None and not isinstance(mi.documentation, str): continue if mi is not MemberInfo.NO_VALUE: yield mi.as_str(' ') else: yield ' pass' yield '' def _lines_with_signature(self): seen_decorators = set() for d in self.signature.decorators: d = str(d) if d not in seen_decorators: seen_decorators.add(d) yield d yield 'def ' + str(self.signature) + ':' if self.documentation: yield ' ' + repr(self.documentation) if self.signature.restype: yield ' ' + self.signature.restype else: yield ' pass' yield '' def as_str(self, indent=''): if self.literal: return indent + self._str_from_literal(self.literal) if self.members: return '\n'.join(indent + s for s in self._lines_with_members()) if self.signature: return '\n'.join(indent + s for s in self._lines_with_signature()) if self.type_name is not None: return indent + self._str_from_typename(self.type_name) if self.value is not None: return indent + self._str_from_value(self.value) return indent + self.name MODULE_MEMBER_SUBSTITUTE = { '__builtins__': MemberInfo('__builtins__', None, literal='{}'), '__spec__': None, '__loader__': None, } CLASS_MEMBER_SUBSTITUTE = { '__bases__': MemberInfo('__bases__', None, literal='()'), '__mro__': MemberInfo('__mro__', None, literal='()'), '__dict__': MemberInfo('__dict__', None, literal='{}'), '__doc__': None, '__new__': None, } class ScrapeState(object): def __init__(self, module_name, module=None): self.root_module = None self.module = module self.module_name = module_name self.imports = set() self.members = [] def initial_import(self, search_path=None): if self.module: return if search_path: sys.path.insert(0, search_path) try: try: mod = __import__(self.module_name) except Exception: ex_msg = str(sys.exc_info()[1]) warnings.warn("Working around " + ex_msg, InspectWarning) if ex_msg == "This must be an MFC application - try 'import win32ui' first": import win32ui elif ex_msg == "Could not find TCL routines" or self.module_name == 'matplotlib.backends._tkagg': if sys.version_info[0] == 2: import Tkinter else: import tkinter else: raise mod = None if not mod: # Try the import again, either fixed or without chaining the # previous exception. mod = __import__(self.module_name) finally: if search_path: del sys.path[0] self.root_module = mod # __import__ gives us the topmost module. We should generally use # getattr() from here to find the child module. However, sometimes # this is unsuccessful. So we work backwards through the full name # to see what is in sys.modules, then getattr() to go forwards. mod_name = self.module_name bits = [] while mod_name and mod_name not in sys.modules: mod_name, _, bit = self.module_name.rpartition('.')[0] bits.insert(0, bit) if mod_name: self.root_module = mod = sys.modules[mod_name] else: bits = self.module_name.split('.')[1:] for bit in bits: mod = getattr(mod, bit) self.module = mod def collect_top_level_members(self): self._collect_members(self.module, self.members, MODULE_MEMBER_SUBSTITUTE, None) if self.module_name == 'sys': sysinfo = [m for m in self.members if m.type_name in SYS_INFO_TYPES] for m in sysinfo: self.members.append(MemberInfo(m.name, None, literal="__" + m.name + "()")) m.name = m.scope_name = m.type_name = '__' + m.type_name m_names = set(m.name for m in self.members) undeclared = [] for m in self.members: if m.value is not None and m.type_name and '.' not in m.type_name and m.type_name not in m_names: undeclared.append(MemberInfo(m.type_name, type(m.value), module=self.module_name)) self.members[:0] = undeclared def _should_collect_members(self, member): if self.module_name in member.need_imports and member.name == member.type_name: return True # Support cffi libs if member.type_name == builtins.__name__ + '.CompiledLib': return True def collect_second_level_members(self): for mi in self.members: if self._should_collect_members(mi): substitutes = dict(CLASS_MEMBER_SUBSTITUTE) substitutes['__class__'] = MemberInfo('__class__', None, literal=mi.type_name) self._collect_members(mi.value, mi.members, substitutes, mi) if mi.scope_name != mi.type_name: # When the scope and type names are different, we have a static # class. To emulate this, we add '@staticmethod' decorators to # all members. for mi2 in mi.members: if mi2.signature: mi2.signature.decorators += '@staticmethod', def _collect_members(self, mod, members, substitutes, outer_member): '''Fills the members attribute with a dictionary containing all members from the module.''' if not mod: raise RuntimeError("failed to import module") if mod is MemberInfo.NO_VALUE: return existing_names = set(m.name for m in members) if outer_member: scope = outer_member.scope_name scope_alias = outer_member.alias else: scope, scope_alias = None, None mod_scope = (self.module_name + '.' + scope) if scope else self.module_name mod_doc = getattr(mod, '__doc__', None) mro = (getattr(mod, '__mro__', None) or ())[1:] for name in dir(mod): if keyword.iskeyword(name): continue try: m = substitutes[name] if m: members.append(m) continue except LookupError: pass try: m = substitutes[mod_scope + '.' + name] if m: members.append(m) continue except LookupError: pass if name in existing_names: continue try: value = getattr(mod, name) except AttributeError: warnings.warn("attribute " + name + " on " + repr(mod) + " was in dir() but not getattr()", InspectWarning) except Exception: warnings.warn("error getting " + name + " for " + repr(mod), InspectWarning) else: if not self._should_add_value(value): continue if name != '__init__' and self._mro_contains(mro, name, value): continue members.append(MemberInfo(name, value, scope=scope, module=self.module_name, module_doc=mod_doc, scope_alias=scope_alias)) def _should_add_value(self, value): try: value_type = type(value) mod = getattr(value_type, '__module__', None) name = value_type.__name__ except Exception: warnings.warn("error getting typename", InspectWarning) return if (mod, name) == (builtins.__name__, 'CompiledLib'): # Always allow CFFI lib return True if issubclass(value_type, (type(sys), type(inspect))): # Disallow nested modules return # By default, include all values return True def _mro_contains(self, mro, name, value): for m in mro: try: mro_value = getattr(m, name) except Exception: pass else: if mro_value is value: return True def translate_members(self): pass def dump(self, out): imports = set() for value in self.members: for mod in value.need_imports: imports.add(mod) imports.discard(self.module_name) if imports: for mod in sorted(imports): print("import " + mod + " as " + safe_module_name(mod), file=out) print("", file=out) for value in self.members: s = value.as_str('') try: print(s, file=out) except TypeError: print(repr(s), file=sys.stderr) raise def add_builtin_objects(state): Signature.KNOWN_RESTYPES.update({ "__Type__.__call__": "cls()", "__Property__.__delete__": "None", "__Float__.__getformat__": "''", "__Bytes__.__getitem__": "__T__()", "__Unicode__.__getitem__": "__T__()", "__Type__.__instancecheck__": "False", "__Tuple__.__iter__": "__TupleIterator__()", "__List__.__iter__": "__ListIterator__()", "__Dict__.__iter__": "__DictKeys__()", "__Set__.__iter__": "__SetIterator__()", "__FrozenSet__.__iter__": "__SetIterator__()", "__Bytes__.__iter__": "__BytesIterator__()", "__Unicode__.__iter__": "__UnicodeIterator__()", "__BytesIterator__.__next__": "0", "__UnicodeIterator__.__next__": "__Unicode__()", "__Type__.__prepare__": "None", "__List__.__reversed__": "__ListIterator__()", "__Float__.__setformat__": "None", "__Type__.__subclasses__": "(cls,)", "__truediv__": "__Float__()", "__Type__.__subclasscheck__": "False", "__subclasshook__": "False", "all": "False", "any": "False", "ascii": "''", "__Set__.add": "None", "__List__.append": "None", "__Float__.as_integer_ratio": "(0, 0)", "bin": "''", "__Int__.bit_length": "0", "callable": "False", "capitalize": "__T__()", "casefold": "__T__()", "center": "__T__()", "chr": "''", "clear": "None", "__Generator__.close": "None", "conjugate": "__Complex__()", "copy": "__T__()", "count": "0", "__Bytes__.decode": "''", "__Property__.deleter": "func", "__Set__.difference": "__T__()", "__FrozenSet__.difference": "__T__()", "__Set__.difference_update": "None", "__Set__.discard": "None", "divmod": "(0, 0)", "__Bytes__.encode": "b''", "__Unicode__.encode": "b''", "endswith": "False", "expandtabs": "__T__()", "__List__.extend": "None", "find": "0", "__Unicode__.format": "__T__()", "__Unicode__.format_map": "__T__()", "__Bool__.from_bytes": "False", "__Int__.from_bytes": "0", "__Long__.from_bytes": "__Long__()", "__Float__.fromhex": "0.0", "__Bytes__.fromhex": "b''", "__Dict__.fromkeys": "{}", "__Dict__.get": "self[0]", "__Property__.getter": "func", "format": "''", "globals": "__Dict__()", "hasattr": "False", "hash": "0", "hex": "''", "id": "0", "index": "0", "input": "''", "__List__.insert": "None", "__Set__.intersection": "__T__()", "__FrozenSet__.intersection": "__T__()", "__Set__.intersection_update": "None", "isalnum": "False", "isalpha": "False", "isdecimal": "False", "isdigit": "False", "islower": "False", "isidentifier": "False", "isnumeric": "False", "isprintable": "False", "isspace": "False", "istitle": "False", "isupper": "False", "__Float__.is_integer": "False", "__Set__.isdisjoint": "False", "__FrozenSet__.isdisjoint": "False", "__DictKeys__.isdisjoint": "False", "__DictItems__.isdisjoint": "False", "__Set__.issubset": "False", "__FrozenSet__.issubset": "False", "__Set__.issuperset": "False", "__FrozenSet__.issuperset": "False", "__Dict__.items": "__DictItems__()", "__Bytes__.join": "b''", "__Unicode__.join": "''", "__Dict__.keys": "__DictKeys__()", "len": "0", "locals": "__Dict__()", "lower": "__T__()", "ljust": "__T__()", "lstrip": "__T__()", "__Bytes__.maketrans": "b''", "__Unicode__.maketrans": "{}", "__Type__.mro": "[__Type__()]", "oct": "''", "partition": "(__T__(), __T__(), __T__())", "__List__.pop": "self[0]", "__Dict__.pop": "self.keys()[0]", "__Set__.pop": "Any", "__Dict__.popitem": "self.items()[0]", "remove": "None", "replace": "__T__()", "repr": "''", "rfind": "0", "__List__.reverse": "None", "rindex": "0", "rjust": "__T__()", "round": "0.0", "rpartition": "(__T__(), __T__(), __T__())", "rsplit": "[__T__()]", "rstrip": "__T__()", "__Generator__.send": "self.__next__()", "__Dict__.setdefault": "self[0]", "__Property__.setter": "func", "__List__.sort": "None", "sorted": "__List__()", "split": "[__T__()]", "splitlines": "[self()]", "startswith": "False", "strip": "__T__()", "swapcase": "__T__()", "__Set__.symmetric_difference": "__T__()", "__FrozenSet__.symmetric_difference": "__T__()", "__Set__.symmetric_difference_update": "None", "__Bytes__.translate": "__T__()", "__Unicode__.translate": "__T__()", "__Generator__.throw": "None", "title": "__T__()", "to_bytes": "b''", "__Set__.union": "__T__()", "__FrozenSet__.union": "__T__()", "__Dict__.update": "None", "__Set__.update": "None", "upper": "__T__()", "__Dict__.values": "__DictValues__()", "zfill": "__T__()", }) Signature.KNOWN_ARGSPECS.update({ "__Type__.__call__": "(cls, *args, **kwargs)", "__Int__.__ceil__": "(self)", "__Int__.__floor__": "(self)", "__Float__.__getformat__": "(typestr)", "__Dict__.__getitem__": "(self, key)", "__Type__.__instancecheck__": "(self, instance)", "__Bool__.__init__": "(self, x)", "__Int__.__init__": "(self, x=0)", "__List__.__init__": "(self, iterable)", "__Tuple__.__init__": "(self, iterable)", "__Type__.__prepare__": "(cls, name, bases, **kwds)", "__Int__.__round__": "(self, ndigits=0)", "__Float__.__round__": "(self, ndigits=0)", "__List__.__reversed__": "(self)", "__Float__.__setformat__": "(typestr, fmt)", "__Dict__.__setitem__": "(self, key, value)", "__Set__.add": "(self, value)", "__List__.append": "(self, value)", "__Float__.as_integer_ratio": "(self)", "__Int__.bit_length": "(self)", "capitalize": "(self)", "casefold": "(self)", "__Bytes__.center": "(self, width, fillbyte=b' ')", "__Unicode__.center": "(self, width, fillchar=' ')", "clear": "(self)", "__Generator__.close": "(self)", "conjugate": "(self)", "copy": "(self)", "count": "(self, x)", "__Bytes__.count": "(self, sub, start=0, end=-1)", "__Unicode__.count": "(self, sub, start=0, end=-1)", "__Bytes__.decode": "(self, encoding='utf-8', errors='strict')", "__Property__.deleter": "(self, func)", "__Set__.difference": "(self, other)", "__FrozenSet__.difference": "(self, other)", "__Set__.difference_update": "(self, *others)", "__Set__.discard": "(self, elem)", "__Unicode__.encode": "(self, encoding='utf-8', errors='strict')", "endswith": "(self, suffix, start=0, end=-1)", "expandtabs": "(self, tabsize=8)", "__List__.extend": "(self, iterable)", "find": "(self, sub, start=0, end=-1)", "__Unicode__.format": "(self, *args, **kwargs)", "__Unicode__.format_map": "(self, mapping)", "__Bool__.from_bytes": "(bytes, byteorder, *, signed=False)", "__Int__.from_bytes": "(bytes, byteorder, *, signed=False)", "__Float__.fromhex": "(string)", "__Dict__.get": "(self, key, d=None)", "__Property__.getter": "(self, func)", "hex": "(self)", "__List__.insert": "(self, index, value)", "index": "(self, v)", "__Bytes__.index": "(self, sub, start=0, end=-1)", "__Unicode__.index": "(self, sub, start=0, end=-1)", "__Set__.intersection": "(self, other)", "__FrozenSet__.intersection": "(self, other)", "__Set__.intersection_update": "(self, *others)", "isalnum": "(self)", "isalpha": "(self)", "isdecimal": "(self)", "isdigit": "(self)", "isidentifier": "(self)", "islower": "(self)", "isnumeric": "(self)", "isprintable": "(self)", "isspace": "(self)", "istitle": "(self)", "isupper": "(self)", "__Float__.is_integer": "(self)", "__Set__.isdisjoint": "(self, other)", "__FrozenSet__.isdisjoint": "(self, other)", "__DictKeys__.isdisjoint": "(self, other)", "__DictItems__.isdisjoint": "(self, other)", "__Set__.issubset": "(self, other)", "__FrozenSet__.issubset": "(self, other)", "__Set__.issuperset": "(self, other)", "__FrozenSet__.issuperset": "(self, other)", "__Dict__.items": "(self)", "__Bytes__.join": "(self, iterable)", "__Unicode__.join": "(self, iterable)", "__Dict__.keys": "(self)", "lower": "(self)", "__Bytes__.ljust": "(self, width, fillbyte=b' ')", "__Unicode__.ljust": "(self, width, fillchar=' ')", "lstrip": "(self, chars)", "__Bytes__.maketrans": "(from_, to)", "__Unicode__.maketrans": "(x, y, z)", "__Type__.mro": "(cls)", "__Bytes__.partition": "(self, sep)", "__Unicode__.partition": "(self, sep)", "__List__.pop": "(self, index=-1)", "__Dict__.pop": "(self, k, d=None)", "__Set__.pop": "(self)", "__Dict__.popitem": "(self, k, d=None)", "__List__.remove": "(self, value)", "__Set__.remove": "(self, elem)", "replace": "(self, old, new, count=-1)", "__List__.reverse": "(self)", "rfind": "(self, sub, start=0, end=-1)", "rindex": "(self, sub, start=0, end=-1)", "__Bytes__.rjust": "(self, width, fillbyte=b' ')", "__Unicode__.rjust": "(self, width, fillchar=' ')", "__Bytes__.rpartition": "(self, sep)", "__Unicode__.rpartition": "(self, sep)", "rsplit": "(self, sep=None, maxsplit=-1)", "rstrip": "(self, chars=None)", "__Generator__.send": "(self, value)", "__Dict__.setdefault": "(self, k, d)", "__Property__.setter": "(self, func)", "__List__.sort": "(self)", "split": "(self, sep=None, maxsplit=-1)", "splitlines": "(self, keepends=False)", "strip": "(self, chars=None)", "startswith": "(self, prefix, start=0, end=-1)", "swapcase": "(self)", "__Set__.symmetric_difference": "(self, other)", "__FrozenSet__.symmetric_difference": "(self, other)", "__Set__.symmetric_difference_update": "(self, *others)", "__Generator__.throw": "(self, type, value=None, traceback=None)", "title": "(self)", "__Int__.to_bytes": "(bytes, byteorder, *, signed=False)", "__Bytes__.translate": "(self, table, delete=b'')", "__Unicode__.translate": "(self, table)", "__Set__.union": "(self, *others)", "__FrozenSet__.union": "(self, *others)", "__Dict__.update": "(self, d)", "__Set__.update": "(self, *others)", "upper": "(self)", "__Dict__.values": "(self)", "zfill": "(self, width)", }) if sys.version[0] == '2': Signature.KNOWN_RESTYPES.update({ "__BytesIterator__.__next__": None, "__BytesIterator__.next": "b''", "__UnicodeIterator__.__next__": None, "__UnicodeIterator__.next": "u''", "__Generator__.send": "self.next()", "__Function__.func_closure": "()", "__Function__.func_doc": "b''", "__Function__.func_name": "b''", "input": None, "raw_input": "b''", }) Signature.KNOWN_ARGSPECS.update({ "__BytesIterator__.next": "(self)", "__UnicodeIterator__.next": "(self)", }) already_added = set() def add_simple(name, doc, *members): mi = MemberInfo(name, MemberInfo.NO_VALUE) mi.documentation = doc mi.need_imports = (state.module_name,) mi.members.extend(members) state.members.append(mi) def add_literal(name, literal): state.members.append(MemberInfo(name, None, literal=literal)) def add_type(alias, type_obj): if type_obj.__name__ in already_added: add_literal(alias, type_obj.__name__) return already_added.add(type_obj.__name__) mi = MemberInfo(type_obj.__name__, type_obj, module=builtins.__name__, alias=alias) state.members.append(mi) state.members.append(MemberInfo(alias, None, literal=mi.name)) add_simple('__Unknown__', '<unknown>', MemberInfo("__name__", None, literal='"<unknown>"')) add_simple('__NoneType__', 'the type of the None object', MemberInfo.NO_VALUE) # NoneType and None are explicitly defined to avoid parser errors # because of None being a keyword. #add_literal('NoneType', '__NoneType__') #add_literal('None', '__NoneType__()') add_type('__Object__', object) add_type('__Type__', type) add_type('__Int__', int) if type(bool()) is int: add_literal('__Bool__', '__Int__') else: add_type('__Bool__', bool) try: long except NameError: add_literal('__Long__', '__Int__') else: add_type('__Long__', long) add_type("__Float__", float) add_type("__Complex__", complex) add_type("__Tuple__", tuple) add_type("__List__", list) add_type("__Dict__", dict) add_type("__Set__", set) add_type("__FrozenSet__", frozenset) if bytes is not str: add_type("__Bytes__", bytes) add_type("__BytesIterator__", type(iter(bytes()))) add_type("__Unicode__", str) add_type("__UnicodeIterator__", type(iter(str()))) add_literal("__Str__", "__Unicode__") add_literal("__StrIterator__", "__UnicodeIterator__") else: add_type("__Bytes__", str) add_type("__BytesIterator__", type(iter(str()))) add_type("__Unicode__", unicode) add_type("__UnicodeIterator__", type(iter(unicode()))) add_literal("__Str__", "__Bytes__") add_literal("__StrIterator__", "__BytesIterator__") add_type("__Module__", type(inspect)) add_type("__Function__", type(add_simple)) add_type("__BuiltinMethodDescriptor__", type(object.__hash__)) add_type("__BuiltinFunction__", type(abs)) add_type("__Generator__", type((_ for _ in []))) add_type("__Property__", property) add_type("__ClassMethod__", classmethod) add_type("__StaticMethod__", staticmethod) add_type("__Ellipsis__", type(Ellipsis)) add_type("__TupleIterator__", type(iter(()))) add_type("__ListIterator__", type(iter([]))) add_type("__DictKeys__", type({}.keys())) add_type("__DictValues__", type({}.values())) add_type("__DictItems__", type({}.items())) add_type("__SetIterator__", type(iter(set()))) add_type("__CallableIterator__", type(iter((lambda: None), None))) # Also write out the builtin module names here so that we cache them try: builtin_module_names = sys.builtin_module_names except AttributeError: pass else: add_literal('__builtin_module_names__', '"' + ','.join(builtin_module_names) + '"') if __name__ == '__main__': EXCLUDED_MEMBERS = () outfile = sys.stdout if '-u8' in sys.argv: sys.argv.remove('-u8') try: b_outfile = outfile.buffer except AttributeError: warnings.warn("cannot enable UTF-8 output", InspectWarning) pass # on Python 2, so hopefully everything is valid ASCII... else: import io outfile = io.TextIOWrapper(b_outfile, encoding='utf-8', errors='replace') if len(sys.argv) == 1: state = ScrapeState(builtins.__name__, builtins) add_builtin_objects(state) EXCLUDED_MEMBERS += ('None', 'False', 'True', '__debug__') if sys.version_info[0] == 2: EXCLUDED_MEMBERS += ('print',) elif len(sys.argv) >= 2: state = ScrapeState(sys.argv[1]) if len(sys.argv) >= 3: state.initial_import(sys.argv[2]) else: state.initial_import() state.collect_top_level_members() state.members[:] = [m for m in state.members if m.name not in EXCLUDED_MEMBERS] state.collect_second_level_members() state.dump(outfile) #import io #state.dump(io.BytesIO())
34.96384
138
0.547198
794d61c188c002c8dfb2fc89ec998d79aea1bad5
19,597
py
Python
cave/com.raytheon.viz.gfe/localization/gfe/userPython/textUtilities/FormatterRunner.py
srcarter3/awips2
37f31f5e88516b9fd576eaa49d43bfb762e1d174
[ "Apache-2.0" ]
null
null
null
cave/com.raytheon.viz.gfe/localization/gfe/userPython/textUtilities/FormatterRunner.py
srcarter3/awips2
37f31f5e88516b9fd576eaa49d43bfb762e1d174
[ "Apache-2.0" ]
null
null
null
cave/com.raytheon.viz.gfe/localization/gfe/userPython/textUtilities/FormatterRunner.py
srcarter3/awips2
37f31f5e88516b9fd576eaa49d43bfb762e1d174
[ "Apache-2.0" ]
1
2021-10-30T00:03:05.000Z
2021-10-30T00:03:05.000Z
## # This software was developed and / or modified by Raytheon Company, # pursuant to Contract DG133W-05-CQ-1067 with the US Government. # # U.S. EXPORT CONTROLLED TECHNICAL DATA # This software product contains export-restricted data whose # export/transfer/disclosure is restricted by U.S. law. Dissemination # to non-U.S. persons whether in the United States or abroad requires # an export license or other authorization. # # Contractor Name: Raytheon Company # Contractor Address: 6825 Pine Street, Suite 340 # Mail Stop B8 # Omaha, NE 68106 # 402.291.0100 # # See the AWIPS II Master Rights File ("Master Rights File.pdf") for # further licensing information. ## import TimeRange, AbsTime import logging import TextFormatter import time, os, string, inspect, sys import JUtil, VarDictGroker import RedirectLogging import UFStatusHandler from com.raytheon.uf.viz.core import VizApp from com.raytheon.uf.common.gfe.ifpclient import PyFPClient # # Runs the text formatter to generate text products # # # SOFTWARE HISTORY # # Date Ticket# Engineer Description # ------------- -------- --------- --------------------------------------------- # May 29, 2008 njensen Initial Creation. # Dev 10, 2014 14946 ryu Add getTimeZones() function. # Apr 16, 2015 14946 ryu Fix getTimeZones to return the office TZ if timezone # is not set for any zone in a segment. # Apr 20, 2015 4027 randerso Fixes for formatter autotests # Apr 25, 2015 4952 njensen Updated for new JEP API # May 06, 2015 4467 randerso Convert to upper case before writing to files if # mixed case is not enabled for the product. # Cleaned up file writing code # Jul 29, 2015 4263 dgilling Support updated TextProductManager. # Nov 30, 2015 5129 dgilling Support new IFPClient. # Sep 28, 2016 19293 randerso Log formatter exceptions to formatter log file # Feb 07, 2017 6092 randerso Changed startTime and endTime to be time.struct_times # Feb 26, 2018 7230 mapeters Don't reset DRT time to real time # ## ## # This is a base file that is not intended to be overridden. ## displayNameDict = {} # Set up logging info PLUGIN_NAME = 'com.raytheon.viz.gfe' CATEGORY = 'GFE' DEFAULT_LOG_FILENAME = '/tmp/gfe.log' FILEMODE='w' PATH_MGR = None try: logging.basicConfig(level=logging.INFO) logger = logging.getLogger("FormatterRunner") formatter = logging.Formatter("%(asctime)s:%(name)s:%(levelname)s:%(message)s") # Get the information for the file logger from com.raytheon.uf.common.localization import PathManagerFactory from com.raytheon.uf.common.localization import LocalizationContext LocalizationType = LocalizationContext.LocalizationType LocalizationLevel = LocalizationContext.LocalizationLevel PATH_MGR = PathManagerFactory.getPathManager() except: logging.basicConfig(filename=DEFAULT_LOG_FILENAME,level=logging.DEBUG) logger = logging.getLogger() logger.exception("Exception occurred") ## TODO: Remove use of DataManager in this code. Will need to coordinate with ## the field developers to ensure local site overrides aren't relying on having ## access to it. def executeFromJava(databaseID, site, username, dataMgr, forecastList, logFile, cmdLineVarDict=None, drtTime=None, vtecMode=None, vtecActiveTable="active", testMode=0 ): if type(forecastList) is not list: forecastList = [str(forecastList)] # Set up the file logger for this product # ctx = PATH_MGR.getContext(LocalizationType.valueOf('CAVE_STATIC'), LocalizationLevel.valueOf('USER')) # logFile = PATH_MGR.getFile(ctx, os.path.join('gfe', 'logs', forecastList[0])).getPath() logger.info("logFile: " + str(logFile)) fh = logging.FileHandler(filename=logFile, mode=FILEMODE) fh.setLevel(logging.INFO) fh.setFormatter(formatter) logger.addHandler(fh) # redirect stdout and stderr to logger RedirectLogging.redirect(logger, stdout=True, stderr=True) logger.info(forecastList[0]) site = str(site) databaseID = str(databaseID) username = str(username) startTime = time.time() logger.info("Text Formatter Starting") try: forecasts = runFormatter(databaseID=databaseID, site=site, forecastList=forecastList, testMode=testMode, cmdLineVarDict=cmdLineVarDict, vtecMode=vtecMode, username=username, dataMgr=dataMgr, drtTime=drtTime, vtecActiveTable=vtecActiveTable) except: logger.exception("Error generating text product") raise elapsedTime = (time.time() - startTime)*1000 logger.info("Text Formatter Finished, took: %d ms",elapsedTime) RedirectLogging.restore() return forecasts def getPid(forecast): # taken from ProductParser.py import re sl = r'^' # start of line el = r'\s*?\n' # end of line id3 = r'[A-Za-z]{3}' # 3 charater word empty = r'^\s*' + el # empty line wmoid = r'(?P<wmoid>[A-Z]{4}\d{2})' # wmoid fsid = r'(?P<fsid>[A-Z]{4})' # full station id pit = r'(?P<pit>\d{6})' # product issuance time UTC ff = r'(?P<funnyfield> ' + id3 + ')?' # "funny" field # CI block ci_start = sl + wmoid + ' ' + fsid + ' ' + pit + ff + el awipsid = r'(?P<pil>(?P<cat>[A-Z0-9]{3})(?P<lid>[A-Z0-9]{1,3}))' + el ci_block = r'(?P<ciblock>' + ci_start + awipsid + '\n?)' ci_re = re.compile(ci_block) pid = None m = ci_re.search(forecast) if m is not None: pid = m.group('cat') return pid def runFormatter(databaseID, site, forecastList, cmdLineVarDict, vtecMode, username, dataMgr, serverFile=None, editAreas=[], timeRanges=[], timePeriod=None, drtTime=None, vtecActiveTable='active', testMode=0, experimentalMode=0, serverOutputFile=None, startTime=None, endTime=None, language=None, outputFile=None, appendFile=None ): if cmdLineVarDict: exec "cmdLineVarDict = " + cmdLineVarDict else: cmdLineVarDict = {} # Set default Forecast Type if len(forecastList) == 0: usage() logger.error("ForecastList [-t] is empty or missing") return # Can't have both T and E modes if testMode and experimentalMode: usage() logger.error("Can't have both -T and -E switches") return if drtTime: import offsetTime offsetTime.setDrtOffset(drtTime) # Create Time Range useRawTR = 0 if startTime is not None and endTime is not None: start = decodeTimeStruct(startTime) end = decodeTimeStruct(endTime) timeRange = TimeRange.TimeRange(start, end) # Set so this time range will override all others useRawTR = 1 else: timeRange = None # Handle the VTEC modes if vtecMode is not None and vtecMode not in ['X','O','T','E']: usage() logger.error("-v vtecMode must be ['X', 'O', 'T', 'E']") sys.exit(1) #force VTEC mode to "T" if in TEST mode and another vtecCode is specified if testMode and vtecMode is not None: vtecMode = "T" #force VTEC mode to "E" if in EXPERIMENTAL mode and another vtecCode #is specified elif experimentalMode and vtecMode is not None: vtecMode = "E" #force into TEST mode, if vtec code is 'T' if vtecMode == "T": testMode = 1 experimentalMode = 0 elif vtecMode == "E": experimentalMode = 1 testMode = 0 # Create an ifpClient ifpClient = PyFPClient(VizApp.getWsId(), site) global GridLoc GridLoc = ifpClient.getDBGridLocation() #importer = TextIFPImporter(ifpClient) #importer.install() import Utility import ForecastNarrative import ForecastTable import Analysis site = str(ifpClient.getSiteID()[0]) # Create dictionary of arguments argDict = { #"host" : host, #"port" : port, "databaseID": databaseID, "site" : site, "cmdLineVarDict": cmdLineVarDict, "serverFile": serverFile, "editAreas": editAreas, "timeRanges": timeRanges, "timeRange": timeRange, "timePeriod": timePeriod, "useRawTR": useRawTR, "vtecMode": vtecMode, "vtecActiveTable": vtecActiveTable, "testMode": testMode, "experimentalMode": experimentalMode, "serverOutputFile": serverOutputFile, } # Handle command line switches for variables that can be # set elsewhere i.e. in the command line varDict OR the # product definition section. # If there was a command line switch for these items, # make an entry in argDict. Otherwise, do not. for item in ["language", "outputFile", "appendFile"]: exec "if " + item + " is not None: argDict['" + item + "'] = " + item logger.info("Arguments: " + str(argDict)) argDict["ifpClient"] = ifpClient argDict["utility"] = Utility.Utility(None, None, ifpClient) #argDict["AFPS"] = AFPS #argDict["AFPSSup"] = AFPSSup argDict["Analysis"] = Analysis argDict["ForecastNarrative"] = ForecastNarrative argDict["ForecastTable"] = ForecastTable # get product creation time to the minute - almost all fmtrs use this argDict['creationTime'] = int(time.time()/60)*60.0 # Set the Site Time Zone tz = str(ifpClient.getSiteTimeZone()) os.environ['TZ'] = tz time.tzset() # Create the formatter formatter = TextFormatter.TextFormatter(dataMgr, ifpClient) # For each Forecast Type, # Create generate forecast forecasts = "" # returned value outForecasts = "" # written to output files for forecastType in forecastList: forecast = formatter.getForecast(forecastType, argDict) forecasts = forecasts + forecast # Convert data written to files to upper case if required mixedCase = False pid = getPid(forecast) if pid is None: logger.warning("Unable to determine PID: defaulting to upper case") else: from com.raytheon.uf.common.dataplugin.text.db import MixedCaseProductSupport mixedCase = MixedCaseProductSupport.isMixedCase(str(pid)) if mixedCase: outForecasts = outForecasts + forecast else: outForecasts = outForecasts + forecast.upper() logger.info("Text:\n" + str(forecasts)) try: outputFile = argDict["outputFile"] success = writeToFile(outForecasts, outputFile, "w") if success == 0: print "Couldn't open output file", outputFile logger.error("Couldn't open output file: ", outputFile) sys.exit(1) except: pass try: outputFile = argDict["serverOutputFile"] success = writeToFile(outForecasts, outputFile, "w") if success == 0: print "Couldn't open output file", outputFile logger.error("Couldn't open output file: ", outputFile) sys.exit(1) except: pass try: appendFile = argDict["appendFile"] success = writeToFile(outForecasts, appendFile, "a") if success == 0: print "Couldn't open append file", appendFile logger.error("Couldn't write to append file: ", appendFile) sys.exit(1) except: pass try: serverFile = argDict["serverFile"] writeToSite = (username == "SITE") success = writeToServerFile(outForecasts, serverFile, writeToSite) if success == 0: print "Couldn't open server output file", serverFile logger.error("Couldn't open server output file: ", serverFile) sys.exit(1) except: pass del outForecasts # Remove any lat/lon areas created temporarily #global LatLonIds #argDict["ifpClient"].deleteReferenceData(LatLonIds) # Somebody is holding onto an ifpClient and thus the C++ # object is not being destroyed. This causes the network # connection to stay open. Below is a kludge to force # the destruction of the C++ object. #del ifpClient.this # This also means that you may not import any new modules after this # point!!!!!!!!!!!!!!! return forecasts def decodeTimeStruct(timeStruct): return AbsTime.absTimeYMD(timeStruct.tm_year, timeStruct.tm_mon, timeStruct.tm_mday, timeStruct.tm_hour, timeStruct.tm_min) def writeToFile(forecasts, outputFile, mode): if outputFile: logger.info("Writing forecast to " + outputFile) try: with open(outputFile, mode) as outfile: outfile.write(forecasts) os.chmod(outputFile, 0644) except: logger.exception("Error writing forecast to "+outputFile) return 0 return 1 def writeToServerFile(forecasts, outputFile, writeToSite): if outputFile: try: if writeToSite: ctx = PATH_MGR.getContext(LocalizationType.COMMON_STATIC, LocalizationLevel.SITE) else: ctx = PATH_MGR.getContext(LocalizationType.COMMON_STATIC, LocalizationLevel.USER) filePath = PATH_MGR.SEPARATOR.join(["gfe", "text", "PRODGEN", outputFile + ".PRODGEN"]) lFile = PATH_MGR.getLocalizationFile(ctx, filePath) logger.info("Writing forecast to " + str(lFile)) from LockingFile import File with File(lFile.getFile(), "", 'w') as outfile: outfile.write(forecasts) return lFile.save() except: logger.exception("Error writing forecast to " + str(lFile)) return 0 return 1 def importModules(paths): global displayNameDict displayNameDict = {} split = paths.split(os.path.pathsep) for path in split: if not path in sys.path: sys.path.append(path) inv = [] if os.path.exists(path): inv = os.listdir(path) inv = filter(filterScripts, inv) for pid in inv: name = os.path.splitext(pid)[0] if sys.modules.has_key(name): del sys.modules[name] try: mod = __import__(name) except: logger.exception("Import Failed " + name) mod = None definition = None if mod is not None: d = mod.__dict__ #search for Definition at top-level definition = d.get('Definition', None) if definition is None: # search for definition within class name definition = d.get(name, None) if definition is None: tp = d.get('TextProduct', None) if tp is not None: #search for definition within TextProduct class definition = getattr(tp, 'Definition', None) if definition is None or type(definition) is not dict: logger.info("Formatter: No Definition Found " + name) continue dspName = getDisplayName(definition) if dspName is None or dspName == "None": continue displayNameDict[dspName] = (mod, definition) def getScripts(paths, getVtecCodes): from java.util import ArrayList from com.raytheon.uf.common.dataplugin.gfe.textproduct import ProductDefinition from com.raytheon.viz.gfe.textformatter import TextProductConfigData from com.raytheon.viz.gfe.textformatter import TextProductMetadata logger.info("TextProduct FormatterLauncher Processing....") importModules(paths) textProducts = ArrayList() for (displayName, value) in displayNameDict.items(): (module, definition) = value moduleName = module.__name__ pdef = ProductDefinition(JUtil.pyDictToJavaMap(definition)) productMetadata = TextProductMetadata(moduleName, displayName, pdef) textProducts.add(productMetadata) vtecCodes = {} if getVtecCodes: import VTECMessageType vtecCodes = VTECMessageType.VTECMessageTypeDict logger.info("TextProduct FormatterLauncher Done....") return TextProductConfigData(JUtil.pyValToJavaObj(vtecCodes), textProducts) def filterScripts(name): (filename, ext) = os.path.splitext(name) return ext == ".py" and not filename.endswith("Definition") def getDisplayName(definition): try: dspName = definition['displayName'] except: dspName = None return dspName def ppDef(definition): "pretty prints the definition to make it more readable. Returns string." s = "\n" if definition is None: return "<Definition is None>" if type(definition) == dict and len(definition.keys()): keys = definition.keys() keys.sort() #get maximum length of key maxL = 0 for k in keys: maxL = max(len(k), maxL) # output the data, formatted fmt = "%-" + `maxL` + "s" for k in keys: s = s + fmt % k + ": " + str(definition[k]) + '\n' return s else: return "<Definition not dictionary>\n" + `definition` ## TODO: Investigate if the dependency on DataManager can be removed here. ## At the moment this passes through to ValuesDialog for building special ## widgets in the DialogAreaComposite. def getVarDict(paths, dspName, dataMgr, ifpClient, issuedBy, dataSource): importModules(paths) tz = str(ifpClient.getSiteTimeZone()) os.environ['TZ'] = tz time.tzset() productDef = displayNameDict[dspName][1] productDef['database'] = dataSource vdg = VarDictGroker.VarDictGroker(displayNameDict[dspName][0], productDef, dspName, issuedBy, dataMgr) return vdg.getVarDict() def getVTECMessageType(productCategory): import VTECMessageType return VTECMessageType.getVTECMessageType(productCategory) def getTimeZones(zones, officeTZ): import AreaDictionary timezones = [] if zones is not None: for zone in zones: zdict = AreaDictionary.AreaDictionary.get(zone, {}) tzs = zdict.get("ugcTimeZone", []) if type(tzs) is str: tzs = [tzs] for tz in tzs: if tz not in timezones: timezones.append(tz) if officeTZ in timezones and officeTZ != timezones[0]: timezones.remove(officeTZ) timezones.insert(0, officeTZ) if len(timezones) == 0: timezones.append(officeTZ) return JUtil.pylistToJavaStringList(timezones) def reloadModule(moduleName): # m = __import__(moduleName) # reload(m) if sys.modules.has_key(moduleName): del sys.modules[moduleName] try: __import__(moduleName) except: logger.exception("Import Failed " + moduleName)
35.30991
112
0.618054
794d62418c269dca14a72d1f15a6c4568fd4dea5
1,775
py
Python
loginpass/github.py
authlib/socialism
635823a78a2a92cf8630f9935aebb9afcccb8656
[ "BSD-3-Clause" ]
223
2018-04-23T14:48:17.000Z
2022-03-19T09:27:36.000Z
loginpass/github.py
authlib/socialism
635823a78a2a92cf8630f9935aebb9afcccb8656
[ "BSD-3-Clause" ]
68
2018-04-22T13:55:25.000Z
2022-03-15T15:28:42.000Z
loginpass/github.py
authlib/socialism
635823a78a2a92cf8630f9935aebb9afcccb8656
[ "BSD-3-Clause" ]
86
2018-04-24T21:09:26.000Z
2022-03-17T08:55:34.000Z
""" loginpass.github ~~~~~~~~~~~~~~~~ Loginpass Backend of GitHub (https://github.com). Useful Links: - Create App: https://github.com/settings/developers - API documentation: https://developer.github.com/v3/ :copyright: (c) 2018 by Hsiaoming Yang :license: BSD, see LICENSE for more details. """ from authlib.oidc.core import UserInfo class GitHub(object): NAME = 'github' OAUTH_CONFIG = { 'api_base_url': 'https://api.github.com/', 'access_token_url': 'https://github.com/login/oauth/access_token', 'authorize_url': 'https://github.com/login/oauth/authorize', 'client_kwargs': {'scope': 'user:email'}, 'userinfo_endpoint': 'https://api.github.com/user', } def userinfo(self, **kwargs): resp = self.get(self.OAUTH_CONFIG['userinfo_endpoint'], **kwargs) data = resp.json() params = { 'sub': str(data['id']), 'name': data['name'], 'email': data.get('email'), 'preferred_username': data['login'], 'profile': data['html_url'], 'picture': data['avatar_url'], 'website': data.get('blog'), } # The email can be be None despite the scope being 'user:email'. # That is because a user can choose to make his/her email private. # If that is the case we get all the users emails regardless if private or note # and use the one he/she has marked as `primary` if params.get('email') is None: resp = self.get('user/emails', **kwargs) resp.raise_for_status() data = resp.json() params["email"] = next(email['email'] for email in data if email['primary']) return UserInfo(params)
32.87037
88
0.584789
794d6267b7ebc7454874741e5e3495050115697a
2,168
py
Python
venv/Lib/site-packages/IPython/core/release.py
kiensamsk/Fantasy-Premier-League
49802e4af7f08d914157240b49df90579dba5154
[ "MIT" ]
null
null
null
venv/Lib/site-packages/IPython/core/release.py
kiensamsk/Fantasy-Premier-League
49802e4af7f08d914157240b49df90579dba5154
[ "MIT" ]
null
null
null
venv/Lib/site-packages/IPython/core/release.py
kiensamsk/Fantasy-Premier-League
49802e4af7f08d914157240b49df90579dba5154
[ "MIT" ]
1
2022-02-10T01:22:40.000Z
2022-02-10T01:22:40.000Z
# -*- coding: utf-8 -*- """Release data for the IPython project.""" #----------------------------------------------------------------------------- # Copyright (c) 2008, IPython Development Team. # Copyright (c) 2001, Fernando Perez <fernando.perez@colorado.edu> # Copyright (c) 2001, Janko Hauser <jhauser@zscout.de> # Copyright (c) 2001, Nathaniel Gray <n8gray@caltech.edu> # # Distributed under the terms of the Modified BSD License. # # The full license is in the file COPYING.txt, distributed with this software. #----------------------------------------------------------------------------- # IPython version information. An empty _version_extra corresponds to a full # release. 'dev' as a _version_extra string means this is a development # version _version_major = 8 _version_minor = 0 _version_patch = 0 _version_extra = ".dev" # _version_extra = "rc1" _version_extra = "" # Uncomment this for full releases # Construct full version string from these. _ver = [_version_major, _version_minor, _version_patch] __version__ = '.'.join(map(str, _ver)) if _version_extra: __version__ = __version__ + _version_extra version = __version__ # backwards compatibility name version_info = (_version_major, _version_minor, _version_patch, _version_extra) # Change this when incrementing the kernel protocol version kernel_protocol_version_info = (5, 0) kernel_protocol_version = "%i.%i" % kernel_protocol_version_info license = 'BSD' authors = {'Fernando' : ('Fernando Perez','fperez.net@gmail.com'), 'Janko' : ('Janko Hauser','jhauser@zscout.de'), 'Nathan' : ('Nathaniel Gray','n8gray@caltech.edu'), 'Ville' : ('Ville Vainio','vivainio@gmail.com'), 'Brian' : ('Brian E Granger', 'ellisonbg@gmail.com'), 'Min' : ('Min Ragan-Kelley', 'benjaminrk@gmail.com'), 'Thomas' : ('Thomas A. Kluyver', 'takowl@gmail.com'), 'Jorgen' : ('Jorgen Stenarson', 'jorgen.stenarson@bostream.nu'), 'Matthias' : ('Matthias Bussonnier', 'bussonniermatthias@gmail.com'), } author = 'The IPython Development Team' author_email = 'ipython-dev@python.org'
39.418182
80
0.640221
794d62c7292746165bca7f304d0efd837bbc03f2
3,665
py
Python
tensorbay/opendataset/LIP/loader.py
machearn/tensorbay-python-sdk
5c96a5f4c0028c7bec0764f2d0142b29597ec3a9
[ "MIT" ]
73
2021-02-24T12:23:26.000Z
2022-03-12T13:00:31.000Z
tensorbay/opendataset/LIP/loader.py
machearn/tensorbay-python-sdk
5c96a5f4c0028c7bec0764f2d0142b29597ec3a9
[ "MIT" ]
681
2021-02-25T07:34:17.000Z
2022-03-25T07:08:23.000Z
tensorbay/opendataset/LIP/loader.py
machearn/tensorbay-python-sdk
5c96a5f4c0028c7bec0764f2d0142b29597ec3a9
[ "MIT" ]
35
2021-02-24T12:00:45.000Z
2022-03-30T06:43:13.000Z
#!/usr/bin/env python3 # # Copyright 2021 Graviti. Licensed under MIT License. # # pylint: disable=invalid-name """Dataloader of LIP dataset.""" import csv import os from itertools import islice from typing import List from tensorbay.dataset import Data, Dataset from tensorbay.geometry import Keypoint2D from tensorbay.label import LabeledKeypoints2D, SemanticMask from tensorbay.utility import chunked DATASET_NAME = "LIP" _SEGMENT_NAMES = ("train", "val", "test") def LIP(path: str) -> Dataset: """`LIP <https://github.com/Engineering-Course/LIP_SSL>`_ dataset. The file structure should be like:: <path> Testing_images/ testing_images/ 315_462476.jpg ... test_id.txt TrainVal_images/ TrainVal_images/ train_images/ 77_471474.jpg ... val_images/ 36_453991.jpg ... train_id.txt val_id.txt TrainVal_parsing_annotations/ TrainVal_parsing_annotations/ train_segmentations/ 77_471474.png ... val_segmentations/ 36_453991.png ... TrainVal_pose_annotations/ lip_train_set.csv lip_val_set.csv Arguments: path: The root directory of the dataset. Returns: Loaded `~tensorbay.dataset.dataset.Dataset` instance. """ root_path = os.path.abspath(os.path.expanduser(path)) test_path = os.path.join(root_path, "Testing_images") trainval_image_path = os.path.join(root_path, "TrainVal_images", "TrainVal_images") trainval_parsing_path = os.path.join( root_path, "TrainVal_parsing_annotations", "TrainVal_parsing_annotations" ) pose_path = os.path.join(root_path, "TrainVal_pose_annotations") dataset = Dataset(DATASET_NAME) dataset.load_catalog(os.path.join(os.path.dirname(__file__), "catalog.json")) for segment_name in _SEGMENT_NAMES: segment = dataset.create_segment(segment_name) if segment_name == "test": image_path = os.path.join(test_path, "testing_images") with open(os.path.join(test_path, "test_id.txt"), encoding="utf-8") as fp: for stem in fp: segment.append(Data(os.path.join(image_path, f"{stem.rstrip()}.jpg"))) else: image_path = os.path.join(trainval_image_path, f"{segment_name}_images") parsing_path = os.path.join(trainval_parsing_path, f"{segment_name}_segmentations") with open( os.path.join(pose_path, f"lip_{segment_name}_set.csv"), encoding="utf-8" ) as csvfile: for keypoints_info in csv.reader(csvfile): segment.append(_get_data(keypoints_info, image_path, parsing_path)) return dataset def _get_data(keypoints_info: List[str], image_path: str, parsing_path: str) -> Data: stem = os.path.splitext(keypoints_info[0])[0] data = Data(os.path.join(image_path, f"{stem}.jpg")) label = data.label label.semantic_mask = SemanticMask(os.path.join(parsing_path, f"{stem}.png")) keypoints = LabeledKeypoints2D() for x, y, v in chunked(islice(keypoints_info, 1, None), 3): keypoints.append( Keypoint2D(float(x), float(y), 1 - int(v)) if x.isnumeric() else Keypoint2D(0, 0, 0) ) label.keypoints2d = [keypoints] return data
35.240385
96
0.602729
794d63aee140a1aa354ec723e28a3a9d534624b8
414
py
Python
src/pathlibext/rmtree.py
gpcimino/pathlibext
10c3d102a5d8a074a9a2f5cdb0356d0187b3eada
[ "MIT" ]
2
2021-12-13T21:32:50.000Z
2021-12-23T11:17:06.000Z
src/pathlibext/rmtree.py
gpcimino/pathlibext
10c3d102a5d8a074a9a2f5cdb0356d0187b3eada
[ "MIT" ]
null
null
null
src/pathlibext/rmtree.py
gpcimino/pathlibext
10c3d102a5d8a074a9a2f5cdb0356d0187b3eada
[ "MIT" ]
null
null
null
import shutil def _rmtree(self, ignore_errors: bool = False, onerror=None): """Delete an entire directory tree. This is a wrapper around shutil.rmtree""" if not self.is_dir(): raise ValueError("rmtree called on Path that is not a directory") if not self.exists(): raise ValueError(f"File system path {self} doesn't exists") return shutil.rmtree(str(self), ignore_errors, onerror)
37.636364
81
0.702899
794d6436b61ab02e0d8f6de8d32f421a23a8a200
11,749
py
Python
Lib/test/test_py_compile.py
oleksandr-pavlyk/cpython
eb002dbe0da9622245a355db5f0cd5aa2fc70b40
[ "0BSD" ]
5
2021-12-03T23:11:53.000Z
2022-01-08T21:02:50.000Z
Lib/test/test_py_compile.py
dalakatt/cpython
2f49b97cc5426087b46515254b9a97a22ee8c807
[ "0BSD" ]
8
2022-01-07T11:31:11.000Z
2022-03-04T00:07:16.000Z
Lib/test/test_py_compile.py
dalakatt/cpython
2f49b97cc5426087b46515254b9a97a22ee8c807
[ "0BSD" ]
3
2017-10-18T09:35:14.000Z
2018-09-09T16:40:13.000Z
import functools import importlib.util import os import py_compile import shutil import stat import subprocess import sys import tempfile import unittest from test import support from test.support import os_helper, script_helper def without_source_date_epoch(fxn): """Runs function with SOURCE_DATE_EPOCH unset.""" @functools.wraps(fxn) def wrapper(*args, **kwargs): with os_helper.EnvironmentVarGuard() as env: env.unset('SOURCE_DATE_EPOCH') return fxn(*args, **kwargs) return wrapper def with_source_date_epoch(fxn): """Runs function with SOURCE_DATE_EPOCH set.""" @functools.wraps(fxn) def wrapper(*args, **kwargs): with os_helper.EnvironmentVarGuard() as env: env['SOURCE_DATE_EPOCH'] = '123456789' return fxn(*args, **kwargs) return wrapper # Run tests with SOURCE_DATE_EPOCH set or unset explicitly. class SourceDateEpochTestMeta(type(unittest.TestCase)): def __new__(mcls, name, bases, dct, *, source_date_epoch): cls = super().__new__(mcls, name, bases, dct) for attr in dir(cls): if attr.startswith('test_'): meth = getattr(cls, attr) if source_date_epoch: wrapper = with_source_date_epoch(meth) else: wrapper = without_source_date_epoch(meth) setattr(cls, attr, wrapper) return cls class PyCompileTestsBase: def setUp(self): self.directory = tempfile.mkdtemp(dir=os.getcwd()) self.source_path = os.path.join(self.directory, '_test.py') self.pyc_path = self.source_path + 'c' self.cache_path = importlib.util.cache_from_source(self.source_path) self.cwd_drive = os.path.splitdrive(os.getcwd())[0] # In these tests we compute relative paths. When using Windows, the # current working directory path and the 'self.source_path' might be # on different drives. Therefore we need to switch to the drive where # the temporary source file lives. drive = os.path.splitdrive(self.source_path)[0] if drive: os.chdir(drive) with open(self.source_path, 'w') as file: file.write('x = 123\n') def tearDown(self): shutil.rmtree(self.directory) if self.cwd_drive: os.chdir(self.cwd_drive) def test_absolute_path(self): py_compile.compile(self.source_path, self.pyc_path) self.assertTrue(os.path.exists(self.pyc_path)) self.assertFalse(os.path.exists(self.cache_path)) def test_do_not_overwrite_symlinks(self): # In the face of a cfile argument being a symlink, bail out. # Issue #17222 try: os.symlink(self.pyc_path + '.actual', self.pyc_path) except (NotImplementedError, OSError): self.skipTest('need to be able to create a symlink for a file') else: assert os.path.islink(self.pyc_path) with self.assertRaises(FileExistsError): py_compile.compile(self.source_path, self.pyc_path) @unittest.skipIf(not os.path.exists(os.devnull) or os.path.isfile(os.devnull), 'requires os.devnull and for it to be a non-regular file') def test_do_not_overwrite_nonregular_files(self): # In the face of a cfile argument being a non-regular file, bail out. # Issue #17222 with self.assertRaises(FileExistsError): py_compile.compile(self.source_path, os.devnull) def test_cache_path(self): py_compile.compile(self.source_path) self.assertTrue(os.path.exists(self.cache_path)) def test_cwd(self): with os_helper.change_cwd(self.directory): py_compile.compile(os.path.basename(self.source_path), os.path.basename(self.pyc_path)) self.assertTrue(os.path.exists(self.pyc_path)) self.assertFalse(os.path.exists(self.cache_path)) def test_relative_path(self): py_compile.compile(os.path.relpath(self.source_path), os.path.relpath(self.pyc_path)) self.assertTrue(os.path.exists(self.pyc_path)) self.assertFalse(os.path.exists(self.cache_path)) @unittest.skipIf(hasattr(os, 'geteuid') and os.geteuid() == 0, 'non-root user required') @unittest.skipIf(os.name == 'nt', 'cannot control directory permissions on Windows') def test_exceptions_propagate(self): # Make sure that exceptions raised thanks to issues with writing # bytecode. # http://bugs.python.org/issue17244 mode = os.stat(self.directory) os.chmod(self.directory, stat.S_IREAD) try: with self.assertRaises(IOError): py_compile.compile(self.source_path, self.pyc_path) finally: os.chmod(self.directory, mode.st_mode) def test_bad_coding(self): bad_coding = os.path.join(os.path.dirname(__file__), 'bad_coding2.py') with support.captured_stderr(): self.assertIsNone(py_compile.compile(bad_coding, doraise=False)) self.assertFalse(os.path.exists( importlib.util.cache_from_source(bad_coding))) def test_source_date_epoch(self): py_compile.compile(self.source_path, self.pyc_path) self.assertTrue(os.path.exists(self.pyc_path)) self.assertFalse(os.path.exists(self.cache_path)) with open(self.pyc_path, 'rb') as fp: flags = importlib._bootstrap_external._classify_pyc( fp.read(), 'test', {}) if os.environ.get('SOURCE_DATE_EPOCH'): expected_flags = 0b11 else: expected_flags = 0b00 self.assertEqual(flags, expected_flags) @unittest.skipIf(sys.flags.optimize > 0, 'test does not work with -O') def test_double_dot_no_clobber(self): # http://bugs.python.org/issue22966 # py_compile foo.bar.py -> __pycache__/foo.cpython-34.pyc weird_path = os.path.join(self.directory, 'foo.bar.py') cache_path = importlib.util.cache_from_source(weird_path) pyc_path = weird_path + 'c' head, tail = os.path.split(cache_path) penultimate_tail = os.path.basename(head) self.assertEqual( os.path.join(penultimate_tail, tail), os.path.join( '__pycache__', 'foo.bar.{}.pyc'.format(sys.implementation.cache_tag))) with open(weird_path, 'w') as file: file.write('x = 123\n') py_compile.compile(weird_path) self.assertTrue(os.path.exists(cache_path)) self.assertFalse(os.path.exists(pyc_path)) def test_optimization_path(self): # Specifying optimized bytecode should lead to a path reflecting that. self.assertIn('opt-2', py_compile.compile(self.source_path, optimize=2)) def test_invalidation_mode(self): py_compile.compile( self.source_path, invalidation_mode=py_compile.PycInvalidationMode.CHECKED_HASH, ) with open(self.cache_path, 'rb') as fp: flags = importlib._bootstrap_external._classify_pyc( fp.read(), 'test', {}) self.assertEqual(flags, 0b11) py_compile.compile( self.source_path, invalidation_mode=py_compile.PycInvalidationMode.UNCHECKED_HASH, ) with open(self.cache_path, 'rb') as fp: flags = importlib._bootstrap_external._classify_pyc( fp.read(), 'test', {}) self.assertEqual(flags, 0b1) def test_quiet(self): bad_coding = os.path.join(os.path.dirname(__file__), 'bad_coding2.py') with support.captured_stderr() as stderr: self.assertIsNone(py_compile.compile(bad_coding, doraise=False, quiet=2)) self.assertIsNone(py_compile.compile(bad_coding, doraise=True, quiet=2)) self.assertEqual(stderr.getvalue(), '') with self.assertRaises(py_compile.PyCompileError): py_compile.compile(bad_coding, doraise=True, quiet=1) class PyCompileTestsWithSourceEpoch(PyCompileTestsBase, unittest.TestCase, metaclass=SourceDateEpochTestMeta, source_date_epoch=True): pass class PyCompileTestsWithoutSourceEpoch(PyCompileTestsBase, unittest.TestCase, metaclass=SourceDateEpochTestMeta, source_date_epoch=False): pass class PyCompileCLITestCase(unittest.TestCase): def setUp(self): self.directory = tempfile.mkdtemp() self.source_path = os.path.join(self.directory, '_test.py') self.cache_path = importlib.util.cache_from_source(self.source_path) with open(self.source_path, 'w') as file: file.write('x = 123\n') def tearDown(self): os_helper.rmtree(self.directory) @support.requires_subprocess() def pycompilecmd(self, *args, **kwargs): # assert_python_* helpers don't return proc object. We'll just use # subprocess.run() instead of spawn_python() and its friends to test # stdin support of the CLI. if args and args[0] == '-' and 'input' in kwargs: return subprocess.run([sys.executable, '-m', 'py_compile', '-'], input=kwargs['input'].encode(), capture_output=True) return script_helper.assert_python_ok('-m', 'py_compile', *args, **kwargs) def pycompilecmd_failure(self, *args): return script_helper.assert_python_failure('-m', 'py_compile', *args) def test_stdin(self): result = self.pycompilecmd('-', input=self.source_path) self.assertEqual(result.returncode, 0) self.assertEqual(result.stdout, b'') self.assertEqual(result.stderr, b'') self.assertTrue(os.path.exists(self.cache_path)) def test_with_files(self): rc, stdout, stderr = self.pycompilecmd(self.source_path, self.source_path) self.assertEqual(rc, 0) self.assertEqual(stdout, b'') self.assertEqual(stderr, b'') self.assertTrue(os.path.exists(self.cache_path)) def test_bad_syntax(self): bad_syntax = os.path.join(os.path.dirname(__file__), 'badsyntax_3131.py') rc, stdout, stderr = self.pycompilecmd_failure(bad_syntax) self.assertEqual(rc, 1) self.assertEqual(stdout, b'') self.assertIn(b'SyntaxError', stderr) def test_bad_syntax_with_quiet(self): bad_syntax = os.path.join(os.path.dirname(__file__), 'badsyntax_3131.py') rc, stdout, stderr = self.pycompilecmd_failure('-q', bad_syntax) self.assertEqual(rc, 1) self.assertEqual(stdout, b'') self.assertEqual(stderr, b'') def test_file_not_exists(self): should_not_exists = os.path.join(os.path.dirname(__file__), 'should_not_exists.py') rc, stdout, stderr = self.pycompilecmd_failure(self.source_path, should_not_exists) self.assertEqual(rc, 1) self.assertEqual(stdout, b'') self.assertIn(b'no such file or directory', stderr.lower()) def test_file_not_exists_with_quiet(self): should_not_exists = os.path.join(os.path.dirname(__file__), 'should_not_exists.py') rc, stdout, stderr = self.pycompilecmd_failure('-q', self.source_path, should_not_exists) self.assertEqual(rc, 1) self.assertEqual(stdout, b'') self.assertEqual(stderr, b'') if __name__ == "__main__": unittest.main()
40.236301
97
0.637671
794d644c908b03b1dad35fbaacbc5b79589a6f67
11,307
py
Python
report.py
mattmerrifield/fake-lab-reports
1785e8ae5cf6ebfce46931e73a6bf7b147bdb4af
[ "MIT" ]
null
null
null
report.py
mattmerrifield/fake-lab-reports
1785e8ae5cf6ebfce46931e73a6bf7b147bdb4af
[ "MIT" ]
null
null
null
report.py
mattmerrifield/fake-lab-reports
1785e8ae5cf6ebfce46931e73a6bf7b147bdb4af
[ "MIT" ]
null
null
null
import enum import itertools import platform import random import datetime as dt import statistics import sys from dataclasses import dataclass, asdict, field from pathlib import Path from typing import List, Optional import pdfkit from jinja2 import Template if platform.system() == "Windows": # ugh. Sorry. I need a better OS on this box, but this is a quick dirty hack path_wkhtmltopdf = r'C:\Program Files\wkhtmltopdf\bin\wkhtmltopdf.exe' PDF_CONFIG = pdfkit.configuration(wkhtmltopdf=path_wkhtmltopdf) else: PDF_CONFIG = pdfkit.configuration() TEMPLATES = Path(__file__).parent with open(TEMPLATES / "template.jinja2") as fh: template = Template(source=fh.read()) # The probability any given test will fall within the "normal" range P_PASS = 0.8 # Some numbers are formatted as a single decimal ONE_DECIMAL = "{:0.1f}" @dataclass class Between: """ The normal range of a test result. """ low: float high: float fmt_precision: str = "{:0.0f}" fmt_result: str = "{}" # extra stuff on the report? fmt_range: str = "{low} - {high}" BELOW = -1 OK = 0 ABOVE = 1 def __str__(self): """ String representation of the range itself """ high, low = self.fmt_precision.format(self.high), self.fmt_precision.format(self.low) return self.fmt_range.format(high=high, low=low) def sample(self, rand_seed, p_pass) -> 'Sample': """ Use a specific seed to deterministically generate a random-looking result within (or slightly out of) the expected high/low range """ # Bad assumption: average Americans have metabolic panel values in the middle of the range. Haha what a joke. mean = (self.high + self.low) / 2 # Math problem: What standard deviation would give us an out-of-range value `failure_pct` of the time? # Work backwards from z-score P values, and the fact that 1-sigma is 68%. # TODO: implement bias by messing with this distribution function. dist = statistics.NormalDist(0, 1) z_top = dist.inv_cdf((1 + p_pass)/2) # He who controls the spice controls the universe. # By spice I mean psuedo-random number generator seed. z_sample = dist.samples(1, seed=rand_seed)[0] print(f"{z_sample}/{z_top}") allowed_deviation = abs(mean - self.high) actual_deviation = z_sample * allowed_deviation / z_top val = mean + actual_deviation return Sample( range=self, # Contains formatting directives, and in/out of bounds info val=val, ) def check_bounds(self, val: float) -> str: out_of_range = self.contains(val) if out_of_range == Between.BELOW: return f"L" if out_of_range == Between.ABOVE: return f"H" return "" def contains(self, value): # The value is called out with a prefix if it's too high or too low if self.low < value < self.high: return self.OK if value < self.low: return Between.BELOW if self.high < value: return Between.ABOVE @dataclass class LTHigh(Between): """ Expect below. Must provide low anyway, so we can generate a number """ fmt_range: str = "<{high}" def contains(self, value): if self.high < value: return Between.ABOVE return Between.OK @dataclass class GTLow(Between): """ Expect above. Must provide high anyway, so we can generate a number """ fmt_range: str = ">{low}" def contains(self, value): if value < self.low: return Between.BELOW return Between.OK @dataclass class Sample: """ The result of sampling a range, formatted according to that range's conventions. """ range: Between val: float # pre-formatted for precision @property def value(self): precision = self.range.fmt_precision.format(self.val) final = self.range.fmt_result.format(precision) return final @property def ok(self): return self.range.check_bounds(self.val) @dataclass class Test: """ Quantitative description of a diagnostic test to run, including name, expected range, units, etc. """ # Parameters to generate a test result. name: str range: Between units: str def sample(self, rand_seed, p_pass) -> 'Result': """ Psuedo-random result generator """ return Result( test=self, result=self.range.sample(rand_seed, p_pass) ) @dataclass class Result: """ The sampled result of a test """ test: Test result: Sample class GenList: """ Data descriptor to get random data from bunch of test data generators Alsa a data attribute to generate a Result """ def __init__(self, *generators): self.tests = generators def __get__(self, instance: 'LabReport', owner) -> List['Result']: """ Render a result based on the lab report's patient number. Should be random-looking, but deterministic based on the patient number. Result includes test meta-data like expected reference range, test name, etc. It's ready to render on a template. """ # Use the patient's ID number as a random seed. Same person, same results. Every time. # Good point of argument for skeptical patients. # But don't use the literal same seed value for every sample -- use a deterministic array of them, so # they're each in or out of range independently of one another not_rand = random.Random(instance.patient_number) results: List[Result] = [t.sample(not_rand.randint(0, sys.maxsize), instance.p_pass) for t in self.tests] for r in results: if val := instance.override_samples.get(r.test.name, None): r.result.val = val return results @dataclass class LabReport: """ Fake data for a fake lab report """ # Configurable report parameters patient_number: int collected_at: dt.datetime has_disease: bool p_pass: float = P_PASS override_samples: dict = field(default_factory=dict) # Data descriptor for making a list of fake test results. # Use it like a property, e.g. `results = self.metabolic_panel` metabolic_panel = GenList( Test("Sodium", Between(132, 146), "mM"), Test("Potassium", Between(3.4, 5.4, fmt_precision=ONE_DECIMAL), "mM", ), Test("Chloride", Between(99, 109), "mM"), Test("Bicarbonate", Between(19, 33), "mM"), Test("Glucose", Between(73, 105, fmt_result="{}**"), "mg/dL"), Test("Bun", Between(6, 24), "mg/dL", ), Test("Creatine", Between(0.5, 1.2, fmt_precision=ONE_DECIMAL), "mg/dL", ), Test("Calcium", Between(8.3, 10.6, fmt_precision=ONE_DECIMAL), "g/dL", ), Test("Protein, Total", Between(6, 8, fmt_precision=ONE_DECIMAL), "g/dL", ), Test("Albumin", Between(3.5, 5.1, fmt_precision=ONE_DECIMAL), "g/dL", ), Test("Bilirubin, Total", Between(0.3, 1.4, fmt_precision=ONE_DECIMAL), "mg/dl", ), Test("ALP", Between(44, 135), "U/L", ), Test("ALT", Between(7.9, 40.9, fmt_precision=ONE_DECIMAL), "U/L"), Test("AST", Between(0, 35), "U/L"), ) lipid_panel = GenList( Test("Cholesterol, Total", LTHigh(100, 240), "mg/dL"), Test("Triglycerides", LTHigh(100, 200), "mg/dL"), Test("HDL Cholesterol", GTLow(40, 90), "mg/dL"), Test("LDL Cholesterol", LTHigh(85, 130), "mg/dL"), ) def force_result(self, name, value): """ Force a specific test, by name, to have a specific value """ # Ew, gross, just us a hash map why don't you (because this is a hack job) for t in itertools.chain(self.metabolic_panel, self.lipid_panel): if t.test.name == name: t.result.value = value return raise KeyError(f"Test {name} not found") def as_html(self) -> str: """ use the above template to get an HTML report """ ctx = asdict(self) ctx['metabolic_panel'] = self.metabolic_panel ctx['lipid_panel'] = self.lipid_panel ctx['has_disease'] = self.has_disease # PDF requires inline style sheets, which we inject via templating with open(TEMPLATES / "style.css") as fh: ctx['style'] = fh.read() with open(TEMPLATES / "normalize.css") as fh: ctx['normalize'] = fh.read() print(ctx) return template.render(ctx) def save_html(self, filename): with open(filename, 'w') as fh: fh.write(self.as_html()) def save_pdf(self, filename): """ Generate psuedorandom results, and render them as a PDF """ pdfkit.from_string( self.as_html(), filename, configuration=PDF_CONFIG, options={ 'encoding': "UTF-8", 'print-media-type': '', 'page-size': 'A4', 'zoom': '1.1' } ) def generate(patient_number, output_folder, has_disease, p_pass): r = LabReport(patient_number=patient_number, collected_at=dt.datetime.now(), has_disease=has_disease, p_pass=p_pass) out = Path(output_folder) / f"{patient_number}.pdf" r.save_pdf(out) def gen_samples(): output_folder = Path(__file__).parent / "okay" output_folder.mkdir(exist_ok=True) START_AT = 15900 NUM_REPORTS = 60 # Arbitrary range of patient numbers; all healthy for i, patient_number in enumerate(range(START_AT, START_AT + NUM_REPORTS)): r = LabReport(patient_number=patient_number, collected_at=dt.datetime.now(), has_disease=False, p_pass=P_PASS) out = Path(output_folder) / f"report-{i}.pdf" # r.save_pdf(out) # One bad patient, with ID 10 BAD_ID = 10 output_folder = Path(__file__).parent / "bad" output_folder.mkdir(exist_ok=True) r = LabReport(patient_number=START_AT + BAD_ID, collected_at=dt.datetime.now(), has_disease=True, p_pass=P_PASS/2) # Some specific hard-coded changes for this one r.override_samples = { 'Sodium': 162, 'Potassium': 6.8, "Bicarbonate": 40, "Chloride": 118, 'Glucose': 152, 'Bun': 41, 'Creatine': 1.44, 'Calcium': 15, 'Protein, Total': 6.6, 'Albumin': 33, 'Bilirubin, Total': 2.4, "ALP": 188.8, 'ALT': 31, 'AST': 93, "Cholesterol, Total": 259, "Triglycerides": 213, "HDL Cholesterol": 22, "LDL Cholesterol": 158, } out = Path(output_folder) / f"report-{BAD_ID}.pdf" r.save_pdf(out) if __name__ == "__main__": gen_samples() import argparse, sys parser = argparse.ArgumentParser() parser.add_argument('patient_number', action='store') parser.add_argument('--output_folder', '-o', type=str, default='.') parser.add_argument('--has_disease', '-d', action='store_true') args = parser.parse_args(sys.argv[1:]) generate(args.patient_number, args.output_folder, args.has_disease, P_PASS/2 if args.has_disease else P_PASS)
31.940678
120
0.620147
794d648d163162df999cd4128ddb727a5431c5d8
892
py
Python
Lib/TrainProcess_Folder.py
Jack-GVDL/TrainUtility
a326b3d2504f980216d17a4c843bac1f1a8dfec6
[ "MIT" ]
null
null
null
Lib/TrainProcess_Folder.py
Jack-GVDL/TrainUtility
a326b3d2504f980216d17a4c843bac1f1a8dfec6
[ "MIT" ]
null
null
null
Lib/TrainProcess_Folder.py
Jack-GVDL/TrainUtility
a326b3d2504f980216d17a4c843bac1f1a8dfec6
[ "MIT" ]
null
null
null
import os from typing import * from .TrainProcess import TrainProcess from .ModelInfo import ModelInfo class TrainProcess_Folder(TrainProcess): def __init__(self): super().__init__() # data self.name = "Folder" # operation # ... def __del__(self): return # Operation # operation def execute(self, stage: int, info: ModelInfo, data: Dict) -> None: path = info.save_path folder = info.save_folder # check if folder exist or not # if not exist, then create one folder_path = os.path.join(path, folder) if not os.path.isdir(folder_path): os.mkdir(folder_path) # info def getPrintContent(self, stage: int, info: ModelInfo) -> str: return self._getContent_(info) def getLogContent(self, stage: int, info: ModelInfo) -> str: return self._getContent_(info) # Protected def _getContent_(self, info: ModelInfo) -> str: return "Operation: folder"
20.744186
68
0.709641
794d6615aa767a04ebefc9c97e1d7d0631d57169
38,312
py
Python
tensor2tensor/data_generators/image.py
rickyHong/tensor2tensor-repl
b6a57e71b0bb53b35e0f9a1ba8ae75c5169f9af5
[ "Apache-2.0" ]
null
null
null
tensor2tensor/data_generators/image.py
rickyHong/tensor2tensor-repl
b6a57e71b0bb53b35e0f9a1ba8ae75c5169f9af5
[ "Apache-2.0" ]
null
null
null
tensor2tensor/data_generators/image.py
rickyHong/tensor2tensor-repl
b6a57e71b0bb53b35e0f9a1ba8ae75c5169f9af5
[ "Apache-2.0" ]
null
null
null
# coding=utf-8 # Copyright 2017 The Tensor2Tensor Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Data generators for image data-sets.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import gzip import io import json import os import random import struct import tarfile import zipfile # Dependency imports import numpy as np from six.moves import cPickle from six.moves import xrange # pylint: disable=redefined-builtin from six.moves import zip # pylint: disable=redefined-builtin from tensor2tensor.data_generators import generator_utils from tensor2tensor.data_generators import problem from tensor2tensor.data_generators import text_encoder from tensor2tensor.layers import common_layers from tensor2tensor.utils import registry import tensorflow as tf from tensorflow.python.eager import context def resize_by_area(img, size): """image resize function used by quite a few image problems.""" return tf.to_int64( tf.image.resize_images(img, [size, size], tf.image.ResizeMethod.AREA)) class ImageProblem(problem.Problem): def example_reading_spec(self, label_repr=None): data_fields = { "image/encoded": tf.FixedLenFeature((), tf.string), "image/format": tf.FixedLenFeature((), tf.string), } data_items_to_decoders = { "inputs": tf.contrib.slim.tfexample_decoder.Image( image_key="image/encoded", format_key="image/format", channels=3), } return data_fields, data_items_to_decoders @registry.register_problem("image_celeba_tune") class ImageCeleba(ImageProblem): """CelebA dataset, aligned and cropped images.""" IMG_DATA = ("img_align_celeba.zip", "https://drive.google.com/uc?export=download&" "id=0B7EVK8r0v71pZjFTYXZWM3FlRnM") LANDMARKS_DATA = ("celeba_landmarks_align", "https://drive.google.com/uc?export=download&" "id=0B7EVK8r0v71pd0FJY3Blby1HUTQ") ATTR_DATA = ("celeba_attr", "https://drive.google.com/uc?export=download&" "id=0B7EVK8r0v71pblRyaVFSWGxPY0U") LANDMARK_HEADINGS = ("lefteye_x lefteye_y righteye_x righteye_y " "nose_x nose_y leftmouth_x leftmouth_y rightmouth_x " "rightmouth_y").split() ATTR_HEADINGS = ( "5_o_Clock_Shadow Arched_Eyebrows Attractive Bags_Under_Eyes Bald Bangs " "Big_Lips Big_Nose Black_Hair Blond_Hair Blurry Brown_Hair " "Bushy_Eyebrows Chubby Double_Chin Eyeglasses Goatee Gray_Hair " "Heavy_Makeup High_Cheekbones Male Mouth_Slightly_Open Mustache " "Narrow_Eyes No_Beard Oval_Face Pale_Skin Pointy_Nose Receding_Hairline " "Rosy_Cheeks Sideburns Smiling Straight_Hair Wavy_Hair Wearing_Earrings " "Wearing_Hat Wearing_Lipstick Wearing_Necklace Wearing_Necktie Young" ).split() def preprocess_example(self, example, unused_mode, unused_hparams): inputs = example["inputs"] # Remove boundaries in CelebA images. Remove 40 pixels each side # vertically and 20 pixels each side horizontally. inputs = tf.image.crop_to_bounding_box(inputs, 40, 20, 218 - 80, 178 - 40) example["inputs"] = resize_by_area(inputs, 8) example["targets"] = resize_by_area(inputs, 32) return example def hparams(self, defaults, unused_model_hparams): p = defaults p.input_modality = {"inputs": ("image:identity", 256)} p.target_modality = ("image:identity", 256) p.batch_size_multiplier = 256 p.max_expected_batch_size_per_shard = 4 p.input_space_id = 1 p.target_space_id = 1 def generator(self, tmp_dir, how_many, start_from=0): """Image generator for CELEBA dataset. Args: tmp_dir: path to temporary storage directory. how_many: how many images and labels to generate. start_from: from which image to start. Yields: A dictionary representing the images with the following fields: * image/encoded: the string encoding the image as JPEG, * image/format: the string "jpeg" representing image format, """ out_paths = [] for fname, url in [self.IMG_DATA, self.LANDMARKS_DATA, self.ATTR_DATA]: path = generator_utils.maybe_download_from_drive(tmp_dir, fname, url) out_paths.append(path) img_path, landmarks_path, attr_path = out_paths # pylint: disable=unbalanced-tuple-unpacking unzipped_folder = img_path[:-4] if not tf.gfile.Exists(unzipped_folder): zipfile.ZipFile(img_path, "r").extractall(tmp_dir) with tf.gfile.Open(landmarks_path) as f: landmarks_raw = f.read() with tf.gfile.Open(attr_path) as f: attr_raw = f.read() def process_landmarks(raw_data): landmarks = {} lines = raw_data.split("\n") headings = lines[1].strip().split() for line in lines[2:-1]: values = line.strip().split() img_name = values[0] landmark_values = [int(v) for v in values[1:]] landmarks[img_name] = landmark_values return landmarks, headings def process_attrs(raw_data): attrs = {} lines = raw_data.split("\n") headings = lines[1].strip().split() for line in lines[2:-1]: values = line.strip().split() img_name = values[0] attr_values = [int(v) for v in values[1:]] attrs[img_name] = attr_values return attrs, headings img_landmarks, _ = process_landmarks(landmarks_raw) img_attrs, _ = process_attrs(attr_raw) image_files = tf.gfile.Glob(unzipped_folder + "/*.jpg") for filename in image_files[start_from:start_from + how_many]: img_name = os.path.basename(filename) landmarks = img_landmarks[img_name] attrs = img_attrs[img_name] with tf.gfile.Open(filename, "r") as f: encoded_image_data = f.read() yield { "image/encoded": [encoded_image_data], "image/format": ["jpeg"], "attributes": attrs, "landmarks": landmarks, } @property def train_shards(self): return 100 @property def dev_shards(self): return 10 def generate_data(self, data_dir, tmp_dir, task_id=-1): generator_utils.generate_dataset_and_shuffle( self.generator(tmp_dir, 162770), # train self.training_filepaths(data_dir, self.train_shards, shuffled=False), self.generator(tmp_dir, 19867, 162770), # dev self.dev_filepaths(data_dir, self.dev_shards, shuffled=False)) @registry.register_problem class ImageFSNS(ImageProblem): """Problem spec for French Street Name recognition.""" def generate_data(self, data_dir, tmp_dir, task_id=-1): list_url = ("https://raw.githubusercontent.com/tensorflow/models/master/" "street/python/fsns_urls.txt") fsns_urls = generator_utils.maybe_download(tmp_dir, "fsns_urls.txt", list_url) fsns_files = [ f.strip() for f in open(fsns_urls, "r") if f.startswith("http://") ] for url in fsns_files: if "/train/train" in url: generator_utils.maybe_download( data_dir, "image_fsns-train" + url[-len("-00100-of-00512"):], url) elif "/validation/validation" in url: generator_utils.maybe_download( data_dir, "image_fsns-dev" + url[-len("-00100-of-00512"):], url) elif "charset" in url: generator_utils.maybe_download(data_dir, "charset_size134.txt", url) def feature_encoders(self, data_dir): # This vocab file must be present within the data directory. vocab_filename = os.path.join(data_dir, "charset_size134.txt") return { "inputs": text_encoder.ImageEncoder(), "targets": text_encoder.SubwordTextEncoder(vocab_filename) } def hparams(self, defaults, unused_model_hparams): p = defaults p.input_modality = {"inputs": (registry.Modalities.IMAGE, 256)} vocab_size = self._encoders["targets"].vocab_size p.target_modality = (registry.Modalities.SYMBOL, vocab_size) p.batch_size_multiplier = 256 p.max_expected_batch_size_per_shard = 2 p.input_space_id = problem.SpaceID.IMAGE p.target_space_id = problem.SpaceID.EN_TOK def example_reading_spec(self): label_key = "image/unpadded_label" data_fields, data_items_to_decoders = ( super(ImageFSNS, self).example_reading_spec()) data_fields[label_key] = tf.VarLenFeature(tf.int64) data_items_to_decoders[ "targets"] = tf.contrib.slim.tfexample_decoder.Tensor(label_key) return data_fields, data_items_to_decoders class Image2ClassProblem(ImageProblem): """Base class for image classification problems.""" @property def is_small(self): raise NotImplementedError() @property def num_classes(self): raise NotImplementedError() @property def train_shards(self): raise NotImplementedError() @property def dev_shards(self): return 1 @property def class_labels(self): return ["ID_%d" % i for i in range(self.num_classes)] def feature_encoders(self, data_dir): del data_dir return { "inputs": text_encoder.ImageEncoder(), "targets": text_encoder.ClassLabelEncoder(self.class_labels) } def generator(self, data_dir, tmp_dir, is_training): raise NotImplementedError() def example_reading_spec(self): label_key = "image/class/label" data_fields, data_items_to_decoders = ( super(Image2ClassProblem, self).example_reading_spec()) data_fields[label_key] = tf.FixedLenFeature((1,), tf.int64) data_items_to_decoders[ "targets"] = tf.contrib.slim.tfexample_decoder.Tensor(label_key) return data_fields, data_items_to_decoders def hparams(self, defaults, unused_model_hparams): p = defaults p.input_modality = {"inputs": (registry.Modalities.IMAGE, 256)} p.target_modality = (registry.Modalities.CLASS_LABEL, self.num_classes) p.batch_size_multiplier = 4 if self.is_small else 256 p.max_expected_batch_size_per_shard = 8 if self.is_small else 2 p.loss_multiplier = 3.0 if self.is_small else 1.0 if self._was_reversed: p.loss_multiplier = 1.0 p.input_space_id = problem.SpaceID.IMAGE p.target_space_id = problem.SpaceID.IMAGE_LABEL def generate_data(self, data_dir, tmp_dir, task_id=-1): generator_utils.generate_dataset_and_shuffle( self.generator(data_dir, tmp_dir, True), self.training_filepaths(data_dir, self.train_shards, shuffled=False), self.generator(data_dir, tmp_dir, False), self.dev_filepaths(data_dir, self.dev_shards, shuffled=False)) def imagenet_preprocess_example(example, mode, resize_size=None): """Preprocessing used for Imagenet and similar problems.""" if resize_size is None: resize_size = [299, 299] def preprocess(img): img = tf.image.resize_images(img, [360, 360]) img = common_layers.image_augmentation( tf.to_float(img) / 255., crop_size=resize_size) return tf.to_int64(img * 255.) def resize(img): return tf.to_int64(tf.image.resize_images(img, resize_size)) inputs = tf.cast(example["inputs"], tf.int64) if mode == tf.estimator.ModeKeys.TRAIN: example["inputs"] = tf.cond( # Preprocess 90% of the time. tf.less(tf.random_uniform([]), 0.9), lambda img=inputs: preprocess(img), lambda img=inputs: resize(img)) else: example["inputs"] = resize(inputs) return example @registry.register_problem class ImageImagenet(Image2ClassProblem): """Imagenet.""" @property def is_small(self): return False @property def num_classes(self): return 1000 def generate_data(self, data_dir, tmp_dir, task_id=-1): # TODO(lukaszkaiser): find a better way than printing this. print("To generate the ImageNet dataset in the proper format, follow " "instructions at https://github.com/tensorflow/models/blob/master" "/inception/README.md#getting-started") def preprocess_example(self, example, mode, _): return imagenet_preprocess_example(example, mode) @registry.register_problem class ImageImagenet224(ImageImagenet): """Imagenet rescaled to 224x224.""" def dataset_filename(self): return "image_imagenet" # Reuse Imagenet data. def generate_data(self, data_dir, tmp_dir, task_id=-1): tf.logging.warning( "Generate data for image_imagenet224 with image_imagenet") def preprocess_example(self, example, mode, _): return imagenet_preprocess_example(example, mode, resize_size=[224, 224]) @registry.register_problem class ImageImagenet32(Image2ClassProblem): """Imagenet rescaled to 32x32.""" def dataset_filename(self): return "image_imagenet" # Reuse Imagenet data. @property def is_small(self): return True # Modalities like for CIFAR. @property def num_classes(self): return 1000 def generate_data(self, data_dir, tmp_dir, task_id=-1): # TODO(lukaszkaiser): find a better way than printing this. print("To generate the ImageNet dataset in the proper format, follow " "instructions at https://github.com/tensorflow/models/blob/master" "/inception/README.md#getting-started") def preprocess_example(self, example, mode, unused_hparams): # Just resize with area. if self._was_reversed: example["inputs"] = tf.to_int64( tf.image.resize_images(example["inputs"], [32, 32], tf.image.ResizeMethod.AREA)) else: example = imagenet_preprocess_example(example, mode) example["inputs"] = tf.to_int64( tf.image.resize_images(example["inputs"], [32, 32])) return example @registry.register_problem class ImageImagenet64(Image2ClassProblem): """Imagenet rescaled to 64x64.""" def dataset_filename(self): return "image_imagenet" # Reuse Imagenet data. @property def is_small(self): return True # Modalities like for CIFAR. @property def num_classes(self): return 1000 def generate_data(self, data_dir, tmp_dir, task_id=-1): # TODO(lukaszkaiser): find a better way than printing this. print("To generate the ImageNet dataset in the proper format, follow " "instructions at https://github.com/tensorflow/models/blob/master" "/inception/README.md#getting-started") def preprocess_example(self, example, mode, unused_hparams): inputs = example["inputs"] # Just resize with area. if self._was_reversed: example["inputs"] = resize_by_area(inputs, 64) else: example = imagenet_preprocess_example(example, mode) example["inputs"] = example["inputs"] = resize_by_area(inputs, 64) return example @registry.register_problem class Img2imgImagenet(ImageProblem): """Imagenet rescaled to 8x8 for input and 32x32 for output.""" def dataset_filename(self): return "image_imagenet" # Reuse Imagenet data. def preprocess_example(self, example, unused_mode, unused_hparams): inputs = example["inputs"] # For Img2Img resize input and output images as desired. example["inputs"] = resize_by_area(inputs, 8) example["targets"] = resize_by_area(inputs, 32) return example def hparams(self, defaults, unused_model_hparams): p = defaults p.input_modality = {"inputs": ("image:identity", 256)} p.target_modality = ("image:identity", 256) p.batch_size_multiplier = 256 p.max_expected_batch_size_per_shard = 4 p.input_space_id = 1 p.target_space_id = 1 def _encoded_images(images): if context.in_eager_mode(): for image in images: yield tf.image.encode_png(image).numpy() else: (width, height, channels) = images[0].shape with tf.Graph().as_default(): image_t = tf.placeholder(dtype=tf.uint8, shape=(width, height, channels)) encoded_image_t = tf.image.encode_png(image_t) with tf.Session() as sess: for image in images: enc_string = sess.run(encoded_image_t, feed_dict={image_t: image}) yield enc_string def image_generator(images, labels): """Generator for images that takes image and labels lists and creates pngs. Args: images: list of images given as [width x height x channels] numpy arrays. labels: list of ints, same length as images. Yields: A dictionary representing the images with the following fields: * image/encoded: the string encoding the image as PNG, * image/format: the string "png" representing image format, * image/class/label: an integer representing the label, * image/height: an integer representing the height, * image/width: an integer representing the width. Every field is actually a singleton list of the corresponding type. Raises: ValueError: if images is an empty list. """ if not images: raise ValueError("Must provide some images for the generator.") width, height, _ = images[0].shape for (enc_image, label) in zip(_encoded_images(images), labels): yield { "image/encoded": [enc_image], "image/format": ["png"], "image/class/label": [int(label)], "image/height": [height], "image/width": [width] } # URLs and filenames for MNIST data. _MNIST_URL = "http://yann.lecun.com/exdb/mnist/" _MNIST_TRAIN_DATA_FILENAME = "train-images-idx3-ubyte.gz" _MNIST_TRAIN_LABELS_FILENAME = "train-labels-idx1-ubyte.gz" _MNIST_TEST_DATA_FILENAME = "t10k-images-idx3-ubyte.gz" _MNIST_TEST_LABELS_FILENAME = "t10k-labels-idx1-ubyte.gz" _MNIST_IMAGE_SIZE = 28 def _get_mnist(directory): """Download all MNIST files to directory unless they are there.""" for filename in [ _MNIST_TRAIN_DATA_FILENAME, _MNIST_TRAIN_LABELS_FILENAME, _MNIST_TEST_DATA_FILENAME, _MNIST_TEST_LABELS_FILENAME ]: generator_utils.maybe_download(directory, filename, _MNIST_URL + filename) def _extract_mnist_images(filename, num_images): """Extract images from an MNIST file into a numpy array. Args: filename: The path to an MNIST images file. num_images: The number of images in the file. Returns: A numpy array of shape [number_of_images, height, width, channels]. """ with gzip.open(filename) as bytestream: bytestream.read(16) buf = bytestream.read(_MNIST_IMAGE_SIZE * _MNIST_IMAGE_SIZE * num_images) data = np.frombuffer(buf, dtype=np.uint8) data = data.reshape(num_images, _MNIST_IMAGE_SIZE, _MNIST_IMAGE_SIZE, 1) return data def _extract_mnist_labels(filename, num_labels): """Extract labels from an MNIST file into integers. Args: filename: The path to an MNIST labels file. num_labels: The number of labels in the file. Returns: A int64 numpy array of shape [num_labels] """ with gzip.open(filename) as bytestream: bytestream.read(8) buf = bytestream.read(num_labels) labels = np.frombuffer(buf, dtype=np.uint8).astype(np.int64) return labels def mnist_common_generator(tmp_dir, training, how_many, data_filename, label_filename, start_from=0): """Image generator for MNIST. Args: tmp_dir: path to temporary storage directory. training: a Boolean; if true, we use the train set, otherwise the test set. how_many: how many images and labels to generate. data_filename: file that contains features data. label_filename: file that contains labels. start_from: from which image to start. Returns: An instance of image_generator that produces MNIST images. """ data_path = os.path.join(tmp_dir, data_filename) labels_path = os.path.join(tmp_dir, label_filename) images = _extract_mnist_images(data_path, 60000 if training else 10000) labels = _extract_mnist_labels(labels_path, 60000 if training else 10000) # Shuffle the data to make sure classes are well distributed. data = list(zip(images, labels)) random.shuffle(data) images, labels = list(zip(*data)) return image_generator(images[start_from:start_from + how_many], labels[start_from:start_from + how_many]) def mnist_generator(tmp_dir, training, how_many, start_from=0): """Image generator for MNIST. Args: tmp_dir: path to temporary storage directory. training: a Boolean; if true, we use the train set, otherwise the test set. how_many: how many images and labels to generate. start_from: from which image to start. Returns: An instance of image_generator that produces MNIST images. """ _get_mnist(tmp_dir) d = _MNIST_TRAIN_DATA_FILENAME if training else _MNIST_TEST_DATA_FILENAME l = _MNIST_TRAIN_LABELS_FILENAME if training else _MNIST_TEST_LABELS_FILENAME return mnist_common_generator(tmp_dir, training, how_many, d, l, start_from) @registry.register_problem class ImageMnistTune(Image2ClassProblem): """MNIST, tuning data.""" @property def is_small(self): return True @property def num_classes(self): return 10 @property def class_labels(self): return [str(c) for c in range(self.num_classes)] @property def train_shards(self): return 10 def generator(self, data_dir, tmp_dir, is_training): if is_training: return mnist_generator(tmp_dir, True, 55000) else: return mnist_generator(tmp_dir, True, 5000, 55000) @registry.register_problem class ImageMnist(ImageMnistTune): def generator(self, data_dir, tmp_dir, is_training): if is_training: return mnist_generator(tmp_dir, True, 60000) else: return mnist_generator(tmp_dir, False, 10000) # URLs and filenames for MNIST data. _FASHION_MNIST_URL = ("http://fashion-mnist.s3-website.eu-central-1" ".amazonaws.com/") _FASHION_MNIST_LOCAL_FILE_PREFIX = "fashion-" _FASHION_MNIST_IMAGE_SIZE = 28 def _get_fashion_mnist(directory): """Download all FashionMNIST files to directory unless they are there.""" # Fashion mnist files have the same names as MNIST. # We must choose a separate name (by adding 'fashion-' prefix) in the tmp_dir. for filename in [ _MNIST_TRAIN_DATA_FILENAME, _MNIST_TRAIN_LABELS_FILENAME, _MNIST_TEST_DATA_FILENAME, _MNIST_TEST_LABELS_FILENAME ]: generator_utils.maybe_download(directory, _FASHION_MNIST_LOCAL_FILE_PREFIX + filename, _FASHION_MNIST_URL + filename) def fashion_mnist_generator(tmp_dir, training, how_many, start_from=0): """Image generator for FashionMNIST. Args: tmp_dir: path to temporary storage directory. training: a Boolean; if true, we use the train set, otherwise the test set. how_many: how many images and labels to generate. start_from: from which image to start. Returns: An instance of image_generator that produces MNIST images. """ _get_fashion_mnist(tmp_dir) d = _FASHION_MNIST_LOCAL_FILE_PREFIX + ( _MNIST_TRAIN_DATA_FILENAME if training else _MNIST_TEST_DATA_FILENAME) l = _FASHION_MNIST_LOCAL_FILE_PREFIX + ( _MNIST_TRAIN_LABELS_FILENAME if training else _MNIST_TEST_LABELS_FILENAME) return mnist_common_generator(tmp_dir, training, how_many, d, l, start_from) @registry.register_problem class ImageFashionMnist(Image2ClassProblem): """Fashion MNIST.""" @property def is_small(self): return True @property def num_classes(self): return 10 @property def class_labels(self): return [str(c) for c in range(self.num_classes)] @property def train_shards(self): return 10 def generator(self, data_dir, tmp_dir, is_training): if is_training: return fashion_mnist_generator(tmp_dir, True, 60000) else: return fashion_mnist_generator(tmp_dir, False, 10000) # URLs and filenames for CIFAR data. _CIFAR10_URL = "https://www.cs.toronto.edu/~kriz/cifar-10-python.tar.gz" _CIFAR10_PREFIX = "cifar-10-batches-py/" _CIFAR10_TRAIN_FILES = [ "data_batch_1", "data_batch_2", "data_batch_3", "data_batch_4", "data_batch_5" ] _CIFAR10_TEST_FILES = ["test_batch"] _CIFAR10_IMAGE_SIZE = 32 def _get_cifar10(directory): """Download and extract CIFAR to directory unless it is there.""" filename = os.path.basename(_CIFAR10_URL) path = generator_utils.maybe_download(directory, filename, _CIFAR10_URL) tarfile.open(path, "r:gz").extractall(directory) def cifar10_generator(tmp_dir, training, how_many, start_from=0): """Image generator for CIFAR-10. Args: tmp_dir: path to temporary storage directory. training: a Boolean; if true, we use the train set, otherwise the test set. how_many: how many images and labels to generate. start_from: from which image to start. Returns: An instance of image_generator that produces CIFAR-10 images and labels. """ _get_cifar10(tmp_dir) data_files = _CIFAR10_TRAIN_FILES if training else _CIFAR10_TEST_FILES all_images, all_labels = [], [] for filename in data_files: path = os.path.join(tmp_dir, _CIFAR10_PREFIX, filename) with tf.gfile.Open(path, "r") as f: data = cPickle.load(f) images = data["data"] num_images = images.shape[0] images = images.reshape((num_images, 3, _CIFAR10_IMAGE_SIZE, _CIFAR10_IMAGE_SIZE)) all_images.extend([ np.squeeze(images[j]).transpose((1, 2, 0)) for j in xrange(num_images) ]) labels = data["labels"] all_labels.extend([labels[j] for j in xrange(num_images)]) return image_generator(all_images[start_from:start_from + how_many], all_labels[start_from:start_from + how_many]) @registry.register_problem class ImageCifar10Tune(ImageMnistTune): """Cifar-10 Tune.""" @property def class_labels(self): return [ "airplane", "automobile", "bird", "cat", "deer", "dog", "frog", "horse", "ship", "truck" ] def preprocess_example(self, example, mode, unused_hparams): example["inputs"].set_shape([_CIFAR10_IMAGE_SIZE, _CIFAR10_IMAGE_SIZE, 3]) if mode == tf.estimator.ModeKeys.TRAIN: example["inputs"] = common_layers.cifar_image_augmentation( example["inputs"]) example["inputs"] = tf.to_int64(example["inputs"]) return example def generator(self, data_dir, tmp_dir, is_training): if is_training: return cifar10_generator(tmp_dir, True, 48000) else: return cifar10_generator(tmp_dir, True, 2000, 48000) @registry.register_problem class ImageCifar10(ImageCifar10Tune): def generator(self, data_dir, tmp_dir, is_training): if is_training: return cifar10_generator(tmp_dir, True, 50000) else: return cifar10_generator(tmp_dir, False, 10000) @registry.register_problem class ImageCifar10Plain(ImageCifar10): def preprocess_example(self, example, mode, unused_hparams): example["inputs"].set_shape([_CIFAR10_IMAGE_SIZE, _CIFAR10_IMAGE_SIZE, 3]) example["inputs"] = tf.to_int64(example["inputs"]) return example @registry.register_problem class ImageCifar10Plain8(ImageCifar10): """CIFAR-10 rescaled to 8x8 for output: Conditional image generation.""" def dataset_filename(self): return "image_cifar10_plain" # Reuse CIFAR-10 plain data. def preprocess_example(self, example, mode, unused_hparams): example["inputs"] = resize_by_area(example["inputs"], 8) return example @registry.register_problem class Img2imgCifar10(ImageCifar10): """CIFAR-10 rescaled to 8x8 for input and 32x32 for output.""" def dataset_filename(self): return "image_cifar10_plain" # Reuse CIFAR-10 plain data. def preprocess_example(self, example, unused_mode, unused_hparams): inputs = example["inputs"] # For Img2Img resize input and output images as desired. example["inputs"] = resize_by_area(inputs, 8) example["targets"] = resize_by_area(inputs, 32) return example def hparams(self, defaults, unused_model_hparams): p = defaults p.input_modality = {"inputs": ("image:identity", 256)} p.target_modality = ("image:identity", 256) p.batch_size_multiplier = 256 p.max_expected_batch_size_per_shard = 4 p.input_space_id = 1 p.target_space_id = 1 # URLs and filenames for MSCOCO data. _MSCOCO_ROOT_URL = "http://msvocds.blob.core.windows.net/" _MSCOCO_URLS = [ "coco2014/train2014.zip", "coco2014/val2014.zip", "coco2014/test2014.zip", "annotations-1-0-3/captions_train-val2014.zip" ] _MSCOCO_TRAIN_PREFIX = "train2014" _MSCOCO_EVAL_PREFIX = "val2014" _MSCOCO_TRAIN_CAPTION_FILE = "annotations/captions_train2014.json" _MSCOCO_EVAL_CAPTION_FILE = "annotations/captions_val2014.json" def _get_mscoco(directory): """Download and extract MSCOCO datasets to directory unless it is there.""" for url in _MSCOCO_URLS: filename = os.path.basename(url) download_url = os.path.join(_MSCOCO_ROOT_URL, url) path = generator_utils.maybe_download(directory, filename, download_url) unzip_dir = os.path.join(directory, filename.strip(".zip")) if not tf.gfile.Exists(unzip_dir): zipfile.ZipFile(path, "r").extractall(directory) def mscoco_generator(data_dir, tmp_dir, training, how_many, start_from=0, eos_list=None, vocab_filename=None, vocab_size=0): """Image generator for MSCOCO captioning problem with token-wise captions. Args: data_dir: path to the data directory. tmp_dir: path to temporary storage directory. training: a Boolean; if true, we use the train set, otherwise the test set. how_many: how many images and labels to generate. start_from: from which image to start. eos_list: optional list of end of sentence tokens, otherwise use default value `1`. vocab_filename: file within `tmp_dir` to read vocabulary from. vocab_size: integer target to generate vocabulary size to. Yields: A dictionary representing the images with the following fields: * image/encoded: the string encoding the image as JPEG, * image/format: the string "jpeg" representing image format, * image/class/label: a list of integers representing the caption, * image/height: an integer representing the height, * image/width: an integer representing the width. Every field is actually a list of the corresponding type. """ eos_list = [1] if eos_list is None else eos_list if vocab_filename is not None: vocab_symbolizer = generator_utils.get_or_generate_vocab( data_dir, tmp_dir, vocab_filename, vocab_size) _get_mscoco(tmp_dir) caption_filepath = ( _MSCOCO_TRAIN_CAPTION_FILE if training else _MSCOCO_EVAL_CAPTION_FILE) caption_filepath = os.path.join(tmp_dir, caption_filepath) prefix = _MSCOCO_TRAIN_PREFIX if training else _MSCOCO_EVAL_PREFIX caption_file = io.open(caption_filepath) caption_json = json.load(caption_file) # Dictionary from image_id to ((filename, height, width), captions). image_dict = dict() for image in caption_json["images"]: image_dict[image["id"]] = [(image["file_name"], image["height"], image["width"]), []] annotations = caption_json["annotations"] annotation_count = len(annotations) image_count = len(image_dict) tf.logging.info("Processing %d images and %d labels\n" % (image_count, annotation_count)) for annotation in annotations: image_id = annotation["image_id"] image_dict[image_id][1].append(annotation["caption"]) data = list(image_dict.values())[start_from:start_from + how_many] random.shuffle(data) for image_info, labels in data: image_filename = image_info[0] image_filepath = os.path.join(tmp_dir, prefix, image_filename) with tf.gfile.Open(image_filepath, "r") as f: encoded_image_data = f.read() height, width = image_info[1], image_info[2] for label in labels: if vocab_filename is None: label = [ord(c) for c in label] + eos_list else: label = vocab_symbolizer.encode(label) + eos_list yield { "image/encoded": [encoded_image_data], "image/format": ["jpeg"], "image/class/label": label, "image/height": [height], "image/width": [width] } class Image2TextProblem(ImageProblem): """Base class for image-to-text problems.""" @property def is_character_level(self): raise NotImplementedError() @property def targeted_vocab_size(self): raise NotImplementedError() # Not needed if self.is_character_level. @property def target_space_id(self): raise NotImplementedError() @property def train_shards(self): raise NotImplementedError() @property def dev_shards(self): raise NotImplementedError() def generator(self, data_dir, tmp_dir, is_training): raise NotImplementedError() def example_reading_spec(self): label_key = "image/class/label" data_fields, data_items_to_decoders = ( super(Image2TextProblem, self).example_reading_spec()) data_fields[label_key] = tf.FixedLenFeature((1,), tf.int64) data_items_to_decoders[ "targets"] = tf.contrib.slim.tfexample_decoder.Tensor(label_key) return data_fields, data_items_to_decoders def feature_encoders(self, data_dir): if self.is_character_level: encoder = text_encoder.ByteTextEncoder() else: vocab_filename = os.path.join( data_dir, "vocab.endefr.%d" % self.targeted_vocab_size) encoder = text_encoder.SubwordTextEncoder(vocab_filename) return {"targets": encoder} def hparams(self, defaults, unused_model_hparams): p = defaults p.input_modality = {"inputs": (registry.Modalities.IMAGE, 256)} encoder = self._encoders["targets"] p.target_modality = (registry.Modalities.SYMBOL, encoder.vocab_size) p.batch_size_multiplier = 256 p.max_expected_batch_size_per_shard = 2 p.loss_multiplier = 1.0 p.input_space_id = problem.SpaceID.IMAGE p.target_space_id = self.target_space_id def generate_data(self, data_dir, tmp_dir, task_id=-1): generator_utils.generate_dataset_and_shuffle( self.generator(data_dir, tmp_dir, True), self.training_filepaths(data_dir, self.train_shards, shuffled=False), self.generator(data_dir, tmp_dir, False), self.dev_filepaths(data_dir, self.dev_shards, shuffled=False)) @registry.register_problem class ImageMsCocoCharacters(Image2TextProblem): """MSCOCO, character level.""" @property def is_character_level(self): return True @property def target_space_id(self): return problem.SpaceID.EN_CHR @property def train_shards(self): return 100 @property def dev_shards(self): return 10 def preprocess_example(self, example, mode, _): return imagenet_preprocess_example(example, mode) def generator(self, data_dir, tmp_dir, is_training): if is_training: return mscoco_generator(data_dir, tmp_dir, True, 80000) else: return mscoco_generator(data_dir, tmp_dir, False, 40000) raise NotImplementedError() @registry.register_problem class ImageMsCocoTokens8k(ImageMsCocoCharacters): """MSCOCO, 8k tokens vocab.""" @property def is_character_level(self): return False @property def targeted_vocab_size(self): return 2**13 # 8192 @property def target_space_id(self): return problem.SpaceID.EN_TOK @property def train_shards(self): return 100 @property def dev_shards(self): return 10 def generator(self, data_dir, tmp_dir, is_training): vocab_filename = "vocab.endefr.%d" % self.targeted_vocab_size if is_training: return mscoco_generator( data_dir, tmp_dir, True, 80000, vocab_filename=vocab_filename, vocab_size=self.targeted_vocab_size) else: return mscoco_generator( data_dir, tmp_dir, False, 40000, vocab_filename=vocab_filename, vocab_size=self.targeted_vocab_size) @registry.register_problem class ImageMsCocoTokens32k(ImageMsCocoTokens8k): """MSCOCO, 32k tokens vocab.""" @property def targeted_vocab_size(self): return 2**15 # 32768 @registry.register_problem class OcrTest(Image2TextProblem): """OCR test problem.""" @property def is_small(self): return True @property def is_character_level(self): return True @property def target_space_id(self): return problem.SpaceID.EN_CHR @property def train_shards(self): return 1 @property def dev_shards(self): return 1 def preprocess_example(self, example, mode, _): # Resize from usual size ~1350x60 to 90x4 in this test. img = example["inputs"] example["inputs"] = tf.to_int64( tf.image.resize_images(img, [90, 4], tf.image.ResizeMethod.AREA)) return example def generator(self, data_dir, tmp_dir, is_training): # In this test problem, we assume that the data is in tmp_dir/ocr/ in # files names 0.png, 0.txt, 1.png, 1.txt and so on until num_examples. num_examples = 2 ocr_dir = os.path.join(tmp_dir, "ocr/") tf.logging.info("Looking for OCR data in %s." % ocr_dir) for i in xrange(num_examples): image_filepath = os.path.join(ocr_dir, "%d.png" % i) text_filepath = os.path.join(ocr_dir, "%d.txt" % i) with tf.gfile.Open(text_filepath, "rb") as f: label = f.read() with tf.gfile.Open(image_filepath, "rb") as f: encoded_image_data = f.read() # In PNG files width and height are stored in these bytes. width, height = struct.unpack(">ii", encoded_image_data[16:24]) yield { "image/encoded": [encoded_image_data], "image/format": ["png"], "image/class/label": label.strip(), "image/height": [height], "image/width": [width] }
33.343777
97
0.703148
794d6692f8e8a0c8d244241dcd86f497ef7ab6f9
29,848
py
Python
src/obflowsim/mm/mm_run_fits_ldr.py
misken/obflowsim
19cad8e292435082ff47f3a09f68be69edff4361
[ "MIT" ]
null
null
null
src/obflowsim/mm/mm_run_fits_ldr.py
misken/obflowsim
19cad8e292435082ff47f3a09f68be69edff4361
[ "MIT" ]
null
null
null
src/obflowsim/mm/mm_run_fits_ldr.py
misken/obflowsim
19cad8e292435082ff47f3a09f68be69edff4361
[ "MIT" ]
null
null
null
import sys import argparse from pathlib import Path import pickle import pandas as pd import matplotlib.pyplot as plt from obflowsim.mm.mm_fitting import crossval_summarize_mm from obflowsim.mm.mm_process_fitted_models import create_cv_plots, create_coeff_plots from obflowsim.mm.mm_process_fitted_models import create_metrics_df, create_predictions_df # Make sure no interactive plotting happens during plot generation plt.ioff() UNIT = "ldr" def fit_models(experiment, input_path, output_path, figures_path): """ Fit metamodels for LDR unit Parameters ---------- experiment : str, identifier used in folder and filenames Returns ------- """ unit = UNIT # input_path = Path(f"input/{experiment}") # output_path = Path(f"output/{experiment}") # figures_path = Path(f"output/{experiment}", "figures") metrics_path_filename = Path(output_path, f"{experiment}_{unit}_metrics.csv") pickle_path_filename = Path(output_path, f"{experiment}_{unit}_results.pkl") # X matrices X_ldr_noq = pd.read_csv(Path(input_path, f'X_ldr_noq_{experiment}.csv'), index_col=0) X_ldr_basicq = pd.read_csv(Path(input_path, f'X_ldr_basicq_{experiment}.csv'), index_col=0) X_ldr_q = pd.read_csv(Path(input_path, f'X_ldr_q_{experiment}.csv'), index_col=0) X_ldr_occmean_onlyq = pd.read_csv(Path(input_path, f'X_ldr_occmean_onlyq_{experiment}.csv'), index_col=0) X_ldr_occp95_onlyq = pd.read_csv(Path(input_path, f'X_ldr_occp95_onlyq_{experiment}.csv'), index_col=0) X_ldr_probblocked_onlyq = pd.read_csv(Path(input_path, f'X_ldr_probblocked_onlyq_{experiment}.csv'), index_col=0) X_ldr_condmeantimeblocked_onlyq = \ pd.read_csv(Path(input_path, f'X_ldr_condmeantimeblocked_onlyq_{experiment}.csv'), index_col=0) # y vectors y_ldr_occmean = pd.read_csv(Path(input_path, f'y_ldr_occmean_{experiment}.csv'), index_col=0).squeeze("columns") y_ldr_occp95 = pd.read_csv(Path(input_path, f'y_ldr_occp95_{experiment}.csv'), index_col=0).squeeze("columns") y_ldr_probblocked = \ pd.read_csv(Path(input_path, f'y_ldr_probblocked_{experiment}.csv'), index_col=0).squeeze("columns") y_ldr_condmeantimeblocked = \ pd.read_csv(Path(input_path, f'y_ldr_condmeantimeblocked_{experiment}.csv'), index_col=0).squeeze("columns") # Fit models # Queueing models ldr_occmean_q_effload_results = \ crossval_summarize_mm('ldr_occmean_q_effload', 'ldr', 'occmean', X_ldr_q, y_ldr_occmean, scale=False, flavor='load', col_idx_arate=0, col_idx_meansvctime=21) ldr_occp95_q_sqrteffload_results = \ crossval_summarize_mm('ldr_occp95_q_sqrteffload', 'ldr', 'occp95', X_ldr_q, y_ldr_occp95, scale=False, flavor='sqrtload', col_idx_arate=0, col_idx_meansvctime=21, load_pctile=0.95) probblocked_q_erlangc_results = \ crossval_summarize_mm('ldr_probblocked_q_erlangc', 'ldr', 'probblocked', X_ldr_q, y_ldr_probblocked, scale=False, fit_intercept=True, flavor='erlangc', col_idx_arate=0, col_idx_meansvctime=5, col_idx_numservers=7) condmeantimeblocked_q_mgc_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_q_mgc', 'ldr', 'condmeantimeblocked', X_ldr_q, y_ldr_condmeantimeblocked, scale=False, fit_intercept=True, flavor='condmeanwaitldr', col_idx_arate=0, col_idx_meansvctime=5, col_idx_numservers=7, col_idx_cv2svctime=18) # Linear models using only queueing approximation terms ldr_occmean_onlyq_lm_results = \ crossval_summarize_mm('ldr_occmean_onlyq_lm', 'ldr', 'occmean', X_ldr_occmean_onlyq, y_ldr_occmean, scale=False, flavor='lm') ldr_occp95_onlyq_lm_results = \ crossval_summarize_mm('ldr_occp95_onlyq_lm', 'ldr', 'occp95', X_ldr_occp95_onlyq, y_ldr_occp95, scale=False, flavor='lm') probblocked_onlyq_lm_results = \ crossval_summarize_mm('ldr_probblocked_onlyq_lm', 'ldr', 'probblocked', X_ldr_probblocked_onlyq, y_ldr_probblocked, scale=False, flavor='lm') condmeantimeblocked_onlyq_lm_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_onlyq_lm', 'ldr', 'condmeantimeblocked', X_ldr_condmeantimeblocked_onlyq, y_ldr_condmeantimeblocked, scale=False, flavor='lm') # Linear regression (lm) ldr_occmean_basicq_lm_results = \ crossval_summarize_mm('ldr_occmean_basicq_lm', 'ldr', 'occmean', X_ldr_basicq, y_ldr_occmean, scale=False, flavor='lm') ldr_occmean_q_lm_results = \ crossval_summarize_mm('ldr_occmean_q_lm', 'ldr', 'occmean', X_ldr_q, y_ldr_occmean, scale=False, flavor='lm') ldr_occmean_noq_lm_results = \ crossval_summarize_mm('ldr_occmean_noq_lm', 'ldr', 'occmean', X_ldr_noq, y_ldr_occmean, scale=False, flavor='lm') ldr_occp95_basicq_lm_results = \ crossval_summarize_mm('ldr_occp95_basicq_lm', 'ldr', 'occp95', X_ldr_basicq, y_ldr_occp95, scale=False, flavor='lm') ldr_occp95_q_lm_results = \ crossval_summarize_mm('ldr_occp95_q_lm', 'ldr', 'occp95', X_ldr_q, y_ldr_occp95, scale=False, flavor='lm') ldr_occp95_noq_lm_results = \ crossval_summarize_mm('ldr_occp95_noq_lm', 'ldr', 'occp95', X_ldr_noq, y_ldr_occp95, scale=False, flavor='lm') probblocked_basicq_lm_results = \ crossval_summarize_mm('ldr_probblocked_basicq_lm', 'ldr', 'probblocked', X_ldr_basicq, y_ldr_probblocked, scale=False, fit_intercept=True, flavor='lm') probblocked_q_lm_results = \ crossval_summarize_mm('ldr_probblocked_q_lm', 'ldr', 'probblocked', X_ldr_q, y_ldr_probblocked, scale=False, fit_intercept=True, flavor='lm') probblocked_noq_lm_results = \ crossval_summarize_mm('ldr_probblocked_noq_lm', 'ldr', 'probblocked', X_ldr_noq, y_ldr_probblocked, scale=False, fit_intercept=True, flavor='lm') condmeantimeblocked_basicq_lm_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_basicq_lm', 'ldr', 'condmeantimeblocked', X_ldr_basicq, y_ldr_condmeantimeblocked, scale=False, fit_intercept=True, flavor='lm') condmeantimeblocked_q_lm_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_q_lm', 'ldr', 'condmeantimeblocked', X_ldr_q, y_ldr_condmeantimeblocked, scale=False, fit_intercept=True, flavor='lm') condmeantimeblocked_noq_lm_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_noq_lm', 'ldr', 'condmeantimeblocked', X_ldr_noq, y_ldr_condmeantimeblocked, scale=False, fit_intercept=True, flavor='lm') # LassoCV (lassocv) ldr_occmean_basicq_lassocv_results = \ crossval_summarize_mm('ldr_occmean_basicq_lassocv', 'ldr', 'occmean', X_ldr_basicq, y_ldr_occmean, scale=True, flavor='lassocv', lasso_max_iter=3000) ldr_occmean_q_lassocv_results = \ crossval_summarize_mm('ldr_occmean_q_lassocv', 'ldr', 'occmean', X_ldr_q, y_ldr_occmean, scale=True, flavor='lassocv', lasso_max_iter=3000) ldr_occmean_noq_lassocv_results = \ crossval_summarize_mm('ldr_occmean_noq_lassocv', 'ldr', 'occmean', X_ldr_noq, y_ldr_occmean, scale=True, flavor='lassocv', lasso_max_iter=3000) ldr_occp95_basicq_lassocv_results = \ crossval_summarize_mm('ldr_occp95_basicq_lassocv', 'ldr', 'occp95', X_ldr_basicq, y_ldr_occp95, scale=True, flavor='lassocv', lasso_max_iter=3000) ldr_occp95_q_lassocv_results = \ crossval_summarize_mm('ldr_occp95_q_lassocv', 'ldr', 'occp95', X_ldr_q, y_ldr_occp95, scale=True, flavor='lassocv', lasso_max_iter=3000) ldr_occp95_noq_lassocv_results = \ crossval_summarize_mm('ldr_occp95_noq_lassocv', 'ldr', 'occp95', X_ldr_noq, y_ldr_occp95, scale=True, flavor='lassocv', lasso_max_iter=3000) probblocked_basicq_lassocv_results = \ crossval_summarize_mm('ldr_probblocked_basicq_lassocv', 'ldr', 'probblocked', X_ldr_basicq, y_ldr_probblocked, scale=True, flavor='lassocv', lasso_max_iter=3000) probblocked_q_lassocv_results = \ crossval_summarize_mm('ldr_probblocked_q_lassocv', 'ldr', 'probblocked', X_ldr_q, y_ldr_probblocked, scale=True, flavor='lassocv', lasso_max_iter=3000) probblocked_noq_lassocv_results = \ crossval_summarize_mm('ldr_probblocked_noq_lassocv', 'ldr', 'probblocked', X_ldr_noq, y_ldr_probblocked, scale=True, flavor='lassocv', lasso_max_iter=3000) condmeantimeblocked_basicq_lassocv_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_basicq_lassocv', 'ldr', 'condmeantimeblocked', X_ldr_basicq, y_ldr_condmeantimeblocked, scale=True, flavor='lassocv', lasso_max_iter=3000) condmeantimeblocked_q_lassocv_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_q_lassocv', 'ldr', 'condmeantimeblocked', X_ldr_q, y_ldr_condmeantimeblocked, scale=True, flavor='lassocv', lasso_max_iter=3000) condmeantimeblocked_noq_lassocv_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_noq_lassocv', 'ldr', 'condmeantimeblocked', X_ldr_noq, y_ldr_condmeantimeblocked, scale=True, flavor='lassocv', lasso_max_iter=3000) # Polynomial regression (poly) ldr_occmean_basicq_poly_results = \ crossval_summarize_mm('ldr_occmean_basicq_poly', 'ldr', 'occmean', X_ldr_basicq, y_ldr_occmean, scale=False, flavor='poly') ldr_occmean_q_poly_results = \ crossval_summarize_mm('ldr_occmean_q_poly', 'ldr', 'occmean', X_ldr_q, y_ldr_occmean, scale=False, flavor='poly') ldr_occmean_noq_poly_results = \ crossval_summarize_mm('ldr_occmean_noq_poly', 'ldr', 'occmean', X_ldr_noq, y_ldr_occmean, scale=False, flavor='poly') ldr_occp95_basicq_poly_results = \ crossval_summarize_mm('ldr_occp95_basicq_poly', 'ldr', 'occp95', X_ldr_basicq, y_ldr_occp95, scale=False, flavor='poly') ldr_occp95_q_poly_results = \ crossval_summarize_mm('ldr_occp95_q_poly', 'ldr', 'occp95', X_ldr_q, y_ldr_occp95, scale=False, flavor='poly') ldr_occp95_noq_poly_results = \ crossval_summarize_mm('ldr_occp95_noq_poly', 'ldr', 'occp95', X_ldr_noq, y_ldr_occp95, scale=False, flavor='poly') probblocked_basicq_poly_results = \ crossval_summarize_mm('ldr_probblocked_basicq_poly', 'ldr', 'probblocked', X_ldr_basicq, y_ldr_probblocked, scale=False, flavor='poly') probblocked_q_poly_results = \ crossval_summarize_mm('ldr_probblocked_q_poly', 'ldr', 'probblocked', X_ldr_q, y_ldr_probblocked, scale=False, flavor='poly') probblocked_noq_poly_results = \ crossval_summarize_mm('ldr_probblocked_noq_poly', 'ldr', 'probblocked', X_ldr_noq, y_ldr_probblocked, scale=False, flavor='poly') condmeantimeblocked_basicq_poly_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_basicq_poly', 'ldr', 'condmeantimeblocked', X_ldr_basicq, y_ldr_condmeantimeblocked, scale=False, flavor='poly') condmeantimeblocked_q_poly_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_q_poly', 'ldr', 'condmeantimeblocked', X_ldr_q, y_ldr_condmeantimeblocked, scale=False, flavor='poly') condmeantimeblocked_noq_poly_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_noq_poly', 'ldr', 'condmeantimeblocked', X_ldr_noq, y_ldr_condmeantimeblocked, scale=False, flavor='poly') # Random forest (rf) ldr_occmean_basicq_rf_results = \ crossval_summarize_mm('ldr_occmean_basicq_rf', 'ldr', 'occmean', X_ldr_basicq, y_ldr_occmean, scale=False, flavor='rf') ldr_occmean_q_rf_results = \ crossval_summarize_mm('ldr_occmean_q_rf', 'ldr', 'occmean', X_ldr_q, y_ldr_occmean, scale=False, flavor='rf') ldr_occmean_noq_rf_results = \ crossval_summarize_mm('ldr_occmean_noq_rf', 'ldr', 'occmean', X_ldr_noq, y_ldr_occmean, scale=False, flavor='rf') ldr_occp95_basicq_rf_results = \ crossval_summarize_mm('ldr_occp95_basicq_rf', 'ldr', 'occp95', X_ldr_basicq, y_ldr_occp95, scale=False, flavor='rf') ldr_occp95_q_rf_results = \ crossval_summarize_mm('ldr_occp95_q_rf', 'ldr', 'occp95', X_ldr_q, y_ldr_occp95, scale=False, flavor='rf') ldr_occp95_noq_rf_results = \ crossval_summarize_mm('ldr_occp95_noq_rf', 'ldr', 'occp95', X_ldr_noq, y_ldr_occp95, scale=False, flavor='rf') probblocked_basicq_rf_results = \ crossval_summarize_mm('ldr_probblocked_basicq_rf', 'ldr', 'probblocked', X_ldr_basicq, y_ldr_probblocked, scale=False, flavor='rf') probblocked_q_rf_results = \ crossval_summarize_mm('ldr_probblocked_q_rf', 'ldr', 'probblocked', X_ldr_q, y_ldr_probblocked, scale=False, flavor='rf') probblocked_noq_rf_results = \ crossval_summarize_mm('ldr_probblocked_noq_rf', 'ldr', 'probblocked', X_ldr_noq, y_ldr_probblocked, scale=False, flavor='rf') condmeantimeblocked_basicq_rf_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_basicq_rf', 'ldr', 'condmeantimeblocked', X_ldr_basicq, y_ldr_condmeantimeblocked, scale=False, flavor='rf') condmeantimeblocked_q_rf_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_q_rf', 'ldr', 'condmeantimeblocked', X_ldr_q, y_ldr_condmeantimeblocked, scale=False, flavor='rf') condmeantimeblocked_noq_rf_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_noq_rf', 'ldr', 'condmeantimeblocked', X_ldr_noq, y_ldr_condmeantimeblocked, scale=False, flavor='rf') # Support vector regression (svr) ldr_occmean_basicq_svr_results = \ crossval_summarize_mm('ldr_occmean_basicq_svr', 'ldr', 'occmean', X_ldr_basicq, y_ldr_occmean, flavor='svr', scale=True) ldr_occmean_q_svr_results = \ crossval_summarize_mm('ldr_occmean_q_svr', 'ldr', 'occmean', X_ldr_q, y_ldr_occmean, flavor='svr', scale=True) ldr_occmean_noq_svr_results = \ crossval_summarize_mm('ldr_occmean_noq_svr', 'ldr', 'occmean', X_ldr_noq, y_ldr_occmean, flavor='svr', scale=True) ldr_occp95_basicq_svr_results = \ crossval_summarize_mm('ldr_occp95_basicq_svr', 'ldr', 'occp95', X_ldr_basicq, y_ldr_occp95, flavor='svr', scale=True) ldr_occp95_q_svr_results = \ crossval_summarize_mm('ldr_occp95_q_svr', 'ldr', 'occp95', X_ldr_q, y_ldr_occp95, flavor='svr', scale=True) ldr_occp95_noq_svr_results = \ crossval_summarize_mm('ldr_occp95_noq_svr', 'ldr', 'occp95', X_ldr_noq, y_ldr_occp95, flavor='svr', scale=True) probblocked_basicq_svr_results = \ crossval_summarize_mm('ldr_probblocked_basicq_svr', 'ldr', 'probblocked', X_ldr_basicq, y_ldr_probblocked, scale=True, flavor='svr') probblocked_q_svr_results = \ crossval_summarize_mm('ldr_probblocked_q_svr', 'ldr', 'probblocked', X_ldr_q, y_ldr_probblocked, scale=True, flavor='svr') probblocked_noq_svr_results = \ crossval_summarize_mm('ldr_probblocked_noq_svr', 'ldr', 'probblocked', X_ldr_noq, y_ldr_probblocked, scale=True, flavor='svr') condmeantimeblocked_basicq_svr_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_basicq_svr', 'ldr', 'condmeantimeblocked', X_ldr_basicq, y_ldr_condmeantimeblocked, scale=True, flavor='svr') condmeantimeblocked_q_svr_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_q_svr', 'ldr', 'condmeantimeblocked', X_ldr_q, y_ldr_condmeantimeblocked, scale=True, flavor='svr') condmeantimeblocked_noq_svr_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_noq_svr', 'ldr', 'condmeantimeblocked', X_ldr_noq, y_ldr_condmeantimeblocked, scale=True, flavor='svr') # MLPRegressor Neural net (nn) ldr_occmean_basicq_nn_results = \ crossval_summarize_mm('ldr_occmean_basicq_nn', 'ldr', 'occmean', X_ldr_basicq, y_ldr_occmean, flavor='nn', scale=True) ldr_occmean_q_nn_results = \ crossval_summarize_mm('ldr_occmean_q_nn', 'ldr', 'occmean', X_ldr_q, y_ldr_occmean, flavor='nn', scale=True) ldr_occmean_noq_nn_results = \ crossval_summarize_mm('ldr_occmean_noq_nn', 'ldr', 'occmean', X_ldr_noq, y_ldr_occmean, flavor='nn', scale=True) ldr_occp95_basicq_nn_results = \ crossval_summarize_mm('ldr_occp95_basicq_nn', 'ldr', 'occp95', X_ldr_basicq, y_ldr_occp95, flavor='nn', scale=True) ldr_occp95_q_nn_results = \ crossval_summarize_mm('ldr_occp95_q_nn', 'ldr', 'occp95', X_ldr_q, y_ldr_occp95, flavor='nn', scale=True) ldr_occp95_noq_nn_results = \ crossval_summarize_mm('ldr_occp95_noq_nn', 'ldr', 'occp95', X_ldr_noq, y_ldr_occp95, flavor='nn', scale=True) probblocked_basicq_nn_results = \ crossval_summarize_mm('ldr_probblocked_basicq_nn', 'ldr', 'probblocked', X_ldr_basicq, y_ldr_probblocked, scale=True, flavor='nn') probblocked_q_nn_results = \ crossval_summarize_mm('ldr_probblocked_q_nn', 'ldr', 'probblocked', X_ldr_q, y_ldr_probblocked, scale=True, flavor='nn') probblocked_noq_nn_results = \ crossval_summarize_mm('ldr_probblocked_noq_nn', 'ldr', 'probblocked', X_ldr_noq, y_ldr_probblocked, scale=True, flavor='nn') condmeantimeblocked_basicq_nn_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_basicq_nn', 'ldr', 'condmeantimeblocked', X_ldr_basicq, y_ldr_condmeantimeblocked, scale=True, flavor='nn') condmeantimeblocked_q_nn_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_q_nn', 'ldr', 'condmeantimeblocked', X_ldr_q, y_ldr_condmeantimeblocked, scale=True, flavor='nn') condmeantimeblocked_noq_nn_results = \ crossval_summarize_mm('ldr_condmeantimeblocked_noq_nn', 'ldr', 'condmeantimeblocked', X_ldr_noq, y_ldr_condmeantimeblocked, scale=True, flavor='nn') ldr_results = {'ldr_occmean_basicq_lm_results': ldr_occmean_basicq_lm_results, 'ldr_occmean_q_lm_results': ldr_occmean_q_lm_results, 'ldr_occmean_noq_lm_results': ldr_occmean_noq_lm_results, 'ldr_occmean_basicq_lassocv_results': ldr_occmean_basicq_lassocv_results, 'ldr_occmean_q_lassocv_results': ldr_occmean_q_lassocv_results, 'ldr_occmean_noq_lassocv_results': ldr_occmean_noq_lassocv_results, 'ldr_occmean_basicq_poly_results': ldr_occmean_basicq_poly_results, 'ldr_occmean_q_poly_results': ldr_occmean_q_poly_results, 'ldr_occmean_noq_poly_results': ldr_occmean_noq_poly_results, 'ldr_occmean_basicq_rf_results': ldr_occmean_basicq_rf_results, 'ldr_occmean_q_rf_results': ldr_occmean_q_rf_results, 'ldr_occmean_noq_rf_results': ldr_occmean_noq_rf_results, 'ldr_occmean_basicq_svr_results': ldr_occmean_basicq_svr_results, 'ldr_occmean_q_svr_results': ldr_occmean_q_svr_results, 'ldr_occmean_noq_svr_results': ldr_occmean_noq_svr_results, 'ldr_occmean_basicq_nn_results': ldr_occmean_basicq_nn_results, 'ldr_occmean_q_nn_results': ldr_occmean_q_nn_results, 'ldr_occmean_noq_nn_results': ldr_occmean_noq_nn_results, 'ldr_occmean_q_effload_results': ldr_occmean_q_effload_results, 'ldr_occp95_basicq_lm_results': ldr_occp95_basicq_lm_results, 'ldr_occp95_q_lm_results': ldr_occp95_q_lm_results, 'ldr_occp95_noq_lm_results': ldr_occp95_noq_lm_results, 'ldr_occp95_basicq_lassocv_results': ldr_occp95_basicq_lassocv_results, 'ldr_occp95_q_lassocv_results': ldr_occp95_q_lassocv_results, 'ldr_occp95_noq_lassocv_results': ldr_occp95_noq_lassocv_results, 'ldr_occp95_basicq_poly_results': ldr_occp95_basicq_poly_results, 'ldr_occp95_q_poly_results': ldr_occp95_q_poly_results, 'ldr_occp95_noq_poly_results': ldr_occp95_noq_poly_results, 'ldr_occp95_basicq_rf_results': ldr_occp95_basicq_rf_results, 'ldr_occp95_q_rf_results': ldr_occp95_q_rf_results, 'ldr_occp95_noq_rf_results': ldr_occp95_noq_rf_results, 'ldr_occp95_basicq_svr_results': ldr_occp95_basicq_svr_results, 'ldr_occp95_q_svr_results': ldr_occp95_q_svr_results, 'ldr_occp95_noq_svr_results': ldr_occp95_noq_svr_results, 'ldr_occp95_basicq_nn_results': ldr_occp95_basicq_nn_results, 'ldr_occp95_q_nn_results': ldr_occp95_q_nn_results, 'ldr_occp95_noq_nn_results': ldr_occp95_noq_nn_results, 'ldr_occp95_q_sqrteffload_results': ldr_occp95_q_sqrteffload_results, 'ldr_probblocked_basicq_lm_results': probblocked_basicq_lm_results, 'ldr_probblocked_q_lm_results': probblocked_q_lm_results, 'ldr_probblocked_noq_lm_results': probblocked_noq_lm_results, 'ldr_probblocked_basicq_lassocv_results': probblocked_basicq_lassocv_results, 'ldr_probblocked_q_lassocv_results': probblocked_q_lassocv_results, 'ldr_probblocked_noq_lassocv_results': probblocked_noq_lassocv_results, 'ldr_probblocked_basicq_poly_results': probblocked_basicq_poly_results, 'ldr_probblocked_q_poly_results': probblocked_q_poly_results, 'ldr_probblocked_noq_poly_results': probblocked_noq_poly_results, 'ldr_probblocked_basicq_rf_results': probblocked_basicq_rf_results, 'ldr_probblocked_q_rf_results': probblocked_q_rf_results, 'ldr_probblocked_noq_rf_results': probblocked_noq_rf_results, 'ldr_probblocked_basicq_svr_results': probblocked_basicq_svr_results, 'ldr_probblocked_q_svr_results': probblocked_q_svr_results, 'ldr_probblocked_noq_svr_results': probblocked_noq_svr_results, 'ldr_probblocked_basicq_nn_results': probblocked_basicq_nn_results, 'ldr_probblocked_q_nn_results': probblocked_q_nn_results, 'ldr_probblocked_noq_nn_results': probblocked_noq_nn_results, 'ldr_probblocked_q_erlangc_results': probblocked_q_erlangc_results, 'ldr_condmeantimeblocked_basicq_lm_results': condmeantimeblocked_basicq_lm_results, 'ldr_condmeantimeblocked_q_lm_results': condmeantimeblocked_q_lm_results, 'ldr_condmeantimeblocked_noq_lm_results': condmeantimeblocked_noq_lm_results, 'ldr_condmeantimeblocked_basicq_lassocv_results': condmeantimeblocked_basicq_lassocv_results, 'ldr_condmeantimeblocked_q_lassocv_results': condmeantimeblocked_q_lassocv_results, 'ldr_condmeantimeblocked_noq_lassocv_results': condmeantimeblocked_noq_lassocv_results, 'ldr_condmeantimeblocked_basicq_poly_results': condmeantimeblocked_basicq_poly_results, 'ldr_condmeantimeblocked_q_poly_results': condmeantimeblocked_q_poly_results, 'ldr_condmeantimeblocked_noq_poly_results': condmeantimeblocked_noq_poly_results, 'ldr_condmeantimeblocked_basicq_rf_results': condmeantimeblocked_basicq_rf_results, 'ldr_condmeantimeblocked_q_rf_results': condmeantimeblocked_q_rf_results, 'ldr_condmeantimeblocked_noq_rf_results': condmeantimeblocked_noq_rf_results, 'ldr_condmeantimeblocked_basicq_svr_results': condmeantimeblocked_basicq_svr_results, 'ldr_condmeantimeblocked_q_svr_results': condmeantimeblocked_q_svr_results, 'ldr_condmeantimeblocked_noq_svr_results': condmeantimeblocked_noq_svr_results, 'ldr_condmeantimeblocked_basicq_nn_results': condmeantimeblocked_basicq_nn_results, 'ldr_condmeantimeblocked_q_nn_results': condmeantimeblocked_q_nn_results, 'ldr_condmeantimeblocked_noq_nn_results': condmeantimeblocked_noq_nn_results, 'ldr_condmeantimeblocked_q_mgc_results': condmeantimeblocked_q_mgc_results, 'ldr_occmean_onlyq_lm_results': ldr_occmean_onlyq_lm_results, 'ldr_occp95_onlyq_lm_results': ldr_occp95_onlyq_lm_results, 'ldr_probblocked_onlyq_lm_results': probblocked_onlyq_lm_results, 'ldr_condmeantimeblocked_onlyq_lm_results': condmeantimeblocked_onlyq_lm_results } create_cv_plots(experiment, unit, ldr_results, figures_path) create_coeff_plots(experiment, unit, ldr_results, figures_path) metrics_df = create_metrics_df(ldr_results) metrics_df.to_csv(metrics_path_filename, index=False) predictions_df = create_predictions_df(ldr_results) predictions_df.to_csv(Path(output_path, f"{experiment}_{unit}_predictions.csv"), index=False) sys.setrecursionlimit(10000) # Pickle the results with open(pickle_path_filename, 'wb') as persisted_file: pickle.dump(ldr_results, persisted_file) def process_command_line(argv=None): """ Parse command line arguments `argv` is a list of arguments, or `None` for ``sys.argv[1:]``. Return a Namespace representing the argument list. """ # Create the parser parser = argparse.ArgumentParser(prog='mm_run_fits_ldr', description='Fit metamodels for LDR') # Add arguments parser.add_argument( "experiment", type=str, help="Experiment ID; used in folder and filenames" ) parser.add_argument( "input_path", type=str, help="Path containing X and y csv files" ) parser.add_argument( "output_path", type=str, help="Path for model fit outputs (except plots)" ) parser.add_argument( "plots_path", type=str, help="Path for output plots" ) # do the parsing args = parser.parse_args(argv) return args def main(argv=None): """ Main driver program Parameters ---------- argv Returns ------- """ # Parse command line arguments args = process_command_line(argv) # fit models fit_models(args.experiment, Path(args.input_path), Path(args.output_path), Path(args.plots_path)) return 0 if __name__ == '__main__': sys.exit(main())
50.675722
120
0.646207
794d66b7fc28ad18870fdc159e5a60e89b47da18
1,151
py
Python
example/usage/example_vkofficial.py
vodka2/vkaudiotoken-python
5720e4cf77f5e1b20c3bf57f3df0717638a539e0
[ "MIT" ]
32
2020-07-21T18:32:59.000Z
2022-03-20T21:16:11.000Z
example/usage/example_vkofficial.py
vodka2/vkaudiotoken-python
5720e4cf77f5e1b20c3bf57f3df0717638a539e0
[ "MIT" ]
1
2020-10-04T04:41:06.000Z
2020-10-05T11:43:48.000Z
example/usage/example_vkofficial.py
vodka2/vkaudiotoken-python
5720e4cf77f5e1b20c3bf57f3df0717638a539e0
[ "MIT" ]
2
2021-09-21T01:17:05.000Z
2022-03-17T10:17:22.000Z
from __future__ import print_function try: import vkaudiotoken except ImportError: import path_hack from vkaudiotoken import supported_clients import sys import requests import json import re token = sys.argv[1] user_agent = supported_clients.VK_OFFICIAL.user_agent sess = requests.session() sess.headers.update({'User-Agent': user_agent}) def get_mp3_from_m3u8(url): if 'index.m3u8?' not in url: return url if '/audios/' in url: return re.sub(r'^(.+?)/[^/]+?/audios/([^/]+)/.+$', '\\1/audios/\\2.mp3', url) else: return re.sub(r'^(.+?)/(p[0-9]+)/[^/]+?/([^/]+)/.+$', '\\1/\\2/\\3.mp3', url) def prettyprint_mp3(result): print(re.sub(r'(?<=")https://.+?index.m3u8\?.+?(?=")', lambda m: get_mp3_from_m3u8(m.group(0)), json.dumps(json.loads(result.content.decode('utf-8')), indent=2))) prettyprint_mp3(sess.post( "https://api.vk.com/method/audio.getCatalog", data=[('access_token', token), ('https', 1), ('ref', 'search'), ('extended', 1), ('lang', 'en'), ('query', 'Justin Bieber - Baby'), ('v', '5.116')] ))
25.577778
99
0.581234
794d672e31b982f76591efa5b5d8c2704705469d
3,520
py
Python
docs/conf.py
naddeoa/whylogs-python
25c6d38961f2608bf331981667a9fe143e3437ef
[ "Apache-2.0" ]
null
null
null
docs/conf.py
naddeoa/whylogs-python
25c6d38961f2608bf331981667a9fe143e3437ef
[ "Apache-2.0" ]
null
null
null
docs/conf.py
naddeoa/whylogs-python
25c6d38961f2608bf331981667a9fe143e3437ef
[ "Apache-2.0" ]
null
null
null
# Configuration file for the Sphinx documentation builder. # # This file only contains a selection of the most common options. For a full # list see the documentation: # https://www.sphinx-doc.org/en/master/usage/configuration.html # -- Path setup -------------------------------------------------------------- # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. # # import os # import sys # sys.path.insert(0, os.path.abspath('.')) import os import re import sys import uuid from collections import namedtuple import sphinx_rtd_theme from sphinx.ext.autodoc import between # Add package paths to the python path to allow sphinx to import modules sys.path.insert(0, os.path.abspath("../src/")) # -- Project information ----------------------------------------------------- project = "WhyLogs" copyright = "2020, WhyLabs, Inc" author = "WhyLabs" gitstamp_fmt = "%d %b %Y" def read(rel_path): import codecs here = os.path.abspath(os.path.dirname(__file__)) with codecs.open(os.path.join(here, rel_path), "r") as fp: return fp.read() # This string is managed by bump2version version = " 0.1.1" # -- General configuration --------------------------------------------------- # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ "sphinx_gallery.gen_gallery", "autoapi.extension", "sphinx_rtd_theme", "sphinx.ext.autodoc", "sphinx.ext.todo", "sphinx.ext.coverage", # 'sphinx.ext.mathjax' "sphinx.ext.napoleon", "sphinxcontrib.contentui", "sphinx_gitstamp", "sphinx.ext.autosectionlabel", "sphinx_autorun", ] sphinx_gallery_conf = { "examples_dirs": "../examples", # path to your example scripts "gallery_dirs": "auto_examples", # path to where to save gallery generated output # 'filename_pattern': '/plot_', # Default pattern of filenames to execute # 'ignore_pattern': r'__init__\.py', # Default filename ignore pattern # 'filename_pattern': '.*\.py$', # 'ignore_pattern': '.*', } autoapi_type = "python" autoapi_dirs = ["../src/whylogs"] autoapi_options = [ "members", "undoc-members", "private-members", "show-inheritance", "show-module-summary", "special-members", "imported-members", "inherited-members", ] # Add any paths that contain templates here, relative to this directory. templates_path = ["_templates"] # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. # This pattern also affects html_static_path and html_extra_path. exclude_patterns = [] index_doc = "index" # -- Options for HTML output ------------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. # html_theme = "sphinx_rtd_theme" # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ["_static"] html_css_files = ["style.css"] html_theme_options = { "navigation_depth": 6, # 'github_url': 'https://github.com/whylabs/whylogs-python', }
29.579832
86
0.668182
794d6736796a918a98b389f596d157edfa494573
4,307
py
Python
updatedns.py
dnorth98/updatedns
1244b68dbe119f104546d8489d1f10676d739eee
[ "MIT" ]
null
null
null
updatedns.py
dnorth98/updatedns
1244b68dbe119f104546d8489d1f10676d739eee
[ "MIT" ]
null
null
null
updatedns.py
dnorth98/updatedns
1244b68dbe119f104546d8489d1f10676d739eee
[ "MIT" ]
null
null
null
#!/usr/bin/env python """ Update DNS Entry in Amazon Route53 via boto ------------------------------------------------------------------- Creator: Brian Parsons <brian@pmex.com> Modified: Dave North <dnorth98@gmail.com> Requires: boto - http://boto.cloudhackers.com/en/latest/index.html Variables: --------------- awskeyid - Amazon AWS Credentials Security Key ID awskeysecret - Amazon AWS Secret that matches the security key id NOTE: If no keys are provided, instance role credentials will be used """ ## ## SET CONFIGURATION VARIABLES ## awskeyid = "" awskeysecret = "" ## ## END OF VARIABLES ## import re import socket import sys import urllib import syslog import os from boto.route53.connection import Route53Connection from boto.route53.record import ResourceRecordSets from boto.route53.exception import DNSServerError def myLog(message): procName = __file__ syslogMsg = procName + ": " + message syslog.syslog(syslogMsg) print '%s' % message # get the FQDN # the domain is specified on the command line, hostname we get from the system if len(sys.argv) == 2: from socket import gethostname hostname = gethostname() domainname = sys.argv[1] if domainname[0] != ".": domainname = "." + domainname fqdn = hostname + domainname myLog("FQDN: %s" % fqdn) else: myLog("No domain component specified on command line") sys.exit(1) try: hostname except NameError: myLog("Unable to read hostname") sys.exit(1) # Add trailing dot to hostname if it doesn't have one if fqdn[-1:] != ".": fqdn += "." # Get current IP try: myLog("Getting public IP from instance metadata service") ipsite = urllib.urlopen('http://169.254.169.254/latest/meta-data/public-ipv4') except: myLog("Connection error getting IP address") sys.exit(1) response = ipsite.read() ips = re.findall("(?:\d{1,3}\.){3}\d{1,3}", response) if type(ips) in [list, tuple, set]: for record in ips: newip = record try: newip except NameError: myLog("Unable to find IP address in response from check site") sys.exit(1) # Verify it's a good ip try: socket.inet_aton(newip) except socket.error: myLog("Received invalid IP address") sys.exit(1) myLog("Current Public IP: %s" % newip) # Initialize the connection to AWS Route53 if awskeyid and awskeysecret: route53 = Route53Connection(awskeyid, awskeysecret) else: route53 = Route53Connection() # Get the zoneid try: route53zones = route53.get_all_hosted_zones() except DNSServerError, e: myLog("Connection error to AWS. Check your credentials") myLog("Error %s - %s" % (e.code, str(e))) sys.exit(1) for zone in route53zones['ListHostedZonesResponse']['HostedZones']: if zone['Name'][0:-1] in fqdn: zoneid = zone['Id'].replace('/hostedzone/', '') myLog("Found Route53 Zone %s for fqdn %s" % (zoneid, fqdn)) try: zoneid except NameError: myLog("Unable to find Route53 Zone for %s" % fqdn) sys.exit(1) # Find the old record if it exists try: sets = route53.get_all_rrsets(zoneid) except DNSServerError, e: myLog("Connection error to AWS") myLog("Error %s - %s" % (e.code, str(e))) sys.exit(1) for rset in sets: if rset.name == fqdn and rset.type == 'A': curiprecord = rset.resource_records if type(curiprecord) in [list, tuple, set]: for record in curiprecord: curip = record myLog("Current DNS IP: %s" % curip) curttl = rset.ttl myLog("Current DNS TTL: %s" % curttl) if curip != newip: # Remove the old record myLog("Removing old record from Route53...") change1 = ResourceRecordSets(route53, zoneid) removeold = change1.add_change("DELETE", fqdn, "A", curttl) removeold.add_value(curip) change1.commit() else: myLog("IPs match in DNS, no DNS changes being made") sys.exit(0) try: curip except NameError: myLog("FQDN %s not found in current zone record" % fqdn) # Add the new record myLog("Adding %s to DNS as %s..." % ( fqdn, newip)) change2 = ResourceRecordSets(route53, zoneid) change = change2.add_change("CREATE", fqdn, "A", 60) change.add_value(newip) change2.commit()
25.485207
82
0.647318
794d68f8c931838ed47738599d67140d7f1d2a81
1,806
py
Python
thespian/test/test_badmessage.py
dendron2000/Thespian
0acbc5a0803f6d2be3421ea6eb08c6beecbf3802
[ "MIT" ]
210
2015-08-31T19:39:34.000Z
2020-01-10T08:07:48.000Z
thespian/test/test_badmessage.py
dendron2000/Thespian
0acbc5a0803f6d2be3421ea6eb08c6beecbf3802
[ "MIT" ]
85
2017-04-08T19:28:42.000Z
2022-03-23T15:25:49.000Z
thespian/test/test_badmessage.py
dendron2000/Thespian
0acbc5a0803f6d2be3421ea6eb08c6beecbf3802
[ "MIT" ]
47
2015-09-01T19:24:20.000Z
2020-01-02T20:03:05.000Z
from thespian.actors import * from thespian.test import * import time from datetime import timedelta import sys max_response_delay = timedelta(seconds=1.0) class BadMessage(object): def __init__(self, val): self.val = val def __str__(self): return 'Using an invalid member: ' + str(self.this_does_not_exist) class BadMessage2(object): # ok to str() this one def __init__(self, val): self.val = val def __str__(self): return 'BadMsg2=' + str(self.val) class MyActor(Actor): def __init__(self): self.count = 0 def receiveMessage(self, msg, sender): self.count += 1 if isinstance(msg, (BadMessage, BadMessage2)) and (self.count & 1): raise Exception('Got a BadMessage: ' + str(msg)) if not isinstance(msg, ActorSystemMessage): self.send(sender, str(msg)) def test01_actorWorks(asys): mya = asys.createActor(MyActor) r = asys.ask(mya, 123, max_response_delay) assert r == '123' def test02_alwaysBad(asys): mya = asys.createActor(MyActor) r = asys.ask(mya, BadMessage(135), max_response_delay) assert isinstance(r, PoisonMessage) assert isinstance(r.poisonMessage, BadMessage) assert r.poisonMessage.val == 135 def test03_intermittentlyBad(asys): mya = asys.createActor(MyActor) # First one should be OK r = asys.ask(mya, BadMessage2(987), max_response_delay) assert r is not None assert '987' in r # Second one gets the exception the first time around, but the # Actor should be re-instated and the message retried, and it # should work the second time, so the failure is undetectable at # this level. r2 = asys.ask(mya, BadMessage2(654), max_response_delay) assert r2 is not None assert '654' in r2
28.666667
75
0.674419
794d6937e346285ad8b30f9092007b61fdacc7c9
18,573
py
Python
model.py
LeoZDong/shape2prog
2185d1d4eb7a1c4c55e644c6af477fd8e8e70241
[ "BSD-2-Clause" ]
null
null
null
model.py
LeoZDong/shape2prog
2185d1d4eb7a1c4c55e644c6af477fd8e8e70241
[ "BSD-2-Clause" ]
null
null
null
model.py
LeoZDong/shape2prog
2185d1d4eb7a1c4c55e644c6af477fd8e8e70241
[ "BSD-2-Clause" ]
null
null
null
from __future__ import print_function import numpy as np import torch import torch.nn as nn import torch.nn.functional as F from torch.autograd import Variable from misc import render_block class BlockOuterNet(nn.Module): """ predict block-level programs and parameters block-LSTM """ def __init__(self, opt): super(BlockOuterNet, self).__init__() self.feat_size = opt.shape_feat_size self.input_size = opt.outer_input_size self.rnn_size = opt.outer_rnn_size self.num_layers = opt.outer_num_layers self.drop_prob = opt.outer_drop_prob self.seq_length = opt.outer_seq_length self.is_cuda = opt.is_cuda self.shape_feat = Conv3DNet(self.feat_size, input_channel=2, power=2) self.core = nn.LSTM(self.input_size, self.rnn_size, self.num_layers, bias=False, dropout=self.drop_prob) self.inner_net = BlockInnerNet(opt) def init_hidden(self, bsz): weight = next(self.parameters()).data return (Variable(weight.new(self.num_layers, bsz, self.rnn_size).zero_()), Variable(weight.new(self.num_layers, bsz, self.rnn_size).zero_())) def forward(self, x, y, sample_prob=None): batch_size = x.size(0) state = self.init_hidden(batch_size) outputs_pgm = [] outputs_param = [] rendered_shapes = np.zeros((batch_size, 32, 32, 32), dtype=np.uint8) def combine(x, y): y = torch.from_numpy(np.copy(y).astype(np.float32)) y = Variable(torch.unsqueeze(y, 1), requires_grad=False) if self.is_cuda: y = y.cuda() return torch.cat([x, y], dim=1) fc_feats = self.shape_feat(combine(x, rendered_shapes)) seq = y for i in range(seq.size(1)): if i == 0: xt = fc_feats else: prob_pre = torch.exp(outputs_pgm[-1].data) _, it1 = torch.max(prob_pre, dim=2) it2 = outputs_param[-1].data.clone() it1 = it1.cpu().numpy() it2 = it2.cpu().numpy() rendered_shapes = render_block(rendered_shapes, it1, it2) xt = self.shape_feat(combine(x, rendered_shapes)) output, state = self.core(xt.unsqueeze(0), state) output = F.relu(output) pgm, param = self.inner_net(output.squeeze(), y[:, i], sample_prob) outputs_pgm.append(pgm) outputs_param.append(param) return [torch.cat([_.unsqueeze(1) for _ in outputs_pgm], 1).contiguous(), torch.cat([_.unsqueeze(1) for _ in outputs_param], 1).contiguous()] def decode(self, x): batch_size = x.size(0) state = self.init_hidden(batch_size) outputs_pgm = [] outputs_param = [] rendered_shapes = np.zeros((batch_size, 32, 32, 32), dtype=np.uint8) def combine(x, y): y = torch.from_numpy(np.copy(y).astype(np.float32)) y = Variable(torch.unsqueeze(y, 1), requires_grad=False) if self.is_cuda: y = y.cuda() return torch.cat([x, y], dim=1) fc_feats = self.shape_feat(combine(x, rendered_shapes)) for i in range(self.seq_length): if i == 0: xt = fc_feats else: prob_pre = torch.exp(outputs_pgm[-1].data) _, it1 = torch.max(prob_pre, dim=2) it2 = outputs_param[-1].data.clone() it1 = it1.cpu().numpy() it2 = it2.cpu().numpy() rendered_shapes = render_block(rendered_shapes, it1, it2) xt = self.shape_feat(combine(x, rendered_shapes)) output, state = self.core(xt.unsqueeze(0), state) output = F.relu(output) pgm, param = self.inner_net.decode(output.squeeze(0)) outputs_pgm.append(pgm) outputs_param.append(param) return [torch.cat([_.unsqueeze(1) for _ in outputs_pgm], 1).contiguous(), torch.cat([_.unsqueeze(1) for _ in outputs_param], 1).contiguous()] class BlockInnerNet(nn.Module): """ Inner Block Net use last pgm as input for each time step step-LSTM """ def __init__(self, opt): super(BlockInnerNet, self).__init__() self.vocab_size = opt.program_size self.max_param = opt.max_param self.input_size = opt.inner_input_size self.rnn_size = opt.inner_rnn_size self.num_layers = opt.inner_num_layers self.drop_prob = opt.inner_drop_prob self.seq_length = opt.inner_seq_length self.cls_feat_size = opt.inner_cls_feat_size self.reg_feat_size = opt.inner_reg_feat_size self.sample_prob = opt.inner_sample_prob self.is_cuda = opt.is_cuda self.pgm_embed = nn.Embedding(self.vocab_size + 1, self.input_size) self.core = nn.LSTM(self.input_size, self.rnn_size, self.num_layers, bias=False, dropout=self.drop_prob) self.logit1 = nn.Linear(self.rnn_size, self.cls_feat_size) self.logit2 = nn.Linear(self.cls_feat_size, self.vocab_size + 1) self.regress1 = nn.Linear(self.rnn_size, self.reg_feat_size) self.regress2 = nn.Linear(self.vocab_size + 1 + self.reg_feat_size, (self.vocab_size + 1) * self.max_param) self.init_weights() def init_weights(self): initrange = 0.1 self.pgm_embed.weight.data.uniform_(-initrange, initrange) self.logit1.bias.data.fill_(0) self.logit1.weight.data.uniform_(-initrange, initrange) self.logit2.bias.data.fill_(0) self.logit2.weight.data.uniform_(-initrange, initrange) self.regress1.bias.data.fill_(0) self.regress1.weight.data.uniform_(-initrange, initrange) self.regress2.bias.data.fill_(0) self.regress2.weight.data.uniform_(-initrange, initrange) def init_hidden(self, bsz): weight = next(self.parameters()).data return (Variable(weight.new(self.num_layers, bsz, self.rnn_size).zero_()), Variable(weight.new(self.num_layers, bsz, self.rnn_size).zero_())) def forward(self, x, y, sample_prob=None): if sample_prob is not None: self.sample_prob = sample_prob batch_size = x.size(0) state = self.init_hidden(batch_size) outputs_pgm = [] outputs_param = [] seq = y for i in range(seq.size(1)): if i == 0: xt = x else: if i >= 1 and self.sample_prob > 0: sample_prob = x.data.new(batch_size).uniform_(0, 1) sample_mask = sample_prob < self.sample_prob if sample_mask.sum() == 0: it1 = seq[:, i-1].clone() else: sample_ind = sample_mask.nonzero().view(-1) it1 = seq[:, i-1].data.clone() prob_prev = torch.exp(outputs_pgm[-1].data) it1.index_copy_(0, sample_ind, torch.multinomial(prob_prev, 1).view(-1).index_select(0, sample_ind)) it1 = Variable(it1, requires_grad=False) else: print("obtain last ground truth") it1 = seq[:, i-1].clone() xt = self.pgm_embed(it1) output, state = self.core(xt.unsqueeze(0), state) pgm_feat1 = F.relu(self.logit1(output.squeeze(0))) pgm_feat2 = self.logit2(pgm_feat1) pgm_score = F.log_softmax(pgm_feat2, dim=1) trans_prob = F.softmax(pgm_feat2, dim=1).detach() param_feat1 = F.relu(self.regress1(output.squeeze(0))) param_feat2 = torch.cat([trans_prob, param_feat1], dim=1) param_score = self.regress2(param_feat2) param_score = param_score.view(batch_size, self.vocab_size + 1, self.max_param) index = seq[:, i].unsqueeze(1).unsqueeze(2).expand(batch_size, 1, self.max_param).detach() param_score = param_score.gather(1, index).squeeze(1) outputs_pgm.append(pgm_score) outputs_param.append(param_score) return [torch.cat([_.unsqueeze(1) for _ in outputs_pgm], 1).contiguous(), torch.cat([_.unsqueeze(1) for _ in outputs_param], 1).contiguous()] def decode(self, x): batch_size = x.size(0) state = self.init_hidden(batch_size) outputs_pgm = [] outputs_param = [] for i in range(self.seq_length): if i == 0: xt = x else: prob_pre = torch.exp(outputs_pgm[-1].data) _, it1 = torch.max(prob_pre, dim=1) it1 = Variable(it1, requires_grad=False) xt = self.pgm_embed(it1) output, state = self.core(xt.unsqueeze(0), state) pgm_feat1 = F.relu(self.logit1(output.squeeze(0))) pgm_feat2 = self.logit2(pgm_feat1) pgm_score = F.log_softmax(pgm_feat2, dim=1) trans_prob = F.softmax(pgm_feat2, dim=1).detach() param_feat1 = F.relu(self.regress1(output.squeeze(0))) param_feat2 = torch.cat([trans_prob, param_feat1], dim=1) param_score = self.regress2(param_feat2) param_score = param_score.view(batch_size, self.vocab_size + 1, self.max_param) _, index = torch.max(trans_prob, dim=1) index = index.unsqueeze(1).unsqueeze(2).expand(batch_size, 1, self.max_param).detach() param_score = param_score.gather(1, index).squeeze(1) outputs_pgm.append(pgm_score) outputs_param.append(param_score) return [torch.cat([_.unsqueeze(1) for _ in outputs_pgm], 1).contiguous(), torch.cat([_.unsqueeze(1) for _ in outputs_param], 1).contiguous()] class Conv3DNet(nn.Module): """ encode 3D voxelized shape into a vector """ def __init__(self, feat_size, input_channel=1, power=1): super(Conv3DNet, self).__init__() power = int(power) self.conv1 = nn.Conv3d(input_channel, 8*power, kernel_size=(5, 5, 5), stride=(1, 1, 1), padding=(2, 2, 2)) self.conv2 = nn.Conv3d(8*power, 16*power, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1)) self.conv3 = nn.Conv3d(16*power, 16*power, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1)) self.conv4 = nn.Conv3d(16*power, 32*power, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1)) self.conv5 = nn.Conv3d(32*power, 32*power, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1)) self.conv6 = nn.Conv3d(32*power, 64*power, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1)) self.conv7 = nn.Conv3d(64*power, 64*power, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1)) self.conv8 = nn.Conv3d(64*power, 64*power, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1)) # self.BN1 = nn.BatchNorm3d(8*power) # self.BN2 = nn.BatchNorm3d(16*power) # self.BN3 = nn.BatchNorm3d(16*power) # self.BN4 = nn.BatchNorm3d(32*power) # self.BN5 = nn.BatchNorm3d(32*power) # self.BN6 = nn.BatchNorm3d(64*power) # self.BN7 = nn.BatchNorm3d(64*power) # self.BN8 = nn.BatchNorm3d(64*power) self.BN1 = nn.BatchNorm3d(8 * power, track_running_stats=False) self.BN2 = nn.BatchNorm3d(16 * power, track_running_stats=False) self.BN3 = nn.BatchNorm3d(16 * power, track_running_stats=False) self.BN4 = nn.BatchNorm3d(32 * power, track_running_stats=False) self.BN5 = nn.BatchNorm3d(32 * power, track_running_stats=False) self.BN6 = nn.BatchNorm3d(64 * power, track_running_stats=False) self.BN7 = nn.BatchNorm3d(64 * power, track_running_stats=False) self.BN8 = nn.BatchNorm3d(64 * power, track_running_stats=False) self.avgpool = nn.AvgPool3d(kernel_size=(4, 4, 4)) self.fc = nn.Linear(64*power, feat_size) def forward(self, x): x = F.relu(self.BN1(self.conv1(x))) x = F.relu(self.BN2(self.conv2(x))) x = F.relu(self.BN3(self.conv3(x))) x = F.relu(self.BN4(self.conv4(x))) x = F.relu(self.BN5(self.conv5(x))) x = F.relu(self.BN6(self.conv6(x))) x = F.relu(self.BN7(self.conv7(x))) x = F.relu(self.BN8(self.conv8(x))) x = self.avgpool(x) x = x.view(x.size(0), -1) x = F.relu(self.fc(x)) return x class RenderNet(nn.Module): """ Multiple Step Render """ def __init__(self, opt): super(RenderNet, self).__init__() # program LSTM parameter self.vocab_size = opt.program_size self.max_param = opt.max_param self.input_encoding_size = opt.input_encoding_size self.rnn_size = opt.rnn_size self.num_layers = opt.num_layers self.drop_prob_lm = opt.drop_prob_lm self.seq_length = opt.seq_length self.program_vector_size = opt.program_vector_size self.nc = opt.nc self.pgm_embed = nn.Linear(self.vocab_size + 1, int(self.input_encoding_size / 2)) self.param_embed = nn.Linear(self.max_param, self.input_encoding_size - int(self.input_encoding_size / 2)) self.lstm = nn.LSTM(self.input_encoding_size, self.rnn_size, self.num_layers, bias=False, dropout=self.drop_prob_lm) self.pgm_param_feat = nn.Linear(self.rnn_size, self.program_vector_size) self.decoder = nn.Sequential( nn.ConvTranspose3d(self.program_vector_size, 64, 4, 1, 0, bias=False), nn.BatchNorm3d(64), nn.ReLU(True), nn.Conv3d(64, 64, 3, 1, 1, bias=False), nn.BatchNorm3d(64), nn.ReLU(True), # 4 x 4 x 4 nn.ConvTranspose3d(64, 16, 4, 2, 1, bias=False), nn.BatchNorm3d(16), nn.ReLU(True), nn.Conv3d(16, 16, 3, 1, 1, bias=False), nn.BatchNorm3d(16), nn.ReLU(True), # 8 x 8 x 8 nn.ConvTranspose3d(16, 4, 4, 2, 1, bias=False), nn.BatchNorm3d(4), nn.ReLU(True), nn.Conv3d(4, 4, 3, 1, 1, bias=False), nn.BatchNorm3d(4), nn.ReLU(True), # 16 x 16 x 16 nn.ConvTranspose3d(4, self.nc, 4, 2, 1, bias=False), ) def init_hidden(self, bsz): weight = next(self.parameters()).data return (Variable(weight.new(self.num_layers, bsz, self.rnn_size).zero_()), Variable(weight.new(self.num_layers, bsz, self.rnn_size).zero_())) def forward(self, program, parameters, index): program = program.permute(1, 0, 2) parameters = parameters.permute(1, 0, 2) bsz = program.size(1) init = self.init_hidden(bsz) # program linear transform dim1 = program.size() program = program.contiguous().view(-1, self.vocab_size + 1) x1 = F.relu(self.pgm_embed(program)) x1 = x1.view(dim1[0], dim1[1], -1) # parameter linear transform dim2 = parameters.size() parameters = parameters.contiguous().view(-1, self.max_param) x2 = F.relu(self.param_embed(parameters)) x2 = x2.view(dim2[0], dim2[1], -1) # LSTM to aggregate programs and parameters x = torch.cat((x1, x2), dim=2) out, hidden = self.lstm(x, init) # select desired step aggregated features index = index.unsqueeze(1).expand(-1, out.size(2)).unsqueeze(0) pgm_param_feat = out.gather(dim=0, index=index).squeeze() pgm_param_feat = F.relu(self.pgm_param_feat(pgm_param_feat)) pgm_param_feat = pgm_param_feat.view(bsz, self.program_vector_size, 1, 1, 1) shape = self.decoder(pgm_param_feat) return shape def compute_LSTM_feat(self, program, parameters, index): program = program.permute(1, 0, 2) parameters = parameters.permute(1, 0, 2) bsz = program.size(1) init = self.init_hidden(bsz) # program linear transform dim1 = program.size() program = program.contiguous().view(-1, self.vocab_size + 1) x1 = F.relu(self.pgm_embed(program)) x1 = x1.view(dim1[0], dim1[1], -1) # parameter linear transform dim2 = parameters.size() parameters = parameters.contiguous().view(-1, self.max_param) x2 = F.relu(self.param_embed(parameters)) x2 = x2.view(dim2[0], dim2[1], -1) # LSTM to aggregate programs and parameters x = torch.cat((x1, x2), dim=2) out, hidden = self.lstm(x, init) # select desired step aggregated features index = index.unsqueeze(1).expand(-1, out.size(2)).unsqueeze(0) pgm_param_feat = out.gather(dim=0, index=index).squeeze() # pgm_param_feat = F.relu(self.pgm_param_feat(pgm_param_feat)) return pgm_param_feat def compute_shape_from_feat(self, pgm_param_feat): bsz = pgm_param_feat.size(0) pgm_param_feat = F.relu(self.pgm_param_feat(pgm_param_feat)) pgm_param_feat = pgm_param_feat.view(bsz, self.program_vector_size, 1, 1, 1) shape = self.decoder(pgm_param_feat) return shape if __name__ == '__main__': from easydict import EasyDict as edict from programs.label_config import stop_id, max_param opt = edict() opt.shape_feat_size = 64 opt.outer_input_size = 64 opt.outer_rnn_size = 64 opt.outer_num_layers = 1 opt.outer_drop_prob = 0 opt.outer_seq_length = 6 opt.is_cuda = False opt.program_size = stop_id - 1 opt.max_param = max_param - 1 opt.inner_input_size = 64 opt.inner_rnn_size = 64 opt.inner_num_layers = 1 opt.inner_drop_prob = 0 opt.inner_seq_length = 3 opt.inner_cls_feat_size = 64 opt.inner_reg_feat_size = 64 opt.inner_sample_prob = 1.0 net = BlockOuterNet(opt) x = torch.zeros((10, 1, 32, 32, 32)) x = Variable(x, requires_grad=False) y = torch.zeros((10, 6, 3)).long() pgms, params = net(x, y) print(pgms.size()) print(params.size()) pgms, params = net.decode(x) print(pgms.size()) print(params.size())
38.69375
124
0.590481
794d6a00080275d8994bb3a434bf080e3d226882
397
py
Python
upperegypt/wsgi.py
mohammedmosaad400/django-upperegypt
ba7c236ad586ff9666d8b4145d249216b2f307fb
[ "MIT" ]
null
null
null
upperegypt/wsgi.py
mohammedmosaad400/django-upperegypt
ba7c236ad586ff9666d8b4145d249216b2f307fb
[ "MIT" ]
null
null
null
upperegypt/wsgi.py
mohammedmosaad400/django-upperegypt
ba7c236ad586ff9666d8b4145d249216b2f307fb
[ "MIT" ]
null
null
null
""" WSGI config for upperegypt project. It exposes the WSGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/2.2/howto/deployment/wsgi/ """ import os from django.core.wsgi import get_wsgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'upperegypt.settings') application = get_wsgi_application()
23.352941
78
0.788413
794d6a33ae57b2bf445c129cadbf4ddbca484572
31,515
py
Python
sdk/compute/azure-mgmt-compute/azure/mgmt/compute/v2020_12_01/models/_compute_management_client_enums.py
benbp/azure-sdk-for-python
2329ba03e48098dcdc581898f6434d7c2b13a7b9
[ "MIT" ]
null
null
null
sdk/compute/azure-mgmt-compute/azure/mgmt/compute/v2020_12_01/models/_compute_management_client_enums.py
benbp/azure-sdk-for-python
2329ba03e48098dcdc581898f6434d7c2b13a7b9
[ "MIT" ]
null
null
null
sdk/compute/azure-mgmt-compute/azure/mgmt/compute/v2020_12_01/models/_compute_management_client_enums.py
benbp/azure-sdk-for-python
2329ba03e48098dcdc581898f6434d7c2b13a7b9
[ "MIT" ]
null
null
null
# coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for license information. # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- from enum import Enum, EnumMeta from six import with_metaclass class _CaseInsensitiveEnumMeta(EnumMeta): def __getitem__(self, name): return super().__getitem__(name.upper()) def __getattr__(cls, name): """Return the enum member matching `name` We use __getattr__ instead of descriptors or inserting into the enum class' __dict__ in order to support `name` and `value` being both properties for enum members (which live in the class' __dict__) and enum members themselves. """ try: return cls._member_map_[name.upper()] except KeyError: raise AttributeError(name) class AccessLevel(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): NONE = "None" READ = "Read" WRITE = "Write" class AvailabilitySetSkuTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the sku of an Availability Set. Use 'Aligned' for virtual machines with managed disks and 'Classic' for virtual machines with unmanaged disks. Default value is 'Classic'. """ CLASSIC = "Classic" ALIGNED = "Aligned" class CachingTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the caching requirements. :code:`<br>`:code:`<br>` Possible values are: :code:`<br>`:code:`<br>` **None** :code:`<br>`:code:`<br>` **ReadOnly** :code:`<br>`:code:`<br>` **ReadWrite** :code:`<br>`:code:`<br>` Default: **None for Standard storage. ReadOnly for Premium storage** """ NONE = "None" READ_ONLY = "ReadOnly" READ_WRITE = "ReadWrite" class DedicatedHostLicenseTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the software license type that will be applied to the VMs deployed on the dedicated host. :code:`<br>`:code:`<br>` Possible values are: :code:`<br>`:code:`<br>` **None** :code:`<br>`:code:`<br>` **Windows_Server_Hybrid** :code:`<br>`:code:`<br>` **Windows_Server_Perpetual** :code:`<br>`:code:`<br>` Default: **None** """ NONE = "None" WINDOWS_SERVER_HYBRID = "Windows_Server_Hybrid" WINDOWS_SERVER_PERPETUAL = "Windows_Server_Perpetual" class DiffDiskOptions(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the ephemeral disk option for operating system disk. """ LOCAL = "Local" class DiffDiskPlacement(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the ephemeral disk placement for operating system disk. This property can be used by user in the request to choose the location i.e, cache disk or resource disk space for Ephemeral OS disk provisioning. For more information on Ephemeral OS disk size requirements, please refer Ephemeral OS disk size requirements for Windows VM at https://docs.microsoft.com/en- us/azure/virtual-machines/windows/ephemeral-os-disks#size-requirements and Linux VM at https://docs.microsoft.com/en-us/azure/virtual-machines/linux/ephemeral-os-disks#size- requirements """ CACHE_DISK = "CacheDisk" RESOURCE_DISK = "ResourceDisk" class DiskCreateOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """This enumerates the possible sources of a disk's creation. """ EMPTY = "Empty" #: Create an empty data disk of a size given by diskSizeGB. ATTACH = "Attach" #: Disk will be attached to a VM. FROM_IMAGE = "FromImage" #: Create a new disk from a platform image specified by the given imageReference or galleryImageReference. IMPORT_ENUM = "Import" #: Create a disk by importing from a blob specified by a sourceUri in a storage account specified by storageAccountId. COPY = "Copy" #: Create a new disk or snapshot by copying from a disk or snapshot specified by the given sourceResourceId. RESTORE = "Restore" #: Create a new disk by copying from a backup recovery point. UPLOAD = "Upload" #: Create a new disk by obtaining a write token and using it to directly upload the contents of the disk. class DiskCreateOptionTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies how the virtual machine should be created.:code:`<br>`:code:`<br>` Possible values are::code:`<br>`:code:`<br>` **Attach** \u2013 This value is used when you are using a specialized disk to create the virtual machine.:code:`<br>`:code:`<br>` **FromImage** \u2013 This value is used when you are using an image to create the virtual machine. If you are using a platform image, you also use the imageReference element described above. If you are using a marketplace image, you also use the plan element previously described. """ FROM_IMAGE = "FromImage" EMPTY = "Empty" ATTACH = "Attach" class DiskDetachOptionTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the detach behavior to be used while detaching a disk or which is already in the process of detachment from the virtual machine. Supported values: **ForceDetach**. :code:`<br>`:code:`<br>` detachOption: **ForceDetach** is applicable only for managed data disks. If a previous detachment attempt of the data disk did not complete due to an unexpected failure from the virtual machine and the disk is still not released then use force-detach as a last resort option to detach the disk forcibly from the VM. All writes might not have been flushed when using this detach behavior. :code:`<br>`:code:`<br>` This feature is still in preview mode and is not supported for VirtualMachineScaleSet. To force-detach a data disk update toBeDetached to 'true' along with setting detachOption: 'ForceDetach'. """ FORCE_DETACH = "ForceDetach" class DiskEncryptionSetIdentityType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The type of Managed Identity used by the DiskEncryptionSet. Only SystemAssigned is supported for new creations. Disk Encryption Sets can be updated with Identity type None during migration of subscription to a new Azure Active Directory tenant; it will cause the encrypted resources to lose access to the keys. """ SYSTEM_ASSIGNED = "SystemAssigned" NONE = "None" class DiskEncryptionSetType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The type of key used to encrypt the data of the disk. """ ENCRYPTION_AT_REST_WITH_CUSTOMER_KEY = "EncryptionAtRestWithCustomerKey" #: Resource using diskEncryptionSet would be encrypted at rest with Customer managed key that can be changed and revoked by a customer. ENCRYPTION_AT_REST_WITH_PLATFORM_AND_CUSTOMER_KEYS = "EncryptionAtRestWithPlatformAndCustomerKeys" #: Resource using diskEncryptionSet would be encrypted at rest with two layers of encryption. One of the keys is Customer managed and the other key is Platform managed. class DiskSecurityTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the SecurityType of the VM. Applicable for OS disks only. """ TRUSTED_LAUNCH = "TrustedLaunch" #: Trusted Launch provides security features such as secure boot and virtual Trusted Platform Module (vTPM). class DiskState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """This enumerates the possible state of the disk. """ UNATTACHED = "Unattached" #: The disk is not being used and can be attached to a VM. ATTACHED = "Attached" #: The disk is currently mounted to a running VM. RESERVED = "Reserved" #: The disk is mounted to a stopped-deallocated VM. ACTIVE_SAS = "ActiveSAS" #: The disk currently has an Active SAS Uri associated with it. READY_TO_UPLOAD = "ReadyToUpload" #: A disk is ready to be created by upload by requesting a write token. ACTIVE_UPLOAD = "ActiveUpload" #: A disk is created for upload and a write token has been issued for uploading to it. class DiskStorageAccountTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The sku name. """ STANDARD_LRS = "Standard_LRS" #: Standard HDD locally redundant storage. Best for backup, non-critical, and infrequent access. PREMIUM_LRS = "Premium_LRS" #: Premium SSD locally redundant storage. Best for production and performance sensitive workloads. STANDARD_SSD_LRS = "StandardSSD_LRS" #: Standard SSD locally redundant storage. Best for web servers, lightly used enterprise applications and dev/test. ULTRA_SSD_LRS = "UltraSSD_LRS" #: Ultra SSD locally redundant storage. Best for IO-intensive workloads such as SAP HANA, top tier databases (for example, SQL, Oracle), and other transaction-heavy workloads. PREMIUM_ZRS = "Premium_ZRS" #: Premium SSD zone redundant storage. Best for the production workloads that need storage resiliency against zone failures. STANDARD_SSD_ZRS = "StandardSSD_ZRS" #: Standard SSD zone redundant storage. Best for web servers, lightly used enterprise applications and dev/test that need storage resiliency against zone failures. class EncryptionType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The type of key used to encrypt the data of the disk. """ ENCRYPTION_AT_REST_WITH_PLATFORM_KEY = "EncryptionAtRestWithPlatformKey" #: Disk is encrypted at rest with Platform managed key. It is the default encryption type. This is not a valid encryption type for disk encryption sets. ENCRYPTION_AT_REST_WITH_CUSTOMER_KEY = "EncryptionAtRestWithCustomerKey" #: Disk is encrypted at rest with Customer managed key that can be changed and revoked by a customer. ENCRYPTION_AT_REST_WITH_PLATFORM_AND_CUSTOMER_KEYS = "EncryptionAtRestWithPlatformAndCustomerKeys" #: Disk is encrypted at rest with 2 layers of encryption. One of the keys is Customer managed and the other key is Platform managed. class ExecutionState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Script execution status. """ UNKNOWN = "Unknown" PENDING = "Pending" RUNNING = "Running" FAILED = "Failed" SUCCEEDED = "Succeeded" TIMED_OUT = "TimedOut" CANCELED = "Canceled" class ExtendedLocationTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The type of extendedLocation. """ EDGE_ZONE = "EdgeZone" class HyperVGeneration(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The hypervisor generation of the Virtual Machine. Applicable to OS disks only. """ V1 = "V1" V2 = "V2" class HyperVGenerationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the HyperVGeneration Type associated with a resource """ V1 = "V1" V2 = "V2" class HyperVGenerationTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the HyperVGeneration Type """ V1 = "V1" V2 = "V2" class IntervalInMins(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Interval value in minutes used to create LogAnalytics call rate logs. """ THREE_MINS = "ThreeMins" FIVE_MINS = "FiveMins" THIRTY_MINS = "ThirtyMins" SIXTY_MINS = "SixtyMins" class IPVersion(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Available from Api-Version 2017-03-30 onwards, it represents whether the specific ipconfiguration is IPv4 or IPv6. Default is taken as IPv4. Possible values are: 'IPv4' and 'IPv6'. """ I_PV4 = "IPv4" I_PV6 = "IPv6" class LinuxVMGuestPatchMode(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the mode of VM Guest Patching to IaaS virtual machine.:code:`<br />`:code:`<br />` Possible values are::code:`<br />`:code:`<br />` **ImageDefault** - The virtual machine's default patching configuration is used. :code:`<br />`:code:`<br />` **AutomaticByPlatform** - The virtual machine will be automatically updated by the platform. The property provisionVMAgent must be true """ IMAGE_DEFAULT = "ImageDefault" AUTOMATIC_BY_PLATFORM = "AutomaticByPlatform" class MaintenanceOperationResultCodeTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The Last Maintenance Operation Result Code. """ NONE = "None" RETRY_LATER = "RetryLater" MAINTENANCE_ABORTED = "MaintenanceAborted" MAINTENANCE_COMPLETED = "MaintenanceCompleted" class NetworkAccessPolicy(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Policy for accessing the disk via network. """ ALLOW_ALL = "AllowAll" #: The disk can be exported or uploaded to from any network. ALLOW_PRIVATE = "AllowPrivate" #: The disk can be exported or uploaded to using a DiskAccess resource's private endpoints. DENY_ALL = "DenyAll" #: The disk cannot be exported. class OperatingSystemStateTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The OS State. """ GENERALIZED = "Generalized" #: Generalized image. Needs to be provisioned during deployment time. SPECIALIZED = "Specialized" #: Specialized image. Contains already provisioned OS Disk. class OperatingSystemTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The Operating System type. """ WINDOWS = "Windows" LINUX = "Linux" class OrchestrationMode(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the orchestration mode for the virtual machine scale set. """ UNIFORM = "Uniform" FLEXIBLE = "Flexible" class OrchestrationServiceNames(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The name of the service. """ AUTOMATIC_REPAIRS = "AutomaticRepairs" DUMMY_ORCHESTRATION_SERVICE_NAME = "DummyOrchestrationServiceName" class OrchestrationServiceState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The current state of the service. """ NOT_RUNNING = "NotRunning" RUNNING = "Running" SUSPENDED = "Suspended" class OrchestrationServiceStateAction(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The action to be performed. """ RESUME = "Resume" SUSPEND = "Suspend" class PatchAssessmentState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Describes the availability of a given patch. """ UNKNOWN = "Unknown" AVAILABLE = "Available" class PatchInstallationState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The state of the patch after the installation operation completed. """ UNKNOWN = "Unknown" INSTALLED = "Installed" FAILED = "Failed" EXCLUDED = "Excluded" NOT_SELECTED = "NotSelected" PENDING = "Pending" class PatchOperationStatus(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The overall success or failure status of the operation. It remains "InProgress" until the operation completes. At that point it will become "Unknown", "Failed", "Succeeded", or "CompletedWithWarnings." """ UNKNOWN = "Unknown" IN_PROGRESS = "InProgress" FAILED = "Failed" SUCCEEDED = "Succeeded" COMPLETED_WITH_WARNINGS = "CompletedWithWarnings" class PrivateEndpointConnectionProvisioningState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The current provisioning state. """ SUCCEEDED = "Succeeded" CREATING = "Creating" DELETING = "Deleting" FAILED = "Failed" class PrivateEndpointServiceConnectionStatus(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The private endpoint connection status. """ PENDING = "Pending" APPROVED = "Approved" REJECTED = "Rejected" class ProtocolTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the protocol of WinRM listener. :code:`<br>`:code:`<br>` Possible values are: :code:`<br>`\ **http** :code:`<br>`:code:`<br>` **https** """ HTTP = "Http" HTTPS = "Https" class ProximityPlacementGroupType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the type of the proximity placement group. :code:`<br>`:code:`<br>` Possible values are: :code:`<br>`:code:`<br>` **Standard** : Co-locate resources within an Azure region or Availability Zone. :code:`<br>`:code:`<br>` **Ultra** : For future use. """ STANDARD = "Standard" ULTRA = "Ultra" class ResourceIdentityType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The type of identity used for the virtual machine. The type 'SystemAssigned, UserAssigned' includes both an implicitly created identity and a set of user assigned identities. The type 'None' will remove any identities from the virtual machine. """ SYSTEM_ASSIGNED = "SystemAssigned" USER_ASSIGNED = "UserAssigned" SYSTEM_ASSIGNED_USER_ASSIGNED = "SystemAssigned, UserAssigned" NONE = "None" class RollingUpgradeActionType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The last action performed on the rolling upgrade. """ START = "Start" CANCEL = "Cancel" class RollingUpgradeStatusCode(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Code indicating the current status of the upgrade. """ ROLLING_FORWARD = "RollingForward" CANCELLED = "Cancelled" COMPLETED = "Completed" FAULTED = "Faulted" class SettingNames(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the name of the setting to which the content applies. Possible values are: FirstLogonCommands and AutoLogon. """ AUTO_LOGON = "AutoLogon" FIRST_LOGON_COMMANDS = "FirstLogonCommands" class SnapshotStorageAccountTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The sku name. """ STANDARD_LRS = "Standard_LRS" #: Standard HDD locally redundant storage. PREMIUM_LRS = "Premium_LRS" #: Premium SSD locally redundant storage. STANDARD_ZRS = "Standard_ZRS" #: Standard zone redundant storage. class StatusLevelTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The level code. """ INFO = "Info" WARNING = "Warning" ERROR = "Error" class StorageAccountTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the storage account type for the managed disk. Managed OS disk storage account type can only be set when you create the scale set. NOTE: UltraSSD_LRS can only be used with data disks. It cannot be used with OS Disk. Standard_LRS uses Standard HDD. StandardSSD_LRS uses Standard SSD. Premium_LRS uses Premium SSD. UltraSSD_LRS uses Ultra disk. Premium_ZRS uses Premium SSD zone redundant storage. StandardSSD_ZRS uses Standard SSD zone redundant storage. For more information regarding disks supported for Windows Virtual Machines, refer to https://docs.microsoft.com/en-us/azure/virtual-machines/windows/disks-types and, for Linux Virtual Machines, refer to https://docs.microsoft.com/en-us/azure/virtual-machines/linux/disks- types """ STANDARD_LRS = "Standard_LRS" PREMIUM_LRS = "Premium_LRS" STANDARD_SSD_LRS = "StandardSSD_LRS" ULTRA_SSD_LRS = "UltraSSD_LRS" PREMIUM_ZRS = "Premium_ZRS" STANDARD_SSD_ZRS = "StandardSSD_ZRS" class UpgradeMode(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the mode of an upgrade to virtual machines in the scale set.:code:`<br />`:code:`<br />` Possible values are::code:`<br />`:code:`<br />` **Manual** - You control the application of updates to virtual machines in the scale set. You do this by using the manualUpgrade action.:code:`<br />`:code:`<br />` **Automatic** - All virtual machines in the scale set are automatically updated at the same time. """ AUTOMATIC = "Automatic" MANUAL = "Manual" ROLLING = "Rolling" class UpgradeOperationInvoker(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Invoker of the Upgrade Operation """ UNKNOWN = "Unknown" USER = "User" PLATFORM = "Platform" class UpgradeState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Code indicating the current status of the upgrade. """ ROLLING_FORWARD = "RollingForward" CANCELLED = "Cancelled" COMPLETED = "Completed" FAULTED = "Faulted" class VirtualMachineEvictionPolicyTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the eviction policy for the Azure Spot VM/VMSS """ DEALLOCATE = "Deallocate" DELETE = "Delete" class VirtualMachinePriorityTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the priority for a standalone virtual machine or the virtual machines in the scale set. :code:`<br>`:code:`<br>` 'Low' enum will be deprecated in the future, please use 'Spot' as the enum to deploy Azure Spot VM/VMSS. """ REGULAR = "Regular" LOW = "Low" SPOT = "Spot" class VirtualMachineScaleSetScaleInRules(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): DEFAULT = "Default" OLDEST_VM = "OldestVM" NEWEST_VM = "NewestVM" class VirtualMachineScaleSetSkuScaleType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The scale type applicable to the sku. """ AUTOMATIC = "Automatic" NONE = "None" class VirtualMachineSizeTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the size of the virtual machine. :code:`<br>`:code:`<br>` The enum data type is currently deprecated and will be removed by December 23rd 2023. :code:`<br>`:code:`<br>` Recommended way to get the list of available sizes is using these APIs: :code:`<br>`:code:`<br>` `List all available virtual machine sizes in an availability set <https://docs.microsoft.com/rest/api/compute/availabilitysets/listavailablesizes>`_ :code:`<br>`:code:`<br>` `List all available virtual machine sizes in a region <https://docs.microsoft.com/en-us/rest/api/compute/resourceskus/list>`_ :code:`<br>`:code:`<br>` `List all available virtual machine sizes for resizing <https://docs.microsoft.com/rest/api/compute/virtualmachines/listavailablesizes>`_. For more information about virtual machine sizes, see `Sizes for virtual machines <https://docs.microsoft.com/en-us/azure/virtual-machines/sizes>`_. :code:`<br>`:code:`<br>` The available VM sizes depend on region and availability set. """ BASIC_A0 = "Basic_A0" BASIC_A1 = "Basic_A1" BASIC_A2 = "Basic_A2" BASIC_A3 = "Basic_A3" BASIC_A4 = "Basic_A4" STANDARD_A0 = "Standard_A0" STANDARD_A1 = "Standard_A1" STANDARD_A2 = "Standard_A2" STANDARD_A3 = "Standard_A3" STANDARD_A4 = "Standard_A4" STANDARD_A5 = "Standard_A5" STANDARD_A6 = "Standard_A6" STANDARD_A7 = "Standard_A7" STANDARD_A8 = "Standard_A8" STANDARD_A9 = "Standard_A9" STANDARD_A10 = "Standard_A10" STANDARD_A11 = "Standard_A11" STANDARD_A1_V2 = "Standard_A1_v2" STANDARD_A2_V2 = "Standard_A2_v2" STANDARD_A4_V2 = "Standard_A4_v2" STANDARD_A8_V2 = "Standard_A8_v2" STANDARD_A2_M_V2 = "Standard_A2m_v2" STANDARD_A4_M_V2 = "Standard_A4m_v2" STANDARD_A8_M_V2 = "Standard_A8m_v2" STANDARD_B1_S = "Standard_B1s" STANDARD_B1_MS = "Standard_B1ms" STANDARD_B2_S = "Standard_B2s" STANDARD_B2_MS = "Standard_B2ms" STANDARD_B4_MS = "Standard_B4ms" STANDARD_B8_MS = "Standard_B8ms" STANDARD_D1 = "Standard_D1" STANDARD_D2 = "Standard_D2" STANDARD_D3 = "Standard_D3" STANDARD_D4 = "Standard_D4" STANDARD_D11 = "Standard_D11" STANDARD_D12 = "Standard_D12" STANDARD_D13 = "Standard_D13" STANDARD_D14 = "Standard_D14" STANDARD_D1_V2 = "Standard_D1_v2" STANDARD_D2_V2 = "Standard_D2_v2" STANDARD_D3_V2 = "Standard_D3_v2" STANDARD_D4_V2 = "Standard_D4_v2" STANDARD_D5_V2 = "Standard_D5_v2" STANDARD_D2_V3 = "Standard_D2_v3" STANDARD_D4_V3 = "Standard_D4_v3" STANDARD_D8_V3 = "Standard_D8_v3" STANDARD_D16_V3 = "Standard_D16_v3" STANDARD_D32_V3 = "Standard_D32_v3" STANDARD_D64_V3 = "Standard_D64_v3" STANDARD_D2_S_V3 = "Standard_D2s_v3" STANDARD_D4_S_V3 = "Standard_D4s_v3" STANDARD_D8_S_V3 = "Standard_D8s_v3" STANDARD_D16_S_V3 = "Standard_D16s_v3" STANDARD_D32_S_V3 = "Standard_D32s_v3" STANDARD_D64_S_V3 = "Standard_D64s_v3" STANDARD_D11_V2 = "Standard_D11_v2" STANDARD_D12_V2 = "Standard_D12_v2" STANDARD_D13_V2 = "Standard_D13_v2" STANDARD_D14_V2 = "Standard_D14_v2" STANDARD_D15_V2 = "Standard_D15_v2" STANDARD_DS1 = "Standard_DS1" STANDARD_DS2 = "Standard_DS2" STANDARD_DS3 = "Standard_DS3" STANDARD_DS4 = "Standard_DS4" STANDARD_DS11 = "Standard_DS11" STANDARD_DS12 = "Standard_DS12" STANDARD_DS13 = "Standard_DS13" STANDARD_DS14 = "Standard_DS14" STANDARD_DS1_V2 = "Standard_DS1_v2" STANDARD_DS2_V2 = "Standard_DS2_v2" STANDARD_DS3_V2 = "Standard_DS3_v2" STANDARD_DS4_V2 = "Standard_DS4_v2" STANDARD_DS5_V2 = "Standard_DS5_v2" STANDARD_DS11_V2 = "Standard_DS11_v2" STANDARD_DS12_V2 = "Standard_DS12_v2" STANDARD_DS13_V2 = "Standard_DS13_v2" STANDARD_DS14_V2 = "Standard_DS14_v2" STANDARD_DS15_V2 = "Standard_DS15_v2" STANDARD_DS13_4_V2 = "Standard_DS13-4_v2" STANDARD_DS13_2_V2 = "Standard_DS13-2_v2" STANDARD_DS14_8_V2 = "Standard_DS14-8_v2" STANDARD_DS14_4_V2 = "Standard_DS14-4_v2" STANDARD_E2_V3 = "Standard_E2_v3" STANDARD_E4_V3 = "Standard_E4_v3" STANDARD_E8_V3 = "Standard_E8_v3" STANDARD_E16_V3 = "Standard_E16_v3" STANDARD_E32_V3 = "Standard_E32_v3" STANDARD_E64_V3 = "Standard_E64_v3" STANDARD_E2_S_V3 = "Standard_E2s_v3" STANDARD_E4_S_V3 = "Standard_E4s_v3" STANDARD_E8_S_V3 = "Standard_E8s_v3" STANDARD_E16_S_V3 = "Standard_E16s_v3" STANDARD_E32_S_V3 = "Standard_E32s_v3" STANDARD_E64_S_V3 = "Standard_E64s_v3" STANDARD_E32_16_V3 = "Standard_E32-16_v3" STANDARD_E32_8_S_V3 = "Standard_E32-8s_v3" STANDARD_E64_32_S_V3 = "Standard_E64-32s_v3" STANDARD_E64_16_S_V3 = "Standard_E64-16s_v3" STANDARD_F1 = "Standard_F1" STANDARD_F2 = "Standard_F2" STANDARD_F4 = "Standard_F4" STANDARD_F8 = "Standard_F8" STANDARD_F16 = "Standard_F16" STANDARD_F1_S = "Standard_F1s" STANDARD_F2_S = "Standard_F2s" STANDARD_F4_S = "Standard_F4s" STANDARD_F8_S = "Standard_F8s" STANDARD_F16_S = "Standard_F16s" STANDARD_F2_S_V2 = "Standard_F2s_v2" STANDARD_F4_S_V2 = "Standard_F4s_v2" STANDARD_F8_S_V2 = "Standard_F8s_v2" STANDARD_F16_S_V2 = "Standard_F16s_v2" STANDARD_F32_S_V2 = "Standard_F32s_v2" STANDARD_F64_S_V2 = "Standard_F64s_v2" STANDARD_F72_S_V2 = "Standard_F72s_v2" STANDARD_G1 = "Standard_G1" STANDARD_G2 = "Standard_G2" STANDARD_G3 = "Standard_G3" STANDARD_G4 = "Standard_G4" STANDARD_G5 = "Standard_G5" STANDARD_GS1 = "Standard_GS1" STANDARD_GS2 = "Standard_GS2" STANDARD_GS3 = "Standard_GS3" STANDARD_GS4 = "Standard_GS4" STANDARD_GS5 = "Standard_GS5" STANDARD_GS4_8 = "Standard_GS4-8" STANDARD_GS4_4 = "Standard_GS4-4" STANDARD_GS5_16 = "Standard_GS5-16" STANDARD_GS5_8 = "Standard_GS5-8" STANDARD_H8 = "Standard_H8" STANDARD_H16 = "Standard_H16" STANDARD_H8_M = "Standard_H8m" STANDARD_H16_M = "Standard_H16m" STANDARD_H16_R = "Standard_H16r" STANDARD_H16_MR = "Standard_H16mr" STANDARD_L4_S = "Standard_L4s" STANDARD_L8_S = "Standard_L8s" STANDARD_L16_S = "Standard_L16s" STANDARD_L32_S = "Standard_L32s" STANDARD_M64_S = "Standard_M64s" STANDARD_M64_MS = "Standard_M64ms" STANDARD_M128_S = "Standard_M128s" STANDARD_M128_MS = "Standard_M128ms" STANDARD_M64_32_MS = "Standard_M64-32ms" STANDARD_M64_16_MS = "Standard_M64-16ms" STANDARD_M128_64_MS = "Standard_M128-64ms" STANDARD_M128_32_MS = "Standard_M128-32ms" STANDARD_NC6 = "Standard_NC6" STANDARD_NC12 = "Standard_NC12" STANDARD_NC24 = "Standard_NC24" STANDARD_NC24_R = "Standard_NC24r" STANDARD_NC6_S_V2 = "Standard_NC6s_v2" STANDARD_NC12_S_V2 = "Standard_NC12s_v2" STANDARD_NC24_S_V2 = "Standard_NC24s_v2" STANDARD_NC24_RS_V2 = "Standard_NC24rs_v2" STANDARD_NC6_S_V3 = "Standard_NC6s_v3" STANDARD_NC12_S_V3 = "Standard_NC12s_v3" STANDARD_NC24_S_V3 = "Standard_NC24s_v3" STANDARD_NC24_RS_V3 = "Standard_NC24rs_v3" STANDARD_ND6_S = "Standard_ND6s" STANDARD_ND12_S = "Standard_ND12s" STANDARD_ND24_S = "Standard_ND24s" STANDARD_ND24_RS = "Standard_ND24rs" STANDARD_NV6 = "Standard_NV6" STANDARD_NV12 = "Standard_NV12" STANDARD_NV24 = "Standard_NV24" class VmDiskTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """VM disk types which are disallowed. """ NONE = "None" UNMANAGED = "Unmanaged" class VMGuestPatchClassificationLinux(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): CRITICAL = "Critical" SECURITY = "Security" OTHER = "Other" class VMGuestPatchClassificationWindows(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): CRITICAL = "Critical" SECURITY = "Security" UPDATE_ROLL_UP = "UpdateRollUp" FEATURE_PACK = "FeaturePack" SERVICE_PACK = "ServicePack" DEFINITION = "Definition" TOOLS = "Tools" UPDATES = "Updates" class VMGuestPatchRebootBehavior(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Describes the reboot requirements of the patch. """ UNKNOWN = "Unknown" NEVER_REBOOTS = "NeverReboots" ALWAYS_REQUIRES_REBOOT = "AlwaysRequiresReboot" CAN_REQUEST_REBOOT = "CanRequestReboot" class VMGuestPatchRebootSetting(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Defines when it is acceptable to reboot a VM during a software update operation. """ IF_REQUIRED = "IfRequired" NEVER = "Never" ALWAYS = "Always" class VMGuestPatchRebootStatus(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The reboot state of the VM following completion of the operation. """ UNKNOWN = "Unknown" NOT_NEEDED = "NotNeeded" REQUIRED = "Required" STARTED = "Started" FAILED = "Failed" COMPLETED = "Completed" class WindowsVMGuestPatchMode(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Specifies the mode of VM Guest Patching to IaaS virtual machine.:code:`<br />`:code:`<br />` Possible values are::code:`<br />`:code:`<br />` **Manual** - You control the application of patches to a virtual machine. You do this by applying patches manually inside the VM. In this mode, automatic updates are disabled; the property WindowsConfiguration.enableAutomaticUpdates must be false:code:`<br />`:code:`<br />` **AutomaticByOS** - The virtual machine will automatically be updated by the OS. The property WindowsConfiguration.enableAutomaticUpdates must be true. :code:`<br />`:code:`<br />` **AutomaticByPlatform** - the virtual machine will automatically updated by the platform. The properties provisionVMAgent and WindowsConfiguration.enableAutomaticUpdates must be true """ MANUAL = "Manual" AUTOMATIC_BY_OS = "AutomaticByOS" AUTOMATIC_BY_PLATFORM = "AutomaticByPlatform"
42.245308
272
0.725115
794d6a8a6b9cb779daaf374abfd545158917f0dd
335
py
Python
intro-python/dnac_config_cloud.py
anaid30kadi/DevNet
463af310b88d131f3a26df313d6ca29071528953
[ "MIT" ]
null
null
null
intro-python/dnac_config_cloud.py
anaid30kadi/DevNet
463af310b88d131f3a26df313d6ca29071528953
[ "MIT" ]
8
2021-04-07T23:21:09.000Z
2021-09-23T23:29:33.000Z
intro-python/dnac_config_cloud.py
anaid30kadi/DevNet
463af310b88d131f3a26df313d6ca29071528953
[ "MIT" ]
null
null
null
# Step 1 # Change DNAC IP to the one you are using DNAC_IP = "sandboxdnac.cisco.com" DNAC_PORT = 443 # Step 2 # Enter user name and password to get a service ticket # If you assign username, password, and version here, # you don't need to pass parameter when calling DNAC_USER = "devnetuser" DNAC_PASSWORD = "Cisco123!" VERSION = "v1"
27.916667
54
0.743284
794d6b6f7dad9605c29281c02ebcc2200aad5985
153
py
Python
django-like/app_two/urls.py
mirokrastev/flask-structures
fb7ad464c3df85be2e66229b2fd99a8da903b3c9
[ "MIT" ]
1
2021-05-06T09:04:36.000Z
2021-05-06T09:04:36.000Z
django-like/app_two/urls.py
mirokrastev/flask-structures
fb7ad464c3df85be2e66229b2fd99a8da903b3c9
[ "MIT" ]
null
null
null
django-like/app_two/urls.py
mirokrastev/flask-structures
fb7ad464c3df85be2e66229b2fd99a8da903b3c9
[ "MIT" ]
null
null
null
from app_two import app_two_bp from app_two import views path = app_two_bp.add_url_rule urlpatterns = [ path('/', view_func=views.app_two_index) ]
17
44
0.764706
794d6d30203f7471c9f6203775e5519e5d3fb381
56,707
py
Python
Configuration/PyReleaseValidation/python/relval_standard.py
tklijnsma/cmssw
d7f103c459dd195eaa651ce0c30e4bd4bfd0fe43
[ "Apache-2.0" ]
1
2020-07-12T11:51:32.000Z
2020-07-12T11:51:32.000Z
Configuration/PyReleaseValidation/python/relval_standard.py
tklijnsma/cmssw
d7f103c459dd195eaa651ce0c30e4bd4bfd0fe43
[ "Apache-2.0" ]
3
2020-02-19T08:33:16.000Z
2020-03-13T10:02:19.000Z
Configuration/PyReleaseValidation/python/relval_standard.py
tklijnsma/cmssw
d7f103c459dd195eaa651ce0c30e4bd4bfd0fe43
[ "Apache-2.0" ]
1
2018-09-12T13:30:34.000Z
2018-09-12T13:30:34.000Z
# import the definition of the steps and input files: from Configuration.PyReleaseValidation.relval_steps import * # here only define the workflows as a combination of the steps defined above: workflows = Matrix() overridesEv5={'-n':'5'} # each workflow defines a name and a list of steps to be done. # if no explicit name/label given for the workflow (first arg), # the name of step1 will be used ## production tests workflows[1] = ['', ['ProdMinBias','DIGIPROD1','RECOPROD1']] workflows[2] = ['', ['ProdTTbar','DIGIPROD1','RECOPROD1']] workflows[3] = ['', ['ProdQCD_Pt_3000_3500','DIGIPROD1','RECOPROD1']] workflows.addOverride(3,overridesEv5) ####################################################### ###Prod2016 #workflows[1301] = ['', ['ProdMinBias_13','DIGIUP15PROD1','RECOPRODUP15','MINIAODMCUP15']] # Re-add miniaod here, to keep event content scrutinized workflows[1301] = ['', ['ProdMinBias_13','DIGIUP15PROD1','RECOPRODUP15']] ######################################################## workflows[1302] = ['', ['ProdTTbar_13','DIGIUP15PROD1','RECOPRODUP15']] workflows[1303] = ['', ['ProdQCD_Pt_3000_3500_13','DIGIUP15PROD1','RECOPRODUP15']] workflows.addOverride(1303,overridesEv5) #####Prod2017 workflows[1301.17] = ['', ['ProdMinBias_13UP17','DIGIUP17PROD1','RECOPRODUP17']] workflows[1302.17] = ['', ['ProdTTbar_13UP17','DIGIUP17PROD1','RECOPRODUP17']] workflows[1303.17] = ['', ['ProdQCD_Pt_3000_3500_13UP17','DIGIUP17PROD1','RECOPRODUP17']] workflows.addOverride(1303.17,overridesEv5) #####Prod2018 #workflows[1301.18] = ['', ['ProdMinBias_13UP18','DIGIUP18PROD1','RECOPRODUP18']] workflows[1302.18] = ['', ['ProdTTbar_13UP18','DIGIUP18PROD1','RECOPRODUP18','MINIAODMCUP18','NANOPRODUP18']] #workflows[1303.18] = ['', ['ProdQCD_Pt_3000_3500_13UP18','DIGIUP18PROD1','RECOPRODUP18']] workflows[1304.18] = ['', ['ProdZEE_13UP18','DIGIUP18PROD1','RECOPRODUP18','MINIAODMCUP18','NANOPRODUP18']] workflows[1304.182] = ['', ['ProdZEE_13UP18','DIGIUP18PROD1','RECOPRODUP18bParking','MINIAODMCUP18bParking']] #workflows.addOverride(1303.17,overridesEv5) #####Prod2018 with concurrentlumi workflows[1302.181] = ['', ['ProdTTbar_13UP18ml','DIGIUP18PROD1ml','RECOPRODUP18ml','MINIAODMCUP18ml','NANOPRODUP18ml']] workflows[1304.181] = ['', ['ProdZEE_13UP18ml','DIGIUP18PROD1ml','RECOPRODUP18ml','MINIAODMCUP18ml','NANOPRODUP18ml']] ### data ### workflows[4.5] = ['', ['RunCosmicsA','RECOCOSD','ALCACOSD','HARVESTDC']] #workflows[4.6] = ['', ['MinimumBias2010A','RECODR1','HARVESTDR1']] workflows[4.6] = ['', ['MinimumBias2010A','RECOSKIMALCA','HARVESTDR1']] #workflows[4.7] = ['', ['MinimumBias2010B','RECODR1ALCA','HARVESTDR1']] #workflows[4.8] = ['', ['WZMuSkim2010A','RECODR1','HARVESTDR1']] #workflows[4.9] = ['', ['WZEGSkim2010A','RECODR1','HARVESTDR1']] #workflows[4.10] = ['', ['WZMuSkim2010B','RECODR1','HARVESTDR1']] #workflows[4.11] = ['', ['WZEGSkim2010B','RECODR1','HARVESTDR1']] workflows[4.12] = ['', ['RunMinBias2010B','RECODR1','HARVESTDR1']] #workflows[4.13] = ['', ['RunMu2010B','RECODR1','HARVESTDR1']] #workflows[4.14] = ['', ['RunElectron2010B','RECODR1','HARVESTDR1']] #workflows[4.15] = ['', ['RunPhoton2010B','RECODR1','HARVESTDR1']] #workflows[4.16] = ['', ['RunJet2010B','RECODR1','HARVESTDR1']] workflows[4.17] = ['', ['RunMinBias2011A','HLTD','RECODR1reHLT','HARVESTDR1reHLT','SKIMDreHLT']] workflows[4.18] = ['', ['RunMu2011A','RECODR1','HARVESTDR1']] workflows[4.19] = ['', ['RunElectron2011A','RECODR1','HARVESTDR1']] workflows[4.20] = ['', ['RunPhoton2011A','RECODR1','HARVESTDR1']] workflows[4.21] = ['', ['RunJet2011A','RECODR1','HARVESTDR1']] workflows[4.22] = ['', ['RunCosmics2011A','RECOCOSD','ALCACOSD','SKIMCOSD','HARVESTDC']] workflows[4.23] = ['',['ValSkim2011A','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.24] = ['',['WMuSkim2011A','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.25] = ['',['WElSkim2011A','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.26] = ['',['ZMuSkim2011A','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.27] = ['',['ZElSkim2011A','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.28] = ['',['HighMet2011A','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.29] = ['', ['RunMinBias2011B','HLTD','RECODR1reHLT','HARVESTDR1reHLT','SKIMDreHLT']] #workflows[4.291] = ['', ['RunMinBias2011B','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.30] = ['', ['RunMu2011B','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.31] = ['', ['RunElectron2011B','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.32] = ['', ['RunPhoton2011B','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.33] = ['', ['RunJet2011B','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.34] = ['',['ValSkim2011B','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] #workflows[4.35] = ['',['WMuSkim2011B','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.36] = ['',['WElSkim2011B','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.37] = ['',['ZMuSkim2011B','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.38] = ['',['ZElSkim2011B','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.39] = ['',['HighMet2011B','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.40] = ['',['RunMinBias2012A','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.41] = ['',['RunTau2012A','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.42] = ['',['RunMET2012A','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.43] = ['',['RunMu2012A','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.44] = ['',['RunElectron2012A','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.45] = ['',['RunJet2012A','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.51] = ['',['RunMinBias2012B','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.52] = ['',['RunMu2012B','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.53] = ['',['RunPhoton2012B','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.54] = ['',['RunEl2012B','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.55] = ['',['RunJet2012B','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.56] = ['',['ZMuSkim2012B','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.57] = ['',['ZElSkim2012B','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.58] = ['',['WElSkim2012B','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.61] = ['',['RunMinBias2012C','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.62] = ['',['RunMu2012C','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.63] = ['',['RunPhoton2012C','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.64] = ['',['RunEl2012C','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.65] = ['',['RunJet2012C','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.66] = ['',['ZMuSkim2012C','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.67] = ['',['ZElSkim2012C','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.68] = ['',['WElSkim2012C','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.71] = ['',['RunMinBias2012D','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.72] = ['',['RunMu2012D','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.73] = ['',['RunPhoton2012D','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.74] = ['',['RunEl2012D','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.75] = ['',['RunJet2012D','HLTD','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.76] = ['',['ZMuSkim2012D','HLTDSKIM2','RECODR1reHLT2','HARVESTDR1reHLT']] workflows[4.77] = ['',['ZElSkim2012D','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[4.78] = ['',['WElSkim2012D','HLTDSKIM','RECODR1reHLT','HARVESTDR1reHLT']] workflows[140.51] = ['',['RunHI2010','REPACKHID','RECOHID11St3','HARVESTDHI']] workflows[140.52] = ['',['RunHI2010','RECOHID10','RECOHIR10D11','HARVESTDHI']] workflows[140.53] = ['',['RunHI2011','RECOHID11','HARVESTDHI']] workflows[140.54] = ['',['RunPA2013','RECO_PPbData','HARVEST_PPbData']] workflows[140.55] = ['',['RunHI2015VR','HYBRIDRepackHI2015VR','HYBRIDZSHI2015','RECOHID15','HARVESTDHI15']] workflows[140.56] = ['',['RunHI2018','RECOHID18','HARVESTDHI18']] workflows[140.5611] = ['',['RunHI2018AOD','REMINIAODHID18','HARVESTHI18MINIAOD']] workflows[140.57] = ['',['RunHI2018Reduced','RECOHID18','HARVESTDHI18']] ### run2 2015B 50ns ### workflows[134.701] = ['',['RunHLTPhy2015B','HLTDR2_50ns','RECODR2_50nsreHLT_HIPM','HARVESTDR2']] workflows[134.702] = ['',['RunDoubleEG2015B','HLTDR2_50ns','RECODR2_50nsreHLT_HIPM','HARVESTDR2']] workflows[134.703] = ['',['RunDoubleMuon2015B','HLTDR2_50ns','RECODR2_50nsreHLT_HIPM','HARVESTDR2']] workflows[134.704] = ['',['RunJetHT2015B','HLTDR2_50ns','RECODR2_50nsreHLT_HIPM','HARVESTDR2']] workflows[134.705] = ['',['RunMET2015B','HLTDR2_50ns','RECODR2_50nsreHLT_HIPM','HARVESTDR2']] workflows[134.706] = ['',['RunMuonEG2015B','HLTDR2_50ns','RECODR2_50nsreHLT_HIPM','HARVESTDR2']] workflows[134.707] = ['',['RunSingleEl2015B','HLTDR2_50ns','RECODR2_50nsreHLT_HIPM','HARVESTDR2']] workflows[134.708] = ['',['RunSingleMu2015B','HLTDR2_50ns','RECODR2_50nsreHLT_HIPM','HARVESTDR2']] workflows[134.709] = ['',['RunSinglePh2015B','HLTDR2_50ns','RECODR2_50nsreHLT_HIPM','HARVESTDR2']] workflows[134.710] = ['',['RunZeroBias2015B','HLTDR2_50ns','RECODR2_50nsreHLT_ZB_HIPM','HARVESTDR2ZB']] ### run 2015C 25ns ### workflows[134.801] = ['',['RunHLTPhy2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.802] = ['',['RunDoubleEG2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.803] = ['',['RunDoubleMuon2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.804] = ['',['RunJetHT2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.805] = ['',['RunMET2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.806] = ['',['RunMuonEG2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.807] = ['',['RunDoubleEGPrpt2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.808] = ['',['RunSingleMuPrpt2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.809] = ['',['RunSingleEl2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.810] = ['',['RunSingleMu2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.811] = ['',['RunSinglePh2015C','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.812] = ['',['RunZeroBias2015C','HLTDR2_25ns','RECODR2_25nsreHLT_ZB_HIPM','HARVESTDR2ZB']] workflows[134.813] = ['',['RunCosmics2015C','RECOCOSDRUN2','ALCACOSDRUN2','HARVESTDCRUN2']] ### run 2015D 25ns ### workflows[134.901] = ['',['RunHLTPhy2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.902] = ['',['RunDoubleEG2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.903] = ['',['RunDoubleMuon2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.904] = ['',['RunJetHT2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.905] = ['',['RunMET2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.906] = ['',['RunMuonEG2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.907] = ['',['RunDoubleEGPrpt2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.908] = ['',['RunSingleMuPrpt2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.909] = ['',['RunSingleEl2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.910] = ['',['RunSingleMu2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.911] = ['',['RunSinglePh2015D','HLTDR2_25ns','RECODR2_25nsreHLT_HIPM','HARVESTDR2']] workflows[134.912] = ['',['RunZeroBias2015D','HLTDR2_25ns','RECODR2_25nsreHLT_ZB_HIPM','HARVESTDR2ZB']] ### run 2016B ### workflows[136.721] = ['',['RunHLTPhy2016B','HLTDR2_2016','RECODR2_2016reHLT_HIPM','HARVESTDR2']] workflows[136.722] = ['',['RunDoubleEG2016B','HLTDR2_2016','RECODR2_2016reHLT_skimDoubleEG_HIPM','HARVESTDR2']] workflows[136.723] = ['',['RunDoubleMuon2016B','HLTDR2_2016','RECODR2_2016reHLT_HIPM','HARVESTDR2']] workflows[136.724] = ['',['RunJetHT2016B','HLTDR2_2016','RECODR2_2016reHLT_skimJetHT_HIPM','HARVESTDR2']] workflows[136.725] = ['',['RunMET2016B','HLTDR2_2016','RECODR2_2016reHLT_skimMET_HIPM','HARVESTDR2']] workflows[136.726] = ['',['RunMuonEG2016B','HLTDR2_2016','RECODR2_2016reHLT_skimMuonEG_HIPM','HARVESTDR2']] workflows[136.727] = ['',['RunDoubleEGPrpt2016B','HLTDR2_2016','RECODR2_2016reHLT_HIPM','HARVESTDR2']] workflows[136.728] = ['',['RunSingleMuPrpt2016B','HLTDR2_2016','RECODR2_2016reHLT_HIPM','HARVESTDR2']] workflows[136.729] = ['',['RunSingleEl2016B','HLTDR2_2016','RECODR2_2016reHLT_L1TEgDQM_HIPM','HARVEST2016_L1TEgDQM']] workflows[136.730] = ['',['RunSingleMu2016B','HLTDR2_2016','RECODR2_2016reHLT_skimSingleMu_HIPM','HARVEST2016_L1TMuDQM']] workflows[136.731] = ['',['RunSinglePh2016B','HLTDR2_2016','RECODR2_2016reHLT_skimSinglePh_HIPM','HARVESTDR2']] workflows[136.732] = ['',['RunZeroBias2016B','HLTDR2_2016','RECODR2_2016reHLT_ZB_HIPM','HARVESTDR2ZB']] workflows[136.733] = ['',['RunCosmics2016B','RECOCOSDRUN2','ALCACOSDRUN2','HARVESTDCRUN2']] workflows[136.734] = ['',['RunMuOnia2016B','HLTDR2_2016','RECODR2_2016reHLT_skimMuOnia_HIPM','HARVESTDR2']] workflows[136.735] = ['',['RunNoBPTX2016B','HLTDR2_2016','RECODR2reHLTAlCaTkCosmics_HIPM','HARVESTDR2']] # reminiAOD wf on 2016B input, mainly here for PPS testing workflows[136.72411] = ['',['RunJetHT2016B_reminiaodUL','REMINIAOD_data2016UL_HIPM','HARVESTDR2_REMINIAOD_data2016UL_HIPM']] ### run 2016C ### workflows[136.736] = ['',['RunHLTPhy2016C','HLTDR2_2016','RECODR2_2016reHLT_HIPM','HARVESTDR2']] workflows[136.737] = ['',['RunDoubleEG2016C','HLTDR2_2016','RECODR2_2016reHLT_skimDoubleEG_HIPM','HARVESTDR2']] workflows[136.738] = ['',['RunDoubleMuon2016C','HLTDR2_2016','RECODR2_2016reHLT_HIPM','HARVESTDR2']] workflows[136.739] = ['',['RunJetHT2016C','HLTDR2_2016','RECODR2_2016reHLT_skimJetHT_HIPM','HARVESTDR2']] workflows[136.740] = ['',['RunMET2016C','HLTDR2_2016','RECODR2_2016reHLT_skimMET_HIPM','HARVESTDR2']] workflows[136.741] = ['',['RunMuonEG2016C','HLTDR2_2016','RECODR2_2016reHLT_skimMuonEG_HIPM','HARVESTDR2']] workflows[136.742] = ['',['RunSingleEl2016C','HLTDR2_2016','RECODR2_2016reHLT_L1TEgDQM_HIPM','HARVEST2016_L1TEgDQM']] workflows[136.743] = ['',['RunSingleMu2016C','HLTDR2_2016','RECODR2_2016reHLT_L1TMuDQM_HIPM','HARVEST2016_L1TMuDQM']] workflows[136.744] = ['',['RunSinglePh2016C','HLTDR2_2016','RECODR2_2016reHLT_skimSinglePh_HIPM','HARVESTDR2']] workflows[136.745] = ['',['RunZeroBias2016C','HLTDR2_2016','RECODR2_2016reHLT_ZB_HIPM','HARVESTDR2ZB']] workflows[136.746] = ['',['RunMuOnia2016C','HLTDR2_2016','RECODR2_2016reHLT_skimMuOnia_HIPM','HARVESTDR2']] ### run 2016D ### workflows[136.747] = ['',['RunHLTPhy2016D','HLTDR2_2016','RECODR2_2016reHLT_HIPM','HARVESTDR2']] workflows[136.748] = ['',['RunDoubleEG2016D','HLTDR2_2016','RECODR2_2016reHLT_skimDoubleEG_HIPM','HARVESTDR2']] workflows[136.749] = ['',['RunDoubleMuon2016D','HLTDR2_2016','RECODR2_2016reHLT_HIPM','HARVESTDR2']] workflows[136.750] = ['',['RunJetHT2016D','HLTDR2_2016','RECODR2_2016reHLT_skimJetHT_HIPM','HARVESTDR2']] workflows[136.751] = ['',['RunMET2016D','HLTDR2_2016','RECODR2_2016reHLT_skimMET_HIPM','HARVESTDR2']] workflows[136.752] = ['',['RunMuonEG2016D','HLTDR2_2016','RECODR2_2016reHLT_skimMuonEG_HIPM','HARVESTDR2']] workflows[136.753] = ['',['RunSingleEl2016D','HLTDR2_2016','RECODR2_2016reHLT_L1TEgDQM_HIPM','HARVEST2016_L1TEgDQM']] workflows[136.754] = ['',['RunSingleMu2016D','HLTDR2_2016','RECODR2_2016reHLT_L1TMuDQM_HIPM','HARVEST2016_L1TMuDQM']] workflows[136.755] = ['',['RunSinglePh2016D','HLTDR2_2016','RECODR2_2016reHLT_skimSinglePh_HIPM','HARVESTDR2']] workflows[136.756] = ['',['RunZeroBias2016D','HLTDR2_2016','RECODR2_2016reHLT_ZB_HIPM','HARVESTDR2ZB']] workflows[136.757] = ['',['RunMuOnia2016D','HLTDR2_2016','RECODR2_2016reHLT_skimMuOnia_HIPM','HARVESTDR2']] ### run 2016E ### workflows[136.758] = ['',['RunHLTPhy2016E','HLTDR2_2016','RECODR2_2016reHLT_HIPM','HARVESTDR2']] workflows[136.759] = ['',['RunDoubleEG2016E','HLTDR2_2016','RECODR2_2016reHLT_skimDoubleEG_HIPM','HARVESTDR2']] workflows[136.760] = ['',['RunDoubleMuon2016E','HLTDR2_2016','RECODR2_2016reHLT_HIPM','HARVESTDR2']] workflows[136.761] = ['',['RunJetHT2016E','HLTDR2_2016','RECODR2_2016reHLT_skimJetHT_HIPM','HARVESTDR2']] workflows[136.762] = ['',['RunMET2016E','HLTDR2_2016','RECODR2_2016reHLT_skimMET_HIPM','HARVESTDR2']] workflows[136.763] = ['',['RunMuonEG2016E','HLTDR2_2016','RECODR2_2016reHLT_skimMuonEG_HIPM','HARVESTDR2']] workflows[136.764] = ['',['RunSingleEl2016E','HLTDR2_2016','RECODR2_2016reHLT_L1TEgDQM_HIPM','HARVEST2016_L1TEgDQM']] workflows[136.765] = ['',['RunSingleMu2016E','HLTDR2_2016','RECODR2_2016reHLT_L1TMuDQM_HIPM','HARVEST2016_L1TMuDQM']] workflows[136.766] = ['',['RunSinglePh2016E','HLTDR2_2016','RECODR2_2016reHLT_skimSinglePh_HIPM','HARVESTDR2']] workflows[136.767] = ['',['RunZeroBias2016E','HLTDR2_2016','RECODR2_2016reHLT_ZB_HIPM','HARVESTDR2ZB']] workflows[136.768] = ['',['RunMuOnia2016E','HLTDR2_2016','RECODR2_2016reHLT_skimMuOnia_HIPM','HARVESTDR2']] # reminiAOD wf on 2016E input workflows[136.7611] = ['',['RunJetHT2016E_reminiaod','REMINIAOD_data2016_HIPM','HARVESTDR2_REMINIAOD_data2016_HIPM']] workflows[136.76111] = ['',['RunJetHT2016E_reminiaodUL','REMINIAOD_data2016UL_HIPM','HARVESTDR2_REMINIAOD_data2016UL_HIPM']] ### run 2016H ### workflows[136.769] = ['',['RunHLTPhy2016H','HLTDR2_2016','RECODR2_2016reHLT_Prompt','HARVESTDR2']] workflows[136.770] = ['',['RunDoubleEG2016H','HLTDR2_2016','RECODR2_2016reHLT_skimDoubleEG_Prompt','HARVESTDR2']] workflows[136.771] = ['',['RunDoubleMuon2016H','HLTDR2_2016','RECODR2_2016reHLT_Prompt','HARVESTDR2']] workflows[136.772] = ['',['RunJetHT2016H','HLTDR2_2016','RECODR2_2016reHLT_skimJetHT_Prompt','HARVESTDR2']] workflows[136.773] = ['',['RunMET2016H','HLTDR2_2016','RECODR2_2016reHLT_skimMET_Prompt','HARVESTDR2']] workflows[136.774] = ['',['RunMuonEG2016H','HLTDR2_2016','RECODR2_2016reHLT_skimMuonEG_Prompt','HARVESTDR2']] workflows[136.775] = ['',['RunSingleEl2016H','HLTDR2_2016','RECODR2_2016reHLT_Prompt_L1TEgDQM','HARVEST2016_L1TEgDQM']] workflows[136.776] = ['',['RunSingleMu2016H','HLTDR2_2016','RECODR2_2016reHLT_Prompt_Lumi_L1TMuDQM','HARVEST2016_L1TMuDQM']] workflows[136.777] = ['',['RunSinglePh2016H','HLTDR2_2016','RECODR2_2016reHLT_skimSinglePh_Prompt','HARVESTDR2']] workflows[136.778] = ['',['RunZeroBias2016H','HLTDR2_2016','RECODR2_2016reHLT_ZBPrompt','HARVESTDR2ZB']] workflows[136.779] = ['',['RunMuOnia2016H','HLTDR2_2016','RECODR2_2016reHLT_skimMuOnia_Prompt','HARVESTDR2']] # reminiAOD wf on 2016H input workflows[136.7721] = ['',['RunJetHT2016H_reminiaod','REMINIAOD_data2016','HARVESTDR2_REMINIAOD_data2016']] workflows[136.77211] = ['',['RunJetHT2016H_reminiaodUL','REMINIAOD_data2016UL','HARVESTDR2_REMINIAOD_data2016UL']] # nanoAOD wf on 2016H 80X input workflows[136.7722] = ['',['RunJetHT2016H_nano','NANOEDM2016_80X','HARVESTNANOAOD2016_80X']] ### run 2017B ### workflows[136.780] = ['',['RunHLTPhy2017B','HLTDR2_2017','RECODR2_2017reHLT_Prompt','HARVEST2017']] workflows[136.781] = ['',['RunDoubleEG2017B','HLTDR2_2017','RECODR2_2017reHLT_skimDoubleEG_Prompt','HARVEST2017']] workflows[136.782] = ['',['RunDoubleMuon2017B','HLTDR2_2017','RECODR2_2017reHLT_Prompt','HARVEST2017']] workflows[136.783] = ['',['RunJetHT2017B','HLTDR2_2017','RECODR2_2017reHLT_skimJetHT_Prompt','HARVEST2017']] workflows[136.784] = ['',['RunMET2017B','HLTDR2_2017','RECODR2_2017reHLT_skimMET_Prompt','HARVEST2017']] workflows[136.785] = ['',['RunMuonEG2017B','HLTDR2_2017','RECODR2_2017reHLT_skimMuonEG_Prompt','HARVEST2017']] workflows[136.786] = ['',['RunSingleEl2017B','HLTDR2_2017','RECODR2_2017reHLT_Prompt_L1TEgDQM','HARVEST2017_L1TEgDQM']] workflows[136.787] = ['',['RunSingleMu2017B','HLTDR2_2017','RECODR2_2017reHLT_skimSingleMu_Prompt_Lumi','HARVEST2017_L1TMuDQM']] workflows[136.788] = ['',['RunSinglePh2017B','HLTDR2_2017','RECODR2_2017reHLT_skimSinglePh_Prompt','HARVEST2017']] workflows[136.789] = ['',['RunZeroBias2017B','HLTDR2_2017','RECODR2_2017reHLT_ZBPrompt','HARVEST2017ZB']] workflows[136.790] = ['',['RunMuOnia2017B','HLTDR2_2017','RECODR2_2017reHLT_skimMuOnia_Prompt','HARVEST2017']] workflows[136.791] = ['',['RunNoBPTX2017B','HLTDR2_2017','RECODR2_2017reHLTAlCaTkCosmics_Prompt','HARVEST2017']] workflows[136.7801] = ['',['RunHLTPhy2017B_AOD','DQMHLTonAOD_2017','HARVESTDQMHLTonAOD_2017']] workflows[136.7802] = ['',['RunHLTPhy2017B_AODextra','DQMHLTonAODextra_2017','HARVESTDQMHLTonAOD_2017']] workflows[136.7803] = ['',['RunHLTPhy2017B_RAWAOD','DQMHLTonRAWAOD_2017','HARVESTDQMHLTonAOD_2017']] workflows[136.844] = ['',['RunCharmonium2017B','HLTDR2_2017','RECODR2_2017reHLT_skimCharmonium_Prompt','HARVEST2017']] ### run 2017C ### workflows[136.792] = ['',['RunHLTPhy2017C','HLTDR2_2017','RECODR2_2017reHLT_Prompt','HARVEST2017']] workflows[136.793] = ['',['RunDoubleEG2017C','HLTDR2_2017','RECODR2_2017reHLT_skimDoubleEG_Prompt','HARVEST2017']] workflows[136.794] = ['',['RunDoubleMuon2017C','HLTDR2_2017','RECODR2_2017reHLT_Prompt','HARVEST2017']] workflows[136.795] = ['',['RunJetHT2017C','HLTDR2_2017','RECODR2_2017reHLT_skimJetHT_Prompt','HARVEST2017']] workflows[136.796] = ['',['RunMET2017C','HLTDR2_2017','RECODR2_2017reHLT_skimMET_Prompt','HARVEST2017']] workflows[136.797] = ['',['RunMuonEG2017C','HLTDR2_2017','RECODR2_2017reHLT_skimMuonEG_Prompt','HARVEST2017']] workflows[136.798] = ['',['RunSingleEl2017C','HLTDR2_2017','RECODR2_2017reHLT_Prompt_L1TEgDQM','HARVEST2017_L1TEgDQM']] workflows[136.799] = ['',['RunSingleMu2017C','HLTDR2_2017','RECODR2_2017reHLT_skimSingleMu_Prompt_Lumi','HARVEST2017_L1TMuDQM']] workflows[136.800] = ['',['RunSinglePh2017C','HLTDR2_2017','RECODR2_2017reHLT_skimSinglePh_Prompt','HARVEST2017']] workflows[136.801] = ['',['RunZeroBias2017C','HLTDR2_2017','RECODR2_2017reHLT_ZBPrompt','HARVEST2017ZB']] workflows[136.802] = ['',['RunMuOnia2017C','HLTDR2_2017','RECODR2_2017reHLT_skimMuOnia_Prompt','HARVEST2017']] workflows[136.803] = ['',['RunNoBPTX2017C','HLTDR2_2017','RECODR2_2017reHLTAlCaTkCosmics_Prompt','HARVEST2017']] workflows[136.840] = ['',['RunDisplacedJet2017C','HLTDR2_2017','RECODR2_2017reHLT_skimDisplacedJet_Prompt','HARVEST2017']] workflows[136.845] = ['',['RunCharmonium2017C','HLTDR2_2017','RECODR2_2017reHLT_skimCharmonium_Prompt','HARVEST2017']] ### run 2017D ### workflows[136.804] = ['',['RunHLTPhy2017D','HLTDR2_2017','RECODR2_2017reHLT_Prompt','HARVEST2017']] workflows[136.805] = ['',['RunDoubleEG2017D','HLTDR2_2017','RECODR2_2017reHLT_skimDoubleEG_Prompt','HARVEST2017']] workflows[136.806] = ['',['RunDoubleMuon2017D','HLTDR2_2017','RECODR2_2017reHLT_Prompt','HARVEST2017']] workflows[136.807] = ['',['RunJetHT2017D','HLTDR2_2017','RECODR2_2017reHLT_skimJetHT_Prompt','HARVEST2017']] workflows[136.808] = ['',['RunMET2017D','HLTDR2_2017','RECODR2_2017reHLT_skimMET_Prompt','HARVEST2017']] workflows[136.809] = ['',['RunMuonEG2017D','HLTDR2_2017','RECODR2_2017reHLT_skimMuonEG_Prompt','HARVEST2017']] workflows[136.810] = ['',['RunSingleEl2017D','HLTDR2_2017','RECODR2_2017reHLT_Prompt_L1TEgDQM','HARVEST2017_L1TEgDQM']] workflows[136.811] = ['',['RunSingleMu2017D','HLTDR2_2017','RECODR2_2017reHLT_skimSingleMu_Prompt_Lumi','HARVEST2017_L1TMuDQM']] workflows[136.812] = ['',['RunSinglePh2017D','HLTDR2_2017','RECODR2_2017reHLT_skimSinglePh_Prompt','HARVEST2017']] workflows[136.813] = ['',['RunZeroBias2017D','HLTDR2_2017','RECODR2_2017reHLT_ZBPrompt','HARVEST2017ZB']] workflows[136.814] = ['',['RunMuOnia2017D','HLTDR2_2017','RECODR2_2017reHLT_skimMuOnia_Prompt','HARVEST2017']] workflows[136.815] = ['',['RunNoBPTX2017D','HLTDR2_2017','RECODR2_2017reHLTAlCaTkCosmics_Prompt','HARVEST2017']] workflows[136.841] = ['',['RunDisplacedJet2017D','HLTDR2_2017','RECODR2_2017reHLT_skimDisplacedJet_Prompt','HARVEST2017']] workflows[136.846] = ['',['RunCharmonium2017D','HLTDR2_2017','RECODR2_2017reHLT_skimCharmonium_Prompt','HARVEST2017']] ### run 2017E ### workflows[136.816] = ['',['RunHLTPhy2017E','HLTDR2_2017','RECODR2_2017reHLT_Prompt','HARVEST2017']] workflows[136.817] = ['',['RunDoubleEG2017E','HLTDR2_2017','RECODR2_2017reHLT_skimDoubleEG_Prompt','HARVEST2017']] workflows[136.818] = ['',['RunDoubleMuon2017E','HLTDR2_2017','RECODR2_2017reHLT_Prompt','HARVEST2017']] workflows[136.819] = ['',['RunJetHT2017E','HLTDR2_2017','RECODR2_2017reHLT_skimJetHT_Prompt','HARVEST2017']] workflows[136.820] = ['',['RunMET2017E','HLTDR2_2017','RECODR2_2017reHLT_skimMET_Prompt','HARVEST2017']] workflows[136.821] = ['',['RunMuonEG2017E','HLTDR2_2017','RECODR2_2017reHLT_skimMuonEG_Prompt','HARVEST2017']] workflows[136.822] = ['',['RunSingleEl2017E','HLTDR2_2017','RECODR2_2017reHLT_Prompt_L1TEgDQM','HARVEST2017_L1TEgDQM']] workflows[136.823] = ['',['RunSingleMu2017E','HLTDR2_2017','RECODR2_2017reHLT_skimSingleMu_Prompt_Lumi','HARVEST2017_L1TMuDQM']] workflows[136.824] = ['',['RunSinglePh2017E','HLTDR2_2017','RECODR2_2017reHLT_skimSinglePh_Prompt','HARVEST2017']] workflows[136.825] = ['',['RunZeroBias2017E','HLTDR2_2017','RECODR2_2017reHLT_ZBPrompt','HARVEST2017ZB']] workflows[136.826] = ['',['RunMuOnia2017E','HLTDR2_2017','RECODR2_2017reHLT_skimMuOnia_Prompt','HARVEST2017']] workflows[136.827] = ['',['RunNoBPTX2017E','HLTDR2_2017','RECODR2_2017reHLTAlCaTkCosmics_Prompt','HARVEST2017']] workflows[136.842] = ['',['RunDisplacedJet2017E','HLTDR2_2017','RECODR2_2017reHLT_skimDisplacedJet_Prompt','HARVEST2017']] workflows[136.847] = ['',['RunCharmonium2017E','HLTDR2_2017','RECODR2_2017reHLT_skimCharmonium_Prompt','HARVEST2017']] ### run 2017F ### workflows[136.828] = ['',['RunHLTPhy2017F','HLTDR2_2017','RECODR2_2017reHLT_Prompt','HARVEST2017']] workflows[136.829] = ['',['RunDoubleEG2017F','HLTDR2_2017','RECODR2_2017reHLT_skimDoubleEG_Prompt','HARVEST2017']] workflows[136.830] = ['',['RunDoubleMuon2017F','HLTDR2_2017','RECODR2_2017reHLT_Prompt','HARVEST2017']] workflows[136.831] = ['',['RunJetHT2017F','HLTDR2_2017','RECODR2_2017reHLT_skimJetHT_Prompt','HARVEST2017']] workflows[136.832] = ['',['RunMET2017F','HLTDR2_2017','RECODR2_2017reHLT_skimMET_Prompt','HARVEST2017']] workflows[136.833] = ['',['RunMuonEG2017F','HLTDR2_2017','RECODR2_2017reHLT_skimMuonEG_Prompt','HARVEST2017']] workflows[136.834] = ['',['RunSingleEl2017F','HLTDR2_2017','RECODR2_2017reHLT_Prompt_L1TEgDQM','HARVEST2017_L1TEgDQM']] workflows[136.835] = ['',['RunSingleMu2017F','HLTDR2_2017','RECODR2_2017reHLT_skimSingleMu_Prompt_Lumi','HARVEST2017_L1TMuDQM']] workflows[136.836] = ['',['RunSinglePh2017F','HLTDR2_2017','RECODR2_2017reHLT_skimSinglePh_Prompt','HARVEST2017']] workflows[136.837] = ['',['RunZeroBias2017F','HLTDR2_2017','RECODR2_2017reHLT_ZBPrompt','HARVEST2017ZB']] workflows[136.838] = ['',['RunMuOnia2017F','HLTDR2_2017','RECODR2_2017reHLT_skimMuOnia_Prompt','HARVEST2017']] workflows[136.839] = ['',['RunNoBPTX2017F','HLTDR2_2017','RECODR2_2017reHLTAlCaTkCosmics_Prompt','HARVEST2017']] workflows[136.8391] = ['',['RunExpressPhy2017F','HLTDR2_2017','RECODR2_2017reHLTSiPixelCalZeroBias_Prompt','HARVEST2017']] workflows[136.843] = ['',['RunDisplacedJet2017F','HLTDR2_2017','RECODR2_2017reHLT_skimDisplacedJet_Prompt','HARVEST2017']] workflows[136.848] = ['',['RunCharmonium2017F','HLTDR2_2017','RECODR2_2017reHLT_skimCharmonium_Prompt','HARVEST2017']] # reminiAOD wf on 2017F input workflows[136.8311] = ['',['RunJetHT2017F_reminiaod','REMINIAOD_data2017','HARVEST2017_REMINIAOD_data2017']] workflows[136.83111] = ['',['RunJetHT2017F_reminiaodUL','REMINIAOD_data2017UL','HARVEST2017_REMINIAOD_data2017UL']] # NANOAOD wf on 2017C miniAODv2 94X input workflows[136.7952] = ['',['RunJetHT2017C_94Xv2NanoAODINPUT','NANOEDM2017_94XMiniAODv2','HARVESTNANOAOD2017_94XMiniAODv2']] ### run 2018A ### workflows[136.849] = ['',['RunHLTPhy2018A','HLTDR2_2018','RECODR2_2018reHLT_Offline','HARVEST2018']] workflows[136.850] = ['',['RunEGamma2018A','HLTDR2_2018','RECODR2_2018reHLT_skimEGamma_Offline_L1TEgDQM','HARVEST2018_L1TEgDQM']] workflows[136.851] = ['',['RunDoubleMuon2018A','HLTDR2_2018','RECODR2_2018reHLT_Offline','HARVEST2018']] workflows[136.852] = ['',['RunJetHT2018A','HLTDR2_2018','RECODR2_2018reHLT_skimJetHT_Offline','HARVEST2018']] workflows[136.853] = ['',['RunMET2018A','HLTDR2_2018','RECODR2_2018reHLT_skimMET_Offline','HARVEST2018']] workflows[136.854] = ['',['RunMuonEG2018A','HLTDR2_2018','RECODR2_2018reHLT_skimMuonEG_Offline','HARVEST2018']] workflows[136.855] = ['',['RunSingleMu2018A','HLTDR2_2018','RECODR2_2018reHLT_skimSingleMu_Offline_Lumi','HARVEST2018_L1TMuDQM']] workflows[136.856] = ['',['RunZeroBias2018A','HLTDR2_2018','RECODR2_2018reHLT_ZBOffline','HARVEST2018ZB']] workflows[136.857] = ['',['RunMuOnia2018A','HLTDR2_2018','RECODR2_2018reHLT_skimMuOnia_Offline','HARVEST2018']] workflows[136.858] = ['',['RunNoBPTX2018A','HLTDR2_2018','RECODR2_2018reHLTAlCaTkCosmics_Offline','HARVEST2018']] workflows[136.859] = ['',['RunDisplacedJet2018A','HLTDR2_2018','RECODR2_2018reHLT_skimDisplacedJet_Offline','HARVEST2018']] workflows[136.860] = ['',['RunCharmonium2018A','HLTDR2_2018','RECODR2_2018reHLT_skimCharmonium_Offline','HARVEST2018']] ### wf to test 90 m beta* Totem run reconstruction ### workflows[136.8501] = ['',['RunEGamma2018A','HLTDR2_2018','RECODR2_2018reHLT_skimParkingBPH_Offline','HARVEST2018']] workflows[136.8561] = ['',['RunZeroBias_hBStarTk','HLTDR2_2018_hBStar','RECODR2_2018reHLT_Offline_hBStar','HARVEST2018_hBStar']] workflows[136.8562] = ['',['RunZeroBias1_hBStarRP','HLTDR2_2018_hBStar','RECODR2_2018reHLT_Offline_hBStar','HARVEST2018_hBStar']] ### NANOAOD wf on 2018 prompt reco MINIADD workflows[136.8521] = ['',['RunJetHT2018A_nano','NANOEDM2018_102Xv1','HARVESTNANOAOD2018_102Xv1']] ### run 2018B ### workflows[136.861] = ['',['RunHLTPhy2018B','HLTDR2_2018','RECODR2_2018reHLT_Offline','HARVEST2018']] workflows[136.862] = ['',['RunEGamma2018B','HLTDR2_2018','RECODR2_2018reHLT_skimEGamma_Offline_L1TEgDQM','HARVEST2018_L1TEgDQM']] workflows[136.863] = ['',['RunDoubleMuon2018B','HLTDR2_2018','RECODR2_2018reHLT_Offline','HARVEST2018']] workflows[136.864] = ['',['RunJetHT2018B','HLTDR2_2018','RECODR2_2018reHLT_skimJetHT_Offline','HARVEST2018']] workflows[136.865] = ['',['RunMET2018B','HLTDR2_2018','RECODR2_2018reHLT_skimMET_Offline','HARVEST2018']] workflows[136.866] = ['',['RunMuonEG2018B','HLTDR2_2018','RECODR2_2018reHLT_skimMuonEG_Offline','HARVEST2018']] workflows[136.867] = ['',['RunSingleMu2018B','HLTDR2_2018','RECODR2_2018reHLT_skimSingleMu_Offline_Lumi','HARVEST2018_L1TMuDQM']] workflows[136.868] = ['',['RunZeroBias2018B','HLTDR2_2018','RECODR2_2018reHLT_ZBOffline','HARVEST2018ZB']] workflows[136.869] = ['',['RunMuOnia2018B','HLTDR2_2018','RECODR2_2018reHLT_skimMuOnia_Offline','HARVEST2018']] workflows[136.870] = ['',['RunNoBPTX2018B','HLTDR2_2018','RECODR2_2018reHLTAlCaTkCosmics_Offline','HARVEST2018']] workflows[136.871] = ['',['RunDisplacedJet2018B','HLTDR2_2018','RECODR2_2018reHLT_skimDisplacedJet_Offline','HARVEST2018']] workflows[136.872] = ['',['RunCharmonium2018B','HLTDR2_2018','RECODR2_2018reHLT_skimCharmonium_Offline','HARVEST2018']] workflows[136.898] = ['',['RunParkingBPH2018B','HLTDR2_2018','RECODR2_2018reHLT_skimParkingBPH_Offline','HARVEST2018']] workflows[136.8642] = ['',['RunJetHT2018BHEfail','HLTDR2_2018','RECODR2_2018reHLT_skimJetHT_Prompt_HEfail','HARVEST2018_HEfail']] workflows[136.8643] = ['',['RunJetHT2018BBadHcalMitig','HLTDR2_2018_BadHcalMitig','RECODR2_2018reHLT_skimJetHT_Prompt_BadHcalMitig','HARVEST2018_BadHcalMitig']] ### run 2018C ### workflows[136.873] = ['',['RunHLTPhy2018C','HLTDR2_2018','RECODR2_2018reHLT_Offline','HARVEST2018']] workflows[136.874] = ['',['RunEGamma2018C','HLTDR2_2018','RECODR2_2018reHLT_skimEGamma_Offline_L1TEgDQM','HARVEST2018_L1TEgDQM']] workflows[136.875] = ['',['RunDoubleMuon2018C','HLTDR2_2018','RECODR2_2018reHLT_Offline','HARVEST2018']] workflows[136.876] = ['',['RunJetHT2018C','HLTDR2_2018','RECODR2_2018reHLT_skimJetHT_Offline','HARVEST2018']] workflows[136.877] = ['',['RunMET2018C','HLTDR2_2018','RECODR2_2018reHLT_skimMET_Offline','HARVEST2018']] workflows[136.878] = ['',['RunMuonEG2018C','HLTDR2_2018','RECODR2_2018reHLT_skimMuonEG_Offline','HARVEST2018']] workflows[136.879] = ['',['RunSingleMu2018C','HLTDR2_2018','RECODR2_2018reHLT_skimSingleMu_Offline_Lumi','HARVEST2018_L1TMuDQM']] workflows[136.880] = ['',['RunZeroBias2018C','HLTDR2_2018','RECODR2_2018reHLT_ZBOffline','HARVEST2018ZB']] workflows[136.881] = ['',['RunMuOnia2018C','HLTDR2_2018','RECODR2_2018reHLT_skimMuOnia_Offline','HARVEST2018']] workflows[136.882] = ['',['RunNoBPTX2018C','HLTDR2_2018','RECODR2_2018reHLTAlCaTkCosmics_Offline','HARVEST2018']] workflows[136.883] = ['',['RunDisplacedJet2018C','HLTDR2_2018','RECODR2_2018reHLT_skimDisplacedJet_Offline','HARVEST2018']] workflows[136.884] = ['',['RunCharmonium2018C','HLTDR2_2018','RECODR2_2018reHLT_skimCharmonium_Offline','HARVEST2018']] ### run 2018D ### workflows[136.885] = ['',['RunHLTPhy2018D','HLTDR2_2018','RECODR2_2018reHLT_Prompt','HARVEST2018_Prompt']] workflows[136.886] = ['',['RunEGamma2018D','HLTDR2_2018','RECODR2_2018reHLT_skimEGamma_Prompt_L1TEgDQM','HARVEST2018_L1TEgDQM_Prompt']] workflows[136.887] = ['',['RunDoubleMuon2018D','HLTDR2_2018','RECODR2_2018reHLT_Prompt','HARVEST2018_Prompt']] workflows[136.888] = ['',['RunJetHT2018D','HLTDR2_2018','RECODR2_2018reHLT_skimJetHT_Prompt','HARVEST2018_Prompt']] workflows[136.889] = ['',['RunMET2018D','HLTDR2_2018','RECODR2_2018reHLT_skimMET_Prompt','HARVEST2018_Prompt']] workflows[136.890] = ['',['RunMuonEG2018D','HLTDR2_2018','RECODR2_2018reHLT_skimMuonEG_Prompt','HARVEST2018_Prompt']] workflows[136.891] = ['',['RunSingleMu2018D','HLTDR2_2018','RECODR2_2018reHLT_skimSingleMu_Prompt_Lumi','HARVEST2018_L1TMuDQM_Prompt']] workflows[136.892] = ['',['RunZeroBias2018D','HLTDR2_2018','RECODR2_2018reHLT_ZBPrompt','HARVEST2018_PromptZB']] workflows[136.893] = ['',['RunMuOnia2018D','HLTDR2_2018','RECODR2_2018reHLT_skimMuOnia_Prompt','HARVEST2018_Prompt']] workflows[136.894] = ['',['RunNoBPTX2018D','HLTDR2_2018','RECODR2_2018reHLTAlCaTkCosmics_Prompt','HARVEST2018_Prompt']] workflows[136.895] = ['',['RunDisplacedJet2018D','HLTDR2_2018','RECODR2_2018reHLT_skimDisplacedJet_Prompt','HARVEST2018_Prompt']] workflows[136.896] = ['',['RunCharmonium2018D','HLTDR2_2018','RECODR2_2018reHLT_skimCharmonium_Prompt','HARVEST2018_Prompt']] # reminiAOD wf on 2018D input workflows[136.88811] = ['',['RunJetHT2018D_reminiaodUL','REMINIAOD_data2018UL','HARVEST2018_REMINIAOD_data2018UL']] ### run 2018D pixel tracks ### workflows[136.8855] = ['',['RunHLTPhy2018D','HLTDR2_2018','RECODR2_2018reHLT_Prompt_pixelTrackingOnly','HARVEST2018_pixelTrackingOnly']] workflows[136.885501] = ['',['RunHLTPhy2018D','HLTDR2_2018','RECODR2_2018reHLT_Patatrack_PixelOnlyCPU','HARVEST2018_pixelTrackingOnly']] workflows[136.885502] = ['',['RunHLTPhy2018D','HLTDR2_2018','RECODR2_2018reHLT_Patatrack_PixelOnlyGPU','HARVEST2018_pixelTrackingOnly']] workflows[136.8885] = ['',['RunJetHT2018D','HLTDR2_2018','RECODR2_2018reHLT_Prompt_pixelTrackingOnly','HARVEST2018_pixelTrackingOnly']] workflows[136.888501] = ['',['RunJetHT2018D','HLTDR2_2018','RECODR2_2018reHLT_Patatrack_PixelOnlyCPU','HARVEST2018_pixelTrackingOnly']] workflows[136.888502] = ['',['RunJetHT2018D','HLTDR2_2018','RECODR2_2018reHLT_Patatrack_PixelOnlyGPU','HARVEST2018_pixelTrackingOnly']] ### run 2018D ECAL-only ### workflows[136.885511] = ['',['RunHLTPhy2018D','HLTDR2_2018','RECODR2_2018reHLT_ECALOnlyCPU','HARVEST2018_ECALOnly']] workflows[136.885512] = ['',['RunHLTPhy2018D','HLTDR2_2018','RECODR2_2018reHLT_ECALOnlyGPU','HARVEST2018_ECALOnly']] workflows[136.888511] = ['',['RunJetHT2018D','HLTDR2_2018','RECODR2_2018reHLT_ECALOnlyCPU','HARVEST2018_ECALOnly']] workflows[136.888512] = ['',['RunJetHT2018D','HLTDR2_2018','RECODR2_2018reHLT_ECALOnlyGPU','HARVEST2018_ECALOnly']] ### run 2018D HCAL-only ### workflows[136.885521] = ['',['RunHLTPhy2018D','HLTDR2_2018','RECODR2_2018reHLT_HCALOnlyCPU','HARVEST2018_HCALOnly']] workflows[136.885522] = ['',['RunHLTPhy2018D','HLTDR2_2018','RECODR2_2018reHLT_HCALOnlyGPU','HARVEST2018_HCALOnly']] workflows[136.888521] = ['',['RunJetHT2018D','HLTDR2_2018','RECODR2_2018reHLT_HCALOnlyCPU','HARVEST2018_HCALOnly']] workflows[136.888522] = ['',['RunJetHT2018D','HLTDR2_2018','RECODR2_2018reHLT_HCALOnlyGPU','HARVEST2018_HCALOnly']] # multi-run harvesting workflows[137.8] = ['',['RunEGamma2018C','HLTDR2_2018','RECODR2_2018reHLT_skimEGamma_Offline_L1TEgDQM', 'RunEGamma2018D','HLTDR2_2018','RECODR2_2018reHLT_skimEGamma_Prompt_L1TEgDQM','HARVEST2018_L1TEgDQM_MULTIRUN']] ### LS2 - MWGR ### workflows[138.1] = ['',['RunCosmics2020','RECOCOSDRUN3','ALCACOSDRUN3','HARVESTDCRUN3']] #### Test of lumi section boundary crossing with run2 2018D #### workflows[136.8861] = ['',['RunEGamma2018Dml1','HLTDR2_2018ml','RECODR2_2018reHLT_skimEGamma_Prompt_L1TEgDQMml','HARVEST2018_L1TEgDQM_Prompt']] workflows[136.8862] = ['',['RunEGamma2018Dml2','HLTDR2_2018ml','RECODR2_2018reHLT_skimEGamma_Prompt_L1TEgDQMml','HARVEST2018_L1TEgDQM_Prompt']] ### fastsim ### workflows[5.1] = ['TTbar', ['TTbarFS','HARVESTFS']] workflows[5.2] = ['SingleMuPt10', ['SingleMuPt10FS','HARVESTFS']] workflows[5.3] = ['SingleMuPt100', ['SingleMuPt100FS','HARVESTFS']] workflows[5.4] = ['ZEE', ['ZEEFS','HARVESTFS']] workflows[5.5] = ['ZTT',['ZTTFS','HARVESTFS']] workflows[5.6] = ['QCD_FlatPt_15_3000', ['QCDFlatPt153000FS','HARVESTFS']] workflows[5.7] = ['H130GGgluonfusion', ['H130GGgluonfusionFS','HARVESTFS']] ### fastsim_13 TeV ### workflows[135.1] = ['TTbar_13', ['TTbarFS_13','HARVESTUP15FS','MINIAODMCUP15FS']] workflows[135.2] = ['SingleMuPt10_UP15', ['SingleMuPt10FS_UP15','HARVESTUP15FS','MINIAODMCUP15FS']] workflows[135.3] = ['SingleMuPt100_UP15', ['SingleMuPt100FS_UP15','HARVESTUP15FS','MINIAODMCUP15FS']] workflows[135.4] = ['ZEE_13', ['ZEEFS_13','HARVESTUP15FS','MINIAODMCUP15FS']] workflows[135.5] = ['ZTT_13',['ZTTFS_13','HARVESTUP15FS','MINIAODMCUP15FS']] workflows[135.6] = ['QCD_FlatPt_15_3000_13', ['QCDFlatPt153000FS_13','HARVESTUP15FS','MINIAODMCUP15FS']] workflows[135.7] = ['H125GGgluonfusion_13', ['H125GGgluonfusionFS_13','HARVESTUP15FS','MINIAODMCUP15FS']] workflows[135.9] = ['ZMM_13',['ZMMFS_13','HARVESTUP15FS','MINIAODMCUP15FS']] workflows[135.11] = ['SMS-T1tttt_mGl-1500_mLSP-100_13', ['SMS-T1tttt_mGl-1500_mLSP-100FS_13','HARVESTUP15FS','MINIAODMCUP15FS']] workflows[135.12] = ['QCD_Pt_80_120_13', ['QCD_Pt_80_120FS_13','HARVESTUP15FS','MINIAODMCUP15FS']] workflows[135.13] = ['TTbar_13', ['TTbarFS_13_trackingOnlyValidation','HARVESTUP15FS_trackingOnly']] ### MinBias fastsim_13 TeV for mixing ### workflows[135.8] = ['',['MinBiasFS_13_ForMixing']] ### Phase1 FastSim 13TeV ### workflows[2017.1] = ['TTbar_13_UP17', ['TTbarFS_13_UP17','HARVESTUP17FS','MINIAODMCUP17FS']] workflows[2017.2] = ['SingleMuPt10_UP17', ['SingleMuPt10FS_UP17','HARVESTUP17FS','MINIAODMCUP17FS']] workflows[2017.3] = ['SingleMuPt100_UP17', ['SingleMuPt100FS_UP17','HARVESTUP17FS','MINIAODMCUP17FS']] workflows[2017.4] = ['ZEE_13_UP17', ['ZEEFS_13_UP17','HARVESTUP17FS','MINIAODMCUP17FS']] workflows[2017.5] = ['ZTT_13_UP17',['ZTTFS_13_UP17','HARVESTUP17FS','MINIAODMCUP17FS']] workflows[2017.6] = ['QCD_FlatPt_15_3000_13_UP17', ['QCDFlatPt153000FS_13_UP17','HARVESTUP17FS','MINIAODMCUP17FS']] workflows[2017.7] = ['H125GGgluonfusion_13_UP17', ['H125GGgluonfusionFS_13_UP17','HARVESTUP17FS','MINIAODMCUP17FS']] workflows[2017.9] = ['ZMM_13_UP17',['ZMMFS_13_UP17','HARVESTUP17FS','MINIAODMCUP17FS']] workflows[2017.11] = ['SMS-T1tttt_mGl-1500_mLSP-100_13_UP17', ['SMS-T1tttt_mGl-1500_mLSP-100FS_13_UP17','HARVESTUP17FS','MINIAODMCUP17FS']] workflows[2017.12] = ['QCD_Pt_80_120_13_UP17', ['QCD_Pt_80_120FS_13_UP17','HARVESTUP17FS','MINIAODMCUP17FS']] workflows[2017.13] = ['TTbar_13_UP17', ['TTbarFS_13_trackingOnlyValidation_UP17','HARVESTUP17FS_trackingOnly']] ### MinBias fastsim_13 TeV for mixing ### workflows[2017.8] = ['',['MinBiasFS_13_UP17_ForMixing']] ### MinBias fastsim_13 TeV for mixing, 2018 ### workflows[2018.8] = ['',['MinBiasFS_13_UP18_ForMixing']] ### Phase1 FastSim 13TeV, 2018 ### workflows[2018.1] = ['TTbar_13_UP18', ['TTbarFS_13_UP18','HARVESTUP18FS','MINIAODMCUP18FS']] workflows[2018.2] = ['SingleMuPt10_UP18', ['SingleMuPt10FS_UP18','HARVESTUP18FS','MINIAODMCUP18FS']] workflows[2018.3] = ['SingleMuPt100_UP18', ['SingleMuPt100FS_UP18','HARVESTUP18FS','MINIAODMCUP18FS']] workflows[2018.4] = ['ZEE_13_UP18', ['ZEEFS_13_UP18','HARVESTUP18FS','MINIAODMCUP18FS']] workflows[2018.5] = ['ZTT_13_UP18',['ZTTFS_13_UP18','HARVESTUP18FS','MINIAODMCUP18FS']] workflows[2018.6] = ['QCD_FlatPt_15_3000_13_UP18', ['QCDFlatPt153000FS_13_UP18','HARVESTUP18FS','MINIAODMCUP18FS']] workflows[2018.7] = ['H125GGgluonfusion_13_UP18', ['H125GGgluonfusionFS_13_UP18','HARVESTUP18FS','MINIAODMCUP18FS']] workflows[2018.9] = ['ZMM_13_UP18',['ZMMFS_13_UP18','HARVESTUP18FS','MINIAODMCUP18FS']] workflows[2018.11] = ['SMS-T1tttt_mGl-1500_mLSP-100_13_UP18', ['SMS-T1tttt_mGl-1500_mLSP-100FS_13_UP18','HARVESTUP18FS','MINIAODMCUP18FS']] workflows[2018.12] = ['QCD_Pt_80_120_13_UP18', ['QCD_Pt_80_120FS_13_UP18','HARVESTUP18FS','MINIAODMCUP18FS']] workflows[2018.13] = ['TTbar_13_UP18', ['TTbarFS_13_trackingOnlyValidation_UP18','HARVESTUP18FS_trackingOnly']] ### standard set ### ## particle guns workflows[15] = ['', ['SingleElectronPt10','DIGI','RECO','HARVEST']] workflows[16] = ['', ['SingleElectronPt1000','DIGI','RECO','HARVEST']] workflows[17] = ['', ['SingleElectronPt35','DIGI','RECO','HARVEST']] workflows[18] = ['', ['SingleGammaPt10','DIGI','RECO','HARVEST']] workflows[19] = ['', ['SingleGammaPt35','DIGI','RECO','HARVEST']] workflows[6] = ['', ['SingleMuPt1','DIGI','RECO','HARVEST']] workflows[20] = ['', ['SingleMuPt10','DIGI','RECO','HARVEST']] workflows[21] = ['', ['SingleMuPt100','DIGI','RECO','HARVEST']] workflows[22] = ['', ['SingleMuPt1000','DIGI','RECO','HARVEST']] ## particle guns postLS1 workflows[1315] = ['', ['SingleElectronPt10_UP15','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1316] = ['', ['SingleElectronPt1000_UP15','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1317] = ['', ['SingleElectronPt35_UP15','DIGIUP15','RECOUP15','HARVESTUP15','NANOUP15']] workflows[1318] = ['', ['SingleGammaPt10_UP15','DIGIUP15','RECOUP15','HARVESTUP15','NANOUP15']] workflows[1319] = ['', ['SingleGammaPt35_UP15','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1306] = ['', ['SingleMuPt1_UP15','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1320] = ['', ['SingleMuPt10_UP15','DIGIUP15','RECOUP15','HARVESTUP15','NANOUP15']] workflows[1321] = ['', ['SingleMuPt100_UP15','DIGIUP15','RECOUP15','HARVESTUP15','NANOUP15']] workflows[1322] = ['', ['SingleMuPt1000_UP15','DIGIUP15','RECOUP15','HARVESTUP15','NANOUP15']] workflows[1323] = ['', ['NuGun_UP15','DIGIUP15','RECOUP15','HARVESTUP15']] ## 8 TeV workflows[24] = ['', ['TTbarLepton','DIGI','RECO','HARVEST']] workflows[35] = ['', ['Wjet_Pt_80_120','DIGI','RECO','HARVEST']] workflows[36] = ['', ['Wjet_Pt_3000_3500','DIGI','RECO','HARVEST']] workflows[37] = ['', ['LM1_sfts','DIGI','RECO','HARVEST']] # the input for the following worrkflow is high statistics workflows[38] = ['', ['QCD_FlatPt_15_3000HS','DIGI','RECO','HARVEST']] workflows[9] = ['', ['Higgs200ChargedTaus','DIGI','RECO','HARVEST']] workflows[13] = ['', ['QCD_Pt_3000_3500','DIGI','RECO','HARVEST']] workflows.addOverride(13,overridesEv5) workflows[39] = ['', ['QCD_Pt_600_800','DIGI','RECO','HARVEST']] workflows[23] = ['', ['JpsiMM','DIGI','RECO','HARVEST']] workflows[25] = ['', ['TTbar','DIGI','RECOAlCaCalo','HARVEST','ALCATT']] workflows[26] = ['', ['WE','DIGI','RECOAlCaCalo','HARVEST']] workflows[29] = ['', ['ZEE','DIGI','RECOAlCaCalo','HARVEST']] workflows[31] = ['', ['ZTT','DIGI','RECO','HARVEST']] workflows[32] = ['', ['H130GGgluonfusion','DIGI','RECO','HARVEST']] workflows[33] = ['', ['PhotonJets_Pt_10','DIGI','RECO','HARVEST']] workflows[34] = ['', ['QQH1352T','DIGI','RECO','HARVEST']] #workflows[46] = ['', ['ZmumuJets_Pt_20_300']] workflows[7] = ['', ['Cosmics','DIGICOS','RECOCOS','ALCACOS','HARVESTCOS']] workflows[7.1]= ['', ['CosmicsSPLoose','DIGICOS','RECOCOS','ALCACOS','HARVESTCOS']] workflows[7.20] = ['', ['Cosmics_UP18','DIGICOS_UP18','RECOCOS_UP18','ALCACOS_UP18','HARVESTCOS_UP18']]#2018 workflows[7.21] = ['', ['Cosmics_UP17','DIGICOS_UP17','RECOCOS_UP17','ALCACOS_UP17','HARVESTCOS_UP17']]#2017 workflows[7.22] = ['', ['Cosmics_UP16','DIGICOS_UP16','RECOCOS_UP16','ALCACOS_UP16','HARVESTCOS_UP16']]#2016 workflows[7.23] = ['', ['Cosmics_UP21','DIGICOS_UP21','RECOCOS_UP21','ALCACOS_UP21','HARVESTCOS_UP21']]#2021 workflows[7.24] = ['', ['Cosmics_UP21_0T','DIGICOS_UP21_0T','RECOCOS_UP21_0T','ALCACOS_UP21_0T','HARVESTCOS_UP21_0T']]#2021 0T workflows[7.3] = ['', ['CosmicsSPLoose_UP18','DIGICOS_UP18','RECOCOS_UP18','ALCACOS_UP18','HARVESTCOS_UP18']] workflows[7.4] = ['', ['Cosmics_UP18','DIGICOSPEAK_UP18','RECOCOSPEAK_UP18','ALCACOS_UP18','HARVESTCOS_UP18']] workflows[8] = ['', ['BeamHalo','DIGICOS','RECOCOS','ALCABH','HARVESTCOS']] workflows[8.1] = ['', ['BeamHalo_UP18','DIGICOS_UP18','RECOCOS_UP18','ALCABH_UP18','HARVESTCOS_UP18']] workflows[8.2] = ['', ['BeamHalo_UP21','DIGICOS_UP21','RECOCOS_UP21','ALCABH_UP21','HARVESTCOS_UP21']] workflows[11] = ['', ['MinBias','DIGI','RECOMIN','HARVEST','ALCAMIN']] workflows[28] = ['', ['QCD_Pt_80_120','DIGI','RECO','HARVEST']] workflows[27] = ['', ['WM','DIGI','RECO','HARVEST']] workflows[30] = ['', ['ZMM','DIGI','RECO','HARVEST']] workflows[10] = ['', ['ADDMonoJet_d3MD3','DIGI','RECO','HARVEST']] workflows[12] = ['', ['ZpMM','DIGI','RECO','HARVEST']] workflows[14] = ['', ['WpM','DIGI','RECO','HARVEST']] workflows[43] = ['', ['ZpMM_2250_8TeV','DIGI','RECO','HARVEST']] workflows[44] = ['', ['ZpEE_2250_8TeV','DIGI','RECO','HARVEST']] workflows[45] = ['', ['ZpTT_1500_8TeV','DIGI','RECO','HARVEST']] ## 13 TeV and postLS1 geometry workflows[1324] = ['', ['TTbarLepton_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1335] = ['', ['Wjet_Pt_80_120_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1336] = ['', ['Wjet_Pt_3000_3500_13','DIGIUP15','RECOUP15','HARVESTUP15']] #workflows[1337] = ['', ['LM1_sfts_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1337] = ['', ['SMS-T1tttt_mGl-1500_mLSP-100_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1338] = ['', ['QCD_FlatPt_15_3000HS_13','DIGIUP15','RECOUP15','HARVESTUP15','NANOUP15']] workflows[1309] = ['', ['Higgs200ChargedTaus_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1313] = ['', ['QCD_Pt_3000_3500_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows.addOverride(1313,overridesEv5) workflows[1339] = ['', ['QCD_Pt_600_800_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1347] = ['', ['Upsilon1SToMuMu_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1349] = ['', ['BsToMuMu_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1350] = ['', ['JpsiMuMu_Pt-8','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1364] = ['', ['BdToMuMu_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1365] = ['', ['BuToJpsiK_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1366] = ['', ['BsToJpsiPhi_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1325] = ['', ['TTbar_13','DIGIUP15','RECOUP15','HARVESTUP15','ALCATTUP15','NANOUP15']] # the 3 workflows below are for tracking-specific ib test, not to be run in standard relval set. workflows[1325.1] = ['', ['TTbar_13','DIGIUP15','RECOUP15_trackingOnly','HARVESTUP15_trackingOnly']] workflows[1325.2] = ['', ['TTbar_13','DIGIUP15','RECOUP15_trackingLowPU','HARVESTUP15']] workflows[1325.3] = ['', ['TTbar_13','DIGIUP15','RECOUP15_trackingOnlyLowPU','HARVESTUP15_trackingOnly']] workflows[1325.4] = ['', ['TTbar_13','DIGIUP15','RECOUP15_HIPM','HARVESTUP15']] # reminiaod wf on 80X MC workflows[1325.5] = ['', ['ProdZEE_13_reminiaodINPUT','REMINIAOD_mc2016','HARVESTDR2_REMINIAOD_mc2016']] # reminiaod wf on 94X MC workflows[1325.51] = ['', ['TTbar_13_94XreminiaodINPUT','REMINIAOD_mc2017','HARVESTUP17_REMINIAOD_mc2017']] # reminiaod on UL MC workflows[1325.516] = ['', ['TTbar_13_reminiaod2016UL_preVFP_INPUT','REMINIAOD_mc2016UL_preVFP','HARVESTDR2_REMINIAOD_mc2016UL_preVFP']] workflows[1325.5161] = ['', ['TTbar_13_reminiaod2016UL_postVFP_INPUT','REMINIAOD_mc2016UL_postVFP','HARVESTDR2_REMINIAOD_mc2016UL_postVFP']] workflows[1325.517] = ['', ['TTbar_13_reminiaod2017UL_INPUT','REMINIAOD_mc2017UL','HARVESTUP17_REMINIAOD_mc2017UL']] workflows[1325.518] = ['', ['TTbar_13_reminiaod2018UL_INPUT','REMINIAOD_mc2018UL','HARVESTUP18_REMINIAOD_mc2018UL']] # nanoaod wf without intermediate EDM, starting from existing MINIAOD inputs workflows[1325.6] = ['', ['TTbar_13_94Xv1NanoAODINPUT','NANOAODMC2017_94XMiniAODv1']] # nanoaod wf with intermediate EDM and merge step, starting from existing MINIAOD inputs workflows[1325.7] = ['', ['TTbar_13_94Xv2NanoAODINPUT','NANOEDMMC2017_94XMiniAODv2','HARVESTNANOAODMC2017_94XMiniAODv2']] workflows[1325.8] = ['', ['TTbar_13_94Xv1NanoAODINPUT','NANOEDMMC2017_94XMiniAODv1','HARVESTNANOAODMC2017_94XMiniAODv1']] # special workflow including the SiStrip APV dynamic gain simulation for 2016 MC workflows[1325.9] = ['', ['TTbar_13','DIGIUP15APVSimu','RECOUP15','HARVESTUP15','ALCATTUP15']] workflows[1325.91] = ['', ['TTbar_13','DIGIUP15APVSimu','RECOUP15_HIPM','HARVESTUP15','ALCATTUP15']] #using ZEE as I cannot find TT at CERN workflows[1329.1] = ['', ['ZEE_13_80XNanoAODINPUT','NANOEDMMC2016_80X','HARVESTNANOAODMC2016_80X']] workflows[1326] = ['', ['WE_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1329] = ['', ['ZEE_13','DIGIUP15','RECOUP15_L1TEgDQM','HARVESTUP15_L1TEgDQM','NANOUP15']] workflows[1356] = ['', ['ZEE_13_DBLMINIAOD','DIGIUP15','RECOAODUP15','HARVESTUP15','DBLMINIAODMCUP15NODQM']] workflows[1331] = ['', ['ZTT_13','DIGIUP15','RECOUP15','HARVESTUP15','NANOUP15']] workflows[1332] = ['', ['H125GGgluonfusion_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1333] = ['', ['PhotonJets_Pt_10_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1334] = ['', ['QQH1352T_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1307] = ['', ['CosmicsSPLoose_UP15','DIGICOS_UP15','RECOCOS_UP15','ALCACOS_UP15','HARVESTCOS_UP15']] workflows[1308] = ['', ['BeamHalo_13','DIGIHAL','RECOHAL','ALCAHAL','HARVESTHAL']] workflows[1311] = ['', ['MinBias_13','DIGIUP15','RECOMINUP15','HARVESTMINUP15','ALCAMINUP15']] workflows[1328] = ['', ['QCD_Pt_80_120_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1327] = ['', ['WM_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1330] = ['', ['ZMM_13','DIGIUP15','RECOUP15_L1TMuDQM','HARVESTUP15_L1TMuDQM','NANOUP15']] workflows[1310] = ['', ['ADDMonoJet_d3MD3_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1312] = ['', ['ZpMM_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1314] = ['', ['WpM_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1340] = ['', ['PhiToMuMu_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1341] = ['', ['RSKKGluon_m3000GeV_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1343] = ['', ['ZpMM_2250_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1344] = ['', ['ZpEE_2250_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1345] = ['', ['ZpTT_1500_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1348] = ['', ['EtaBToJpsiJpsi_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1351] = ['', ['BuMixing_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1352] = ['', ['HSCPstop_M_200_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1353] = ['', ['RSGravitonToGaGa_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1354] = ['', ['WpToENu_M-2000_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1355] = ['', ['DisplacedSUSY_stopToBottom_M_300_1000mm_13','DIGIUP15','RECOUP15','HARVESTUP15']] # fullSim 13TeV normal workflows starting from gridpacks LHE generation workflows[1360] = ['', ['TTbar012Jets_NLO_Mad_py8_Evt_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1361] = ['', ['GluGluHToZZTo4L_M125_Pow_py8_Evt_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1362] = ['', ['VBFHToZZTo4Nu_M125_Pow_py8_Evt_13','DIGIUP15','RECOUP15','HARVESTUP15']] workflows[1363] = ['', ['VBFHToBB_M125_Pow_py8_Evt_13','DIGIUP15','RECOUP15','HARVESTUP15']] # 2017 workflows starting from gridpacks LHE generation workflows[1361.17] = ['', ['GluGluHToZZTo4L_M125_Pow_py8_Evt_13UP17','DIGIUP17','RECOUP17','HARVESTUP17','NANOUP17']] workflows[1362.17] = ['', ['VBFHToZZTo4Nu_M125_Pow_py8_Evt_13UP17','DIGIUP17','RECOUP17','HARVESTUP17','NANOUP17']] workflows[1363.17] = ['', ['VBFHToBB_M125_Pow_py8_Evt_13UP17','DIGIUP17','RECOUP17','HARVESTUP17','NANOUP17']] # 2018 workflows starting from gridpacks LHE generation workflows[1361.18] = ['', ['GluGluHToZZTo4L_M125_Pow_py8_Evt_13UP18','DIGIUP18','RECOUP18','HARVESTUP18','NANOUP18']] workflows[1362.18] = ['', ['VBFHToZZTo4Nu_M125_Pow_py8_Evt_13UP18','DIGIUP18','RECOUP18','HARVESTUP18','NANOUP18']] workflows[1363.18] = ['', ['VBFHToBB_M125_Pow_py8_Evt_13UP18','DIGIUP18','RECOUP18','HARVESTUP18','NANOUP18']] #2018 workflows starting from gridpacks LHE generation with multiple concurrent lumi sections workflows[1361.181] = ['', ['GluGluHToZZTo4L_M125_Pow_py8_Evt_13UP18ml','DIGIUP18ml','RECOUP18ml','HARVESTUP18','NANOUP18ml']] # fullSim 13TeV normal workflows starting from pLHE workflows[1370] = ['', ['GluGluHToGG_M125_Pow_MINLO_NNLOPS_py8_13','Hadronizer_TuneCUETP8M1_13TeV_powhegEmissionVeto_2p_HToGG_M125_13','DIGIUP15','RECOUP15','HARVESTUP15','NANOUP15Had']] # 2017 workflows starting from pLHE workflows[1370.17] = ['', ['GluGluHToGG_M125_Pow_MINLO_NNLOPS_py8_13UP17','Hadronizer_TuneCUETP8M1_13TeV_powhegEmissionVeto_2p_HToGG_M125_13UP17','DIGIUP17','RECOUP17','HARVESTUP17','NANOUP17Had']] # 2018 workflows starting from pLHE workflows[1370.18] = ['', ['GluGluHToGG_M125_Pow_MINLO_NNLOPS_py8_13UP18','Hadronizer_TuneCUETP8M1_13TeV_powhegEmissionVeto_2p_HToGG_M125_13UP18','DIGIUP18','RECOUP18','HARVESTUP18','NANOUP18Had']] ### HI test ### ### Run I cond., 2011 workflows[140] = ['',['HydjetQ_B12_5020GeV_2011','DIGIHI2011','RECOHI2011','HARVESTHI2011']] ### Run II cond., 2015 workflows[145] = ['',['HydjetQ_B12_5020GeV_2015','DIGIHI2015','RECOHI2015','HARVESTHI2015']] ### Run II cond., 2017 workflows[148] = ['',['HydjetQ_MinBias_XeXe_5442GeV_2017','DIGIHI2017','RECOHI2017','HARVESTHI2017']] ### Run II cond., 2018 workflows[150] = ['',['HydjetQ_B12_5020GeV_2018','DIGIHI2018','RECOHI2018','HARVESTHI2018']] workflows[158] = ['',['HydjetQ_B12_5020GeV_2018_ppReco','DIGIHI2018PPRECO','RECOHI2018PPRECO','ALCARECOHI2018PPRECO','HARVESTHI2018PPRECO']] workflows[158.01] = ['',['HydjetQ_reminiaodPbPb2018_INPUT','REMINIAODHI2018PPRECO','HARVESTHI2018PPRECOMINIAOD']] workflows[158.1] = ['',['QCD_Pt_80_120_13_HI','DIGIHI2018PPRECO','RECOHI2018PPRECO','HARVESTHI2018PPRECO']] workflows[158.2] = ['',['PhotonJets_Pt_10_13_HI','DIGIHI2018PPRECO','RECOHI2018PPRECO','HARVESTHI2018PPRECO']] workflows[158.3] = ['',['ZEEMM_13_HI','DIGIHI2018PPRECO','RECOHI2018PPRECO','HARVESTHI2018PPRECO']] # 158.0 with multiple concurrent lumi sections workflows[158.181] = ['',['HydjetQ_B12_5020GeV_2018_ppReco_ml','DIGIHI2018PPRECOml','RECOHI2018PPRECOml','ALCARECOHI2018PPRECOml','HARVESTHI2018PPRECO']] workflows[159] = ['',['HydjetQ_B12_5020GeV_2021_ppReco','DIGIHI2021PPRECO','RECOHI2021PPRECO','ALCARECOHI2021PPRECO','HARVESTHI2021PPRECO']] workflows[159.1] = ['',['QCD_Pt_80_120_14_HI_2021','DIGIHI2021PPRECO','RECOHI2021PPRECO','HARVESTHI2021PPRECO']] workflows[159.2] = ['',['PhotonJets_Pt_10_14_HI_2021','DIGIHI2021PPRECO','RECOHI2021PPRECO','HARVESTHI2021PPRECO']] workflows[159.3] = ['',['ZMM_14_HI_2021','DIGIHI2021PPRECO','RECOHI2021PPRECO','HARVESTHI2021PPRECO']] workflows[159.4] = ['',['ZEE_14_HI_2021','DIGIHI2021PPRECO','RECOHI2021PPRECO','HARVESTHI2021PPRECO']] ### pp reference test ### workflows[149] = ['',['QCD_Pt_80_120_13_PPREF','DIGIPPREF2017','RECOPPREF2017','HARVESTPPREF2017']] ### pPb test ### workflows[280]= ['',['AMPT_PPb_5020GeV_MinimumBias','DIGI','RECO','HARVEST']] ### pPb Run2 ### workflows[281]= ['',['EPOS_PPb_8160GeV_MinimumBias','DIGIUP15_PPb','RECOUP15_PPb','HARVESTUP15_PPb']]
78.869263
197
0.736858
794d6d655d57a074625fb041cb97affbf4857f63
179
py
Python
rasalit/apps/nlucluster/common.py
vishnupriyavr/rasalit
8305aa1a4cf2f3d7d79e989b4656d1995c420531
[ "Apache-2.0" ]
261
2020-04-17T14:33:21.000Z
2022-03-30T06:53:24.000Z
rasalit/apps/nlucluster/common.py
vishnupriyavr/rasalit
8305aa1a4cf2f3d7d79e989b4656d1995c420531
[ "Apache-2.0" ]
36
2020-07-21T18:39:32.000Z
2022-02-10T10:35:57.000Z
rasalit/apps/nlucluster/common.py
vishnupriyavr/rasalit
8305aa1a4cf2f3d7d79e989b4656d1995c420531
[ "Apache-2.0" ]
53
2020-04-20T16:49:51.000Z
2022-03-10T09:03:32.000Z
# Because of how tensorflow works (unfortunately!) we can't seperate our concerns as much # as we would like them to. This is mainly due to the `tf.global_variables_initializer`.
59.666667
89
0.782123
794d6daea95bcebd0f9b7ebdcbd8865b40ad8b96
556
py
Python
production/JVcode/Scripts/normalizer.py
ramaganapathy1/AMuDA-Ir-back-end
179bf52fd303440fff661c074c7835c4932d5165
[ "MIT" ]
2
2017-09-20T07:47:08.000Z
2017-09-27T12:26:11.000Z
production/JVcode/Scripts/normalizer.py
ramaganapathy1/AMuDA-Ir-back-end
179bf52fd303440fff661c074c7835c4932d5165
[ "MIT" ]
null
null
null
production/JVcode/Scripts/normalizer.py
ramaganapathy1/AMuDA-Ir-back-end
179bf52fd303440fff661c074c7835c4932d5165
[ "MIT" ]
2
2017-09-27T12:26:13.000Z
2018-12-11T00:37:53.000Z
import os #listOfFiles = os.listdir(os.getcwd()+"/tabfiles") highestT = 0.0 highestD = 0.0 highestC = 0.0 for filename in os.listdir(os.getcwd()): if filename.endswith(".tab"): doc = open(filename,"r") for line in doc: elements = line.split("\t") tfidf = float(elements[1]) cscore = float(elements[3]) dspan = float(elements[2]) if tfidf >= highestT: highestT = tfidf if cscore >= highestC: highestC = cscore if dspan >= highestD: highestD = dspan doc.close() print highestT,"\t", highestD,"\t", highestC,"\n"
21.384615
50
0.638489
794d6f271392ef9919fe81b680455ec4bb835311
1,517
py
Python
doge/util.py
lenards/pretty-doge
cfd4ba8e607f2958c3ef204b6b91ebf3ac5db69f
[ "BSD-3-Clause" ]
2
2017-09-05T12:39:56.000Z
2021-12-06T05:23:13.000Z
doge/util.py
lenards/pretty-doge
cfd4ba8e607f2958c3ef204b6b91ebf3ac5db69f
[ "BSD-3-Clause" ]
8
2021-06-08T19:01:47.000Z
2022-03-11T23:16:34.000Z
doge/util.py
lenards/pretty-doge
cfd4ba8e607f2958c3ef204b6b91ebf3ac5db69f
[ "BSD-3-Clause" ]
null
null
null
""" Utility methods for pretty-doge. NOTE: These methods were used in development but are not currently used. """ WHITE = (255, 255, 255) BLACK = (0, 0, 0) def choose(color): """ Used to convert the png into 4 color. """ r, g, b, a = color if r < 90: return (0, 0, 0, 255) if r >= 77 and r < 160: return (134, 30, 214, 255) if r >= 160 and r < 208: return (172, 143, 239, 255) else: return (255, 255, 255, 255) def put_orig(self, n): im = self._image.copy() new_light, new_dark = get_pallete(n, LIGHT, DARK) c = 0 for i in range(0, 240): for j in range(0, 240): p = im.getpixel((i, j)) if p == LIGHT: c += 1 im.putpixel((i, j), new_light) else: if p == DARK: c += 1 im.putpixel((i, j), new_dark) return im def put_n(self, n): new_light, new_dark = get_pallete(n, LIGHT, DARK) return self.put(new_light, new_dark) def get_h(h, n): if n < 50: return h - h * (n / 100.) else: return h + h * (n / 100.) def get_pallete(n, l, d): hls = rgb_to_hls(l[0]/255., l[1]/255., l[2]/255.) hls_d = rgb_to_hls(d[0]/255., d[1]/255., d[2]/255.) h_mod = get_h(hls[0], n) h_mod_d = get_h(hls_d[0], n) new_light = scrub_rgb(hls_to_rgb(h_mod, hls[1], hls[2])) new_dark = scrub_rgb(hls_to_rgb(h_mod_d, hls_d[1], hls_d[2])) return new_light, new_dark
23.338462
72
0.520105
794d7071bc957713b62b2adb8e5a2cf35df29198
1,290
py
Python
aiokraken/websockets/schemas/tests/strats/st_owntrade.py
asmodehn/aiokraken
b260bd41d5aa091e6a4f1818328426fbe6f625c0
[ "MIT" ]
null
null
null
aiokraken/websockets/schemas/tests/strats/st_owntrade.py
asmodehn/aiokraken
b260bd41d5aa091e6a4f1818328426fbe6f625c0
[ "MIT" ]
82
2019-08-30T09:37:49.000Z
2022-03-29T14:53:22.000Z
aiokraken/websockets/schemas/tests/strats/st_owntrade.py
asmodehn/aiokraken
b260bd41d5aa091e6a4f1818328426fbe6f625c0
[ "MIT" ]
null
null
null
import hypothesis.strategies as st from aiokraken.websockets.schemas.owntrades import ownTradeWS, ownTradeWSSchema @st.composite def st_owntradews(draw): return ownTradeWS( tradeid= draw(st.text()), ordertxid= draw(st.text()), postxid= draw(st.text()), pair= draw(st.text()), time= draw(st.floats(allow_nan=False, allow_infinity=False)), type= draw(st.text()), ordertype= draw(st.text()), price= draw(st.floats(allow_nan=False, allow_infinity=False)), cost= draw(st.floats(allow_nan=False, allow_infinity=False)), fee= draw(st.floats(allow_nan=False, allow_infinity=False)), vol= draw(st.floats(allow_nan=False, allow_infinity=False)), margin=draw(st.floats(allow_nan=False, allow_infinity=False)), # seems to have been added recently (some trades dont have it, despite having a postxid...) posstatus=draw(st.one_of(st.text(), st.none())), ) @st.composite def st_owntradewsdict(draw): model = draw(st_owntradews()) schema = ownTradeWSSchema() return schema.dump(model) if __name__ == '__main__': for n in range(1, 10): print(repr(st_owntradews().example())) for n in range(1, 10): print(repr(st_owntradewsdict().example()))
31.463415
99
0.66124
794d757c91adb0a57c2d40b45ab1f02ad8238b36
484
py
Python
circles/migrations/0009_auto_20200505_2253.py
CoronaCircles/coronacircles
1d86a659def276fcc0ebd7a723a48f729d6e21a1
[ "MIT" ]
1
2020-04-30T16:19:33.000Z
2020-04-30T16:19:33.000Z
circles/migrations/0009_auto_20200505_2253.py
CoronaCircles/coronacircles
1d86a659def276fcc0ebd7a723a48f729d6e21a1
[ "MIT" ]
43
2020-05-01T23:48:01.000Z
2022-03-12T00:35:32.000Z
circles/migrations/0009_auto_20200505_2253.py
CoronaCircles/coronacircles
1d86a659def276fcc0ebd7a723a48f729d6e21a1
[ "MIT" ]
1
2020-05-04T20:07:16.000Z
2020-05-04T20:07:16.000Z
# Generated by Django 3.0.6 on 2020-05-05 20:53 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('circles', '0008_auto_20200505_0026'), ] operations = [ migrations.AlterField( model_name='mailtemplate', name='language_code', field=models.CharField(choices=[('de', 'German'), ('en', 'English')], default='en', max_length=2, verbose_name='Language'), ), ]
25.473684
135
0.61157
794d75bff484bb7546bd96eeec4a7b315ac79023
716
py
Python
tests/integration/long/__init__.py
fatelei/python-driver
3bddef6185f2691e1713dfe51d1fa26d1555724c
[ "Apache-2.0" ]
1
2017-10-17T11:30:52.000Z
2017-10-17T11:30:52.000Z
tests/integration/long/__init__.py
fatelei/python-driver
3bddef6185f2691e1713dfe51d1fa26d1555724c
[ "Apache-2.0" ]
null
null
null
tests/integration/long/__init__.py
fatelei/python-driver
3bddef6185f2691e1713dfe51d1fa26d1555724c
[ "Apache-2.0" ]
null
null
null
# Copyright 2013-2015 DataStax, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. try: from ccmlib import common except ImportError as e: raise unittest.SkipTest('ccm is a dependency for integration tests:', e)
39.777778
76
0.76257
794d7613cdc509271eeabd7574a5f7ce9ed5f632
6,824
py
Python
numba/core/typing/typeof.py
auderson/numba
3d67c9850ab56457f418cf40af6245fd9c337705
[ "BSD-2-Clause" ]
6,620
2015-01-04T08:51:04.000Z
2022-03-31T12:52:18.000Z
numba/core/typing/typeof.py
auderson/numba
3d67c9850ab56457f418cf40af6245fd9c337705
[ "BSD-2-Clause" ]
6,457
2015-01-04T03:18:41.000Z
2022-03-31T17:38:42.000Z
numba/core/typing/typeof.py
auderson/numba
3d67c9850ab56457f418cf40af6245fd9c337705
[ "BSD-2-Clause" ]
930
2015-01-25T02:33:03.000Z
2022-03-30T14:10:32.000Z
from collections import namedtuple from functools import singledispatch import ctypes import enum import numpy as np from numba.core import types, utils, errors from numba.np import numpy_support # terminal color markup _termcolor = errors.termcolor() class Purpose(enum.Enum): # Value being typed is used as an argument argument = 1 # Value being typed is used as a constant constant = 2 _TypeofContext = namedtuple("_TypeofContext", ("purpose",)) def typeof(val, purpose=Purpose.argument): """ Get the Numba type of a Python value for the given purpose. """ # Note the behaviour for Purpose.argument must match _typeof.c. c = _TypeofContext(purpose) ty = typeof_impl(val, c) if ty is None: msg = _termcolor.errmsg( f"Cannot determine Numba type of {type(val)}") raise ValueError(msg) return ty @singledispatch def typeof_impl(val, c): """ Generic typeof() implementation. """ tp = _typeof_buffer(val, c) if tp is not None: return tp # cffi is handled here as it does not expose a public base class # for exported functions or CompiledFFI instances. from numba.core.typing import cffi_utils if cffi_utils.SUPPORTED: if cffi_utils.is_cffi_func(val): return cffi_utils.make_function_type(val) if cffi_utils.is_ffi_instance(val): return types.ffi return getattr(val, "_numba_type_", None) def _typeof_buffer(val, c): from numba.core.typing import bufproto try: m = memoryview(val) except TypeError: return # Object has the buffer protocol try: dtype = bufproto.decode_pep3118_format(m.format, m.itemsize) except ValueError: return type_class = bufproto.get_type_class(type(val)) layout = bufproto.infer_layout(m) return type_class(dtype, m.ndim, layout=layout, readonly=m.readonly) @typeof_impl.register(ctypes._CFuncPtr) def _typeof_ctypes_function(val, c): from .ctypes_utils import is_ctypes_funcptr, make_function_type if is_ctypes_funcptr(val): return make_function_type(val) @typeof_impl.register(type) def _typeof_type(val, c): """ Type various specific Python types. """ if issubclass(val, BaseException): return types.ExceptionClass(val) if issubclass(val, tuple) and hasattr(val, "_asdict"): return types.NamedTupleClass(val) if issubclass(val, np.generic): return types.NumberClass(numpy_support.from_dtype(val)) if issubclass(val, types.Type): return types.TypeRef(val) from numba.typed import Dict if issubclass(val, Dict): return types.TypeRef(types.DictType) from numba.typed import List if issubclass(val, List): return types.TypeRef(types.ListType) @typeof_impl.register(bool) def _typeof_bool(val, c): return types.boolean @typeof_impl.register(float) def _typeof_float(val, c): return types.float64 @typeof_impl.register(complex) def _typeof_complex(val, c): return types.complex128 @typeof_impl.register(int) def _typeof_int(val, c): # As in _typeof.c nbits = utils.bit_length(val) if nbits < 32: typ = types.intp elif nbits < 64: typ = types.int64 elif nbits == 64 and val >= 0: typ = types.uint64 else: raise ValueError("Int value is too large: %s" % val) return typ @typeof_impl.register(np.generic) def _typeof_numpy_scalar(val, c): try: return numpy_support.map_arrayscalar_type(val) except NotImplementedError: pass @typeof_impl.register(str) def _typeof_str(val, c): return types.string @typeof_impl.register(type((lambda a: a).__code__)) def _typeof_code(val, c): return types.code_type @typeof_impl.register(type(None)) def _typeof_none(val, c): return types.none @typeof_impl.register(type(Ellipsis)) def _typeof_ellipsis(val, c): return types.ellipsis @typeof_impl.register(tuple) def _typeof_tuple(val, c): tys = [typeof_impl(v, c) for v in val] if any(ty is None for ty in tys): return return types.BaseTuple.from_types(tys, type(val)) @typeof_impl.register(list) def _typeof_list(val, c): if len(val) == 0: raise ValueError("Cannot type empty list") ty = typeof_impl(val[0], c) if ty is None: raise ValueError( f"Cannot type list element type {type(val[0])}") return types.List(ty, reflected=True) @typeof_impl.register(set) def _typeof_set(val, c): if len(val) == 0: raise ValueError("Cannot type empty set") item = next(iter(val)) ty = typeof_impl(item, c) if ty is None: raise ValueError( f"Cannot type set element type {type(item)}") return types.Set(ty, reflected=True) @typeof_impl.register(slice) def _typeof_slice(val, c): return types.slice2_type if val.step in (None, 1) else types.slice3_type @typeof_impl.register(enum.Enum) @typeof_impl.register(enum.IntEnum) def _typeof_enum(val, c): clsty = typeof_impl(type(val), c) return clsty.member_type @typeof_impl.register(enum.EnumMeta) def _typeof_enum_class(val, c): cls = val members = list(cls.__members__.values()) if len(members) == 0: raise ValueError("Cannot type enum with no members") dtypes = {typeof_impl(mem.value, c) for mem in members} if len(dtypes) > 1: raise ValueError("Cannot type heterogeneous enum: " "got value types %s" % ", ".join(sorted(str(ty) for ty in dtypes))) if issubclass(val, enum.IntEnum): typecls = types.IntEnumClass else: typecls = types.EnumClass return typecls(cls, dtypes.pop()) @typeof_impl.register(np.dtype) def _typeof_dtype(val, c): tp = numpy_support.from_dtype(val) return types.DType(tp) @typeof_impl.register(np.ndarray) def _typeof_ndarray(val, c): try: dtype = numpy_support.from_dtype(val.dtype) except errors.NumbaNotImplementedError: raise errors.NumbaValueError(f"Unsupported array dtype: {val.dtype}") layout = numpy_support.map_layout(val) readonly = not val.flags.writeable return types.Array(dtype, val.ndim, layout, readonly=readonly) @typeof_impl.register(types.NumberClass) def _typeof_number_class(val, c): return val @typeof_impl.register(types.Literal) def _typeof_literal(val, c): return val @typeof_impl.register(types.TypeRef) def _typeof_typeref(val, c): return val @typeof_impl.register(types.Type) def _typeof_nb_type(val, c): if isinstance(val, types.BaseFunction): return val elif isinstance(val, (types.Number, types.Boolean)): return types.NumberClass(val) else: return types.TypeRef(val)
25.462687
77
0.680393
794d778b22da577442583ac2b8aa8c0d89f6c3c7
2,141
py
Python
recipes/cnpy/all/conanfile.py
rockandsalt/conan-center-index
d739adcec3e4dd4c250eff559ceb738e420673dd
[ "MIT" ]
562
2019-09-04T12:23:43.000Z
2022-03-29T16:41:43.000Z
recipes/cnpy/all/conanfile.py
rockandsalt/conan-center-index
d739adcec3e4dd4c250eff559ceb738e420673dd
[ "MIT" ]
9,799
2019-09-04T12:02:11.000Z
2022-03-31T23:55:45.000Z
recipes/cnpy/all/conanfile.py
rockandsalt/conan-center-index
d739adcec3e4dd4c250eff559ceb738e420673dd
[ "MIT" ]
1,126
2019-09-04T11:57:46.000Z
2022-03-31T16:43:38.000Z
import os import glob from conans import ConanFile, CMake, tools class CnpyConan(ConanFile): name = "cnpy" description = "library to read/write .npy and .npz files in C/C++" license = "MIT" topics = ("conan", "cnpy") homepage = "https://github.com/hongyx11/cnpy" url = "https://github.com/conan-io/conan-center-index" exports_sources = ["CMakeLists.txt", "patches/*"] generators = "cmake", "cmake_find_package" settings = "os", "arch", "compiler", "build_type" _cmake = None options = { "shared": [True, False], "fPIC": [True, False] } default_options = { "shared": False, "fPIC": True } @property def _source_subfolder(self): return "source_subfolder" @property def _build_subfolder(self): return "build_subfolder" def config_options(self): if self.settings.os == "Windows": del self.options.fPIC def configure(self): if self.options.shared: del self.options.fPIC def requirements(self): self.requires("zlib/1.2.11") def source(self): tools.get(**self.conan_data["sources"][self.version]) extracted_dir = glob.glob(self.name + "-*")[0] os.rename(extracted_dir, self._source_subfolder) def build(self): for patch in self.conan_data.get("patches", {}).get(self.version, []): tools.patch(**patch) cmake = self._configure_cmake() cmake.build() def _configure_cmake(self): if self._cmake: return self._cmake self._cmake = CMake(self) self._cmake.definitions["ENABLE_STATIC"] = not self.options.shared self._cmake.configure(build_folder=self._build_subfolder) return self._cmake def package(self): self.copy("LICENSE", dst="licenses", src=self._source_subfolder) cmake = self._configure_cmake() cmake.install() def package_info(self): self.cpp_info.names["cmake_find_package"] = "cnpy" self.cpp_info.names["cmake_find_package_multi"] = "cnpy" self.cpp_info.libs = ["cnpy"]
28.932432
78
0.616067
794d7804f3adee4ea5f017562f40f3f5beab5c35
4,002
py
Python
pkgs/ops-pkg/src/genie/libs/ops/nd/iosxr/tests/test_nd.py
jbronikowski/genielibs
200a34e5fe4838a27b5a80d5973651b2e34ccafb
[ "Apache-2.0" ]
94
2018-04-30T20:29:15.000Z
2022-03-29T13:40:31.000Z
pkgs/ops-pkg/src/genie/libs/ops/nd/iosxr/tests/test_nd.py
jbronikowski/genielibs
200a34e5fe4838a27b5a80d5973651b2e34ccafb
[ "Apache-2.0" ]
67
2018-12-06T21:08:09.000Z
2022-03-29T18:00:46.000Z
pkgs/ops-pkg/src/genie/libs/ops/nd/iosxr/tests/test_nd.py
jbronikowski/genielibs
200a34e5fe4838a27b5a80d5973651b2e34ccafb
[ "Apache-2.0" ]
49
2018-06-29T18:59:03.000Z
2022-03-10T02:07:59.000Z
# Python import unittest from unittest.mock import Mock # Pyats from pyats.topology import Device # genie.libs from genie.libs.ops.nd.iosxr.nd import Nd from genie.libs.ops.nd.iosxr.tests.nd_output import NdOutput from genie.libs.parser.iosxr.show_interface import ShowIpv6VrfAllInterface from genie.libs.parser.iosxr.show_ipv6 import ShowIpv6NeighborsDetail, ShowIpv6Neighbors outputs = {} # Set values outputs['show running-config interface'] = NdOutput.ShowRunInterface outputs['show running-config interface GigabitEthernet0/0/0/0.390'] = NdOutput.ShowRunInterface_custom def mapper(key): return outputs[key] class test_nd(unittest.TestCase): def setUp(self): self.device = Device(name='aDevice') self.device.os = 'iosxr' self.device.mapping={} self.device.mapping['cli']='cli' self.device.connectionmgr.connections['cli'] = self.device self.device.execute = Mock() self.device.execute.side_effect = mapper def test_full_nd(self): self.maxDiff = None nd = Nd(device=self.device) nd.maker.outputs[ShowIpv6VrfAllInterface] = {"{'interface':'','vrf':'all'}": NdOutput.ShowIpv6VrfAllInterface} nd.maker.outputs[ShowIpv6Neighbors] = {"{'interface':'','vrf':'all'}": NdOutput.ShowIpv6Neighbors} nd.maker.outputs[ShowIpv6NeighborsDetail] = {"": NdOutput.ShowIpv6NeighborsDetail} # Learn the feature nd.learn() self.assertEqual(nd.info, NdOutput.ndOpsOutput) def test_custom_nd(self): self.maxDiff = None nd = Nd(device=self.device) nd.maker.outputs[ShowIpv6VrfAllInterface] = {"{'interface':'GigabitEthernet0/0/0/0.390','vrf':'VRF1'}": NdOutput.ShowIpv6VrfAllInterface} nd.maker.outputs[ShowIpv6Neighbors] = {"{'interface':'GigabitEthernet0/0/0/0.390','vrf':'VRF1'}": NdOutput.ShowIpv6Neighbors} nd.maker.outputs[ShowIpv6NeighborsDetail] = {"": NdOutput.ShowIpv6NeighborsDetail} # Learn the feature nd.learn(interface='GigabitEthernet0/0/0/0.390', vrf='VRF1') self.assertEqual(nd.info, NdOutput.ndOpsOutput_custom) def test_selective_attribute_nd(self): nd = Nd(device=self.device) nd.maker.outputs[ShowIpv6VrfAllInterface] = {"{'interface':'','vrf':'all'}": NdOutput.ShowIpv6VrfAllInterface} nd.maker.outputs[ShowIpv6Neighbors] = {"{'interface':'','vrf':'all'}": NdOutput.ShowIpv6Neighbors} nd.maker.outputs[ShowIpv6NeighborsDetail] = {"": NdOutput.ShowIpv6NeighborsDetail} # Learn the feature nd.learn() # Check match self.assertEqual('1800', nd.info['interfaces']['GigabitEthernet0/0/0/1.390']['router_advertisement']['lifetime']) # Check does not match self.assertNotEqual('other', nd.info['interfaces']['GigabitEthernet0/0/0/1.390']['neighbors']['fe80::5c00:40ff:fe02:7']['origin']) def test_missing_attributes_nd(self): nd = Nd(device=self.device) nd.maker.outputs[ShowIpv6VrfAllInterface] = {"{'interface':'','vrf':'all'}": {}} nd.maker.outputs[ShowIpv6Neighbors] = {"{'interface':'','vrf':'all'}": NdOutput.ShowIpv6Neighbors} nd.maker.outputs[ShowIpv6NeighborsDetail] = {"": NdOutput.ShowIpv6NeighborsDetail} # Learn the feature nd.learn() with self.assertRaises(KeyError): nd.info['interfaces']['GigabitEthernet0/0/0/1.420']['router_advertisement'] def test_empty_output_nd(self): self.maxDiff = None nd = Nd(device=self.device) # Get outputs nd.maker.outputs[ShowIpv6VrfAllInterface] = {"{'interface':'','vrf':'all'}": {}} nd.maker.outputs[ShowIpv6Neighbors] = {"{'interface':'','vrf':'all'}": {}} nd.maker.outputs[ShowIpv6NeighborsDetail] = {"": {}} # Learn the feature nd.learn() # Check attribute not found with self.assertRaises(AttributeError): nd.info['interfaces'] if __name__ == '__main__': unittest.main()
39.235294
145
0.671414
794d782c657405216e606077ea9846c5d0786350
23,762
py
Python
src/pymas/core.py
mestradam/pymas
528aa81be9848dea65152a359290238f6ba983a7
[ "MIT" ]
null
null
null
src/pymas/core.py
mestradam/pymas
528aa81be9848dea65152a359290238f6ba983a7
[ "MIT" ]
2
2022-03-25T00:46:25.000Z
2022-03-28T03:23:01.000Z
src/pymas/core.py
mestradam/pymas
528aa81be9848dea65152a359290238f6ba983a7
[ "MIT" ]
1
2022-03-27T01:32:20.000Z
2022-03-27T01:32:20.000Z
import json import numpy as np from pymas.primitives import * class Structure: """Model and analyse a framed structure. Attributes ---------- ux : bool Flag analyze translation along x-axis. uy : bool Flag analyze translation along y-axis. uz : bool Flag analyze translation along z-axis. rx : bool Flag analyze rotation around x-axis. ry : bool Flag analyze rotation around y-axis. rz : bool Flag analyze rotation around z-axis. materials : dict Materials. sections : dict Sections. joints : dict Joints. frames : dict Frames. supports : dict Supports. load_patterns : dict Load patterns. displacements : dict Displacements. end_actions : dict End actions. reactions : dict Reactions. internal_forces: dict Internal forces. internal_displacements: dict Internal displacements. Methods ------- add_material(name, *args, **kwargs) Add a material. add_section(name, *args, **kwargs) Add a section. add_rectangular_section(name, width, height) Add a rectangular section. add_joint(name, *args, **kwargs) Add a joint. add_frame(name, *args, **kwargs) Add a frame. add_support(joint, *args, **kwargs) Add a support. add_load_pattern(name) Add a load pattern. add_load_at_joint(load_pattern, joint, *args, **kwargs) Add a load at joint. add_point_load_at_frame(load_pattern, frame, *args, **kwargs) Add a point load at frame. add_distributed_load(load_pattern, frame, *args, **kwargs) Add a distributed load at frame. get_flags_active_joint_displacements() Get flags active joint displacements. set_flags_active_joint_displacements() Set flags active joint displacements. get_number_active_joint_displacements() Get number active joint displacements. get_indexes() Get joint indexes. set_indexes() Set joint indexes. get_stiffness_matrix() Get the stiffness matrix of the structure. get_stiffness_matrix_modified_by_supports() Get the stiffness matrix of the structure modified by supports. set_stiffness_matrix_modified_by_supports() Set the stiffness matrix of the structure modified by supports. solve_load_pattern(load_pattern) Solve a load pattern. solve() Solve the structure. export(filename) Export the model. """ def __init__(self, ux=None, uy=None, uz=None, rx=None, ry=None, rz=None): """ Instantiate a Structure object. Parameters ---------- ux : bool, optional Flag analyze translation along x-axis. uy : bool, optional Flag analyze translation along y-axis. uz : bool, optional Flag analyze translation along z-axis. rx : bool, optional Flag analyze rotation around x-axis. ry : bool, optional Flag analyze rotation around y-axis. rz : bool, optional Flag analyze rotation around z-axis. """ self._active_joint_displacements = np.array([]) self._indexes = {} self._ksupport = np.array([]) # flags active joint displacements self.ux = ux self.uy = uy self.uz = uz self.rx = rx self.ry = ry self.rz = rz # dict materials and sections self.materials = {} self.sections = {} # dict joints and frames self.joints = {} self.frames = {} # dict supports and load patterns self.supports = {} self.load_patterns = {} # dict displacements, end actions and reactions self.displacements = {} self.end_actions = {} self.reactions = {} # dict internal forces and displacements self.internal_forces = {} self.internal_displacements = {} def add_material(self, name, *args, **kwargs): """ Add a material. Parameters ---------- name : str Material name. Returns ------- material : Material Material. """ material = self.materials[name] = Material(self, name, *args, **kwargs) return material def add_section(self, name, *args, **kwargs): """ Add a section. Parameters ---------- name : str Section name. Returns ------- section : Section Section. """ section = self.sections[name] = Section(self, name, *args, **kwargs) return section def add_rectangular_section(self, name, width, height): """ Add a rectangular section. Parameters ---------- name : str Rectangular section name. width : float Width. height : float Height. Returns ------- rect_sect : RectangularSection RectangularSection. """ rect_sect = self.sections[name] = RectangularSection(self, name, width, height) return rect_sect def add_joint(self, name, *args, **kwargs): """ Add a joint. Parameters ---------- name : str Joint name. Returns ------- joint : Joint Joint. """ joint = self.joints[name] = Joint(self, name, *args, **kwargs) return joint def add_frame(self, name, *args, **kwargs): """ Add a frame. Parameters ---------- name : str Frame name. Returns ------- frame : Frame Frame. """ frame = self.frames[name] = Frame(self, name, *args, **kwargs) return frame def add_support(self, joint, *args, **kwargs): """ Add a support. Parameters ---------- joint : str Joint name. Returns ------- support : Support Support. """ support = self.supports[joint] = Support(self, joint, *args, **kwargs) return support def add_load_pattern(self, name): """ Add a load pattern. Parameters ---------- name : str Load pattern name. Returns ------- loadPattern : LoadPattern. Load pattern. """ loadPattern = self.load_patterns[name] = LoadPattern(self, name) return loadPattern def add_load_at_joint(self, load_pattern, joint, *args, **kwargs): """ Add a point load at joint. Parameters ---------- load_pattern : str Load pattern name. joint : str Joint name. Returns ------- pointLoad : PointLoadAtJoint Point load at joint. """ pointLoad = self.load_patterns[load_pattern].add_point_load_at_joint(joint, *args, **kwargs) return pointLoad def add_point_load_at_frame(self, load_pattern, frame, *args, **kwargs): """ Add a point load at frame. Parameters ---------- load_pattern : str Load pattern name. frame : str Frame name. Returns ------- pointLoad : PointLoadAtFrame Point load at frame. """ pointLoad = self.load_patterns[load_pattern].add_point_load_at_frame(frame, *args, **kwargs) return pointLoad def add_distributed_load(self, load_pattern, frame, *args, **kwargs): """ Add a distributed load at frame. Parameters ---------- load_pattern : str Load pattern name. frame : str Frame name. Returns ------- distributedLoad : DistributedLoad Distributed loads. """ distributedLoad = self.load_patterns[load_pattern].add_distributed_load(frame, *args, **kwargs) return distributedLoad def get_flags_active_joint_displacements(self): """Get flags active joint displacements. Returns ------- array Flags active joint displacements. """ return self._active_joint_displacements def set_flags_active_joint_displacements(self): """Set flags active joint displacements""" ux = self.ux if self.ux is not None else False uy = self.uy if self.uy is not None else False uz = self.uz if self.uz is not None else False rx = self.rx if self.rx is not None else False ry = self.ry if self.ry is not None else False rz = self.rz if self.rz is not None else False self._active_joint_displacements = np.array([ux, uy, uz, rx, ry, rz]) def get_number_active_joint_displacements(self): """ Get number active joint displacements. Returns ------- int Number of active joint displacements. """ return np.count_nonzero(self.get_flags_active_joint_displacements()) def get_indexes(self): """Get joint indexes. Returns ------- dict Key value pairs joints and indexes. """ return self._indexes def set_indexes(self): """Set joint indexes""" n = self.get_number_active_joint_displacements() self._indexes = {joint: np.arange(n * i, n * (i + 1)) for i, joint in enumerate(self.joints)} def get_stiffness_matrix(self): """ Get stiffness matrix of the structure. Returns ------- array Stiffness matrix of the structure. """ flags_active_joint_displacements = self.get_flags_active_joint_displacements() no_active_joint_displacements = self.get_number_active_joint_displacements() indexes = self.get_indexes() no_joints = len(self.joints) no_frames = len(self.frames) n = 2 * no_active_joint_displacements n_2 = n ** 2 rows = np.empty(no_frames * n_2, dtype=int) cols = np.empty(no_frames * n_2, dtype=int) data = np.empty(no_frames * n_2) for i, frame in enumerate(self.frames.values()): k_element = frame.get_global_stiffness_matrix() indexes_element = np.concatenate((indexes[frame.joint_j], indexes[frame.joint_k])) indexes_element = np.broadcast_to(indexes_element, (n, n)) rows[i * n_2:(i + 1) * n_2] = indexes_element.flatten('F') cols[i * n_2:(i + 1) * n_2] = indexes_element.flatten() data[i * n_2:(i + 1) * n_2] = k_element.flatten() return coo_matrix((data, (rows, cols)), 2 * (no_active_joint_displacements * no_joints,)).toarray() def get_stiffness_matrix_modified_by_supports(self): """Get stiffness matrix of the structure modified by supports. Returns ------- stiffness_matrix_with_supports : array Stiffness matrix of the structure modified by supports. """ return self._ksupport def set_stiffness_matrix_modified_by_supports(self): """Set stiffness matrix of the structure modified by supports""" flags_active_joint_displacements = self.get_flags_active_joint_displacements() indexes = self.get_indexes() stiffness_matrix = self.get_stiffness_matrix() n = np.shape(stiffness_matrix)[0] for joint, support in self.supports.items(): joint_indexes = indexes[joint] restraints = support.get_restraints() for index in joint_indexes[restraints]: stiffness_matrix[index] = stiffness_matrix[:, index] = np.zeros(n) stiffness_matrix[index, index] = 1 self._ksupport = stiffness_matrix def solve_load_pattern(self, load_pattern): """ Solve the structure due to a load pattern. Parameters ---------- load_pattern : str Load pattern name. """ flags_joint_displacements = self.get_flags_active_joint_displacements() indexes = self.get_indexes() loadPattern = self.load_patterns[load_pattern] f = loadPattern.get_f().toarray() f_support = np.copy(f) for joint, support in self.supports.items(): joint_indexes = indexes[joint] restrains = support.get_restraints() for index in joint_indexes[restrains]: f_support[index, 0] = 0 # find displacements u = np.linalg.solve(self.get_stiffness_matrix_modified_by_supports(), f_support) # store displacements load_pattern_displacements = {} for joint in self.joints: joint_indexes = indexes[joint] displacements = np.array(6 * [None]) displacements[flags_joint_displacements] = u[joint_indexes, 0] load_pattern_displacements[joint] = Displacements(self, load_pattern, joint, *displacements) self.displacements[load_pattern] = load_pattern_displacements # store frame end actions flags_frame_displacements = np.tile(flags_joint_displacements, 2) n = np.count_nonzero(flags_frame_displacements) rows = [] cols = [] data = [] load_pattern_end_actions = {} for key, frame in self.frames.items(): indexes_element = np.concatenate((indexes[frame.joint_j], indexes[frame.joint_k])) t = frame.get_matrix_rotation() k_element = frame.get_local_stiffness_matrix() u_element = np.zeros((12,1)) u_element[flags_frame_displacements] = u[indexes_element] u_element = np.dot(np.transpose(t), u_element) f_fixed = np.zeros((12, 1)) if key in loadPattern.point_loads_at_frames: for point_load in loadPattern.point_loads_at_frames[key]: f_fixed[flags_frame_displacements] += point_load.get_f_fixed() if key in loadPattern.distributed_loads: for distributed_load in loadPattern.distributed_loads[key]: f_fixed[flags_frame_displacements] += distributed_load.get_f_fixed() f_end_actions = np.dot(k_element, u_element).flatten() f_end_actions += np.dot(np.transpose(t), f_fixed).flatten() load_pattern_end_actions[key] = EndActions(self, load_pattern, key, *f_end_actions) # reactions if frame.joint_j in self.supports or frame.joint_k in self.supports: rows.extend(indexes_element) cols.extend(n * [0]) data.extend(np.dot(t, load_pattern_end_actions[key].get_end_actions()).flatten()[flags_frame_displacements]) self.end_actions[load_pattern] = load_pattern_end_actions # store reactions number_joints = len(self.joints) n = np.count_nonzero(flags_joint_displacements) f += loadPattern.get_f_fixed().toarray() f_end_actions = coo_matrix((data, (rows, cols)), (number_joints * n, 1)).toarray() load_pattern_reactions = {} for joint in self.supports: joint_indexes = indexes[joint] _restrains = self.supports[joint].get_restraints() restrains = np.array(6 * [False]) restrains[flags_joint_displacements] = _restrains reactions = np.array(6 * [None]) reactions[flags_joint_displacements & restrains] = f_end_actions[joint_indexes, 0][_restrains] - f[joint_indexes, 0][_restrains] load_pattern_reactions[joint] = Reaction(self, load_pattern, joint, *reactions) self.reactions[load_pattern] = load_pattern_reactions # store internal forces load_pattern_internal_forces = {} for key, frame in self.frames.items(): load_pattern_internal_forces[key] = InternalForces(self, load_pattern, key, **frame.get_internal_forces(load_pattern)) self.internal_forces[load_pattern] = load_pattern_internal_forces # store internal displacements load_pattern_internal_displacements = {} for key, frame in self.frames.items(): load_pattern_internal_displacements[key] = InternalDisplacements(self, load_pattern, key, **frame.get_internal_displacements(load_pattern)) self.internal_displacements[load_pattern] = load_pattern_internal_displacements def solve(self): """Solve the structure""" self.set_flags_active_joint_displacements() self.set_indexes() self.set_stiffness_matrix_modified_by_supports() for load_pattern in self.load_patterns: self.solve_load_pattern(load_pattern) def export(self, filename): """ Save the structure to a file in json format. Parameters ---------- filename : string Filename """ data = {} # save the materials if self.materials: data['materials'] = {} for key, material in self.materials.items(): data['materials'][key] = {attr: value for attr, value in material.__dict__.items() if not attr.startswith('_') and value is not None} # save the sections if self.sections: data['sections'] = {} for key, section in self.sections.items(): data['sections'][key] = {'type': section.__class__.__name__} data['sections'][key].update({attr: value for attr, value in section.__dict__.items() if not attr.startswith('_') and value is not None}) # save the joints if self.joints: data['joints'] = {} for key, joint in self.joints.items(): data['joints'][key] = {attr: value for attr, value in joint.__dict__.items() if not attr.startswith('_') and value is not None} # save the frames if self.frames: data['frames'] = {} for key, frame in self.frames.items(): data['frames'][key] = {attr: value for attr, value in frame.__dict__.items() if not attr.startswith('_') and value is not None} # save the supports if self.supports: data['supports'] = {} for key, support in self.supports.items(): data['supports'][key] = {attr: value for attr, value in support.__dict__.items() if not attr.startswith('_') and value is not None} # save the load patterns if self.load_patterns: data['load_patterns'] = {} for key, loadPattern in self.load_patterns.items(): data['load_patterns'][key] = {'name': loadPattern.name} # save loads at joints if loadPattern.loads_at_joints: data['load_patterns'][key]['joints'] = {} for _joint, point_loads in loadPattern.loads_at_joints.items(): data['load_patterns'][key]['joints'][_joint] = [] for pointLoad in point_loads: data['load_patterns'][key]['joints'][_joint].append({attr: value for attr, value in pointLoad.__dict__.items() if not attr.startswith('_') and value is not None}) # save loads at frames if loadPattern.point_loads_at_frames or loadPattern.distributed_loads: data['load_patterns'][key]['frames'] = {} for _frame, point_loads in loadPattern.point_loads_at_frames.items(): if not _frame in data['load_patterns'][key]['frames']: data['load_patterns'][key]['frames'][_frame] = [] for pointLoadAtFrame in point_loads: _data = {'type': pointLoadAtFrame.__class__.__name__} _data.update({attr: value for attr, value in pointLoadAtFrame.__dict__.items() if not attr.startswith('_') and value is not None}) data['load_patterns'][key]['frames'][_frame].append(_data) for _frame, distributed_loads in loadPattern.distributed_loads.items(): if not _frame in data['load_patterns'][key]['frames']: data['load_patterns'][key]['frames'][_frame] = [] for distributedLoad in distributed_loads: _data = {'type': distributedLoad.__class__.__name__} _data.update({attr: value for attr, value in distributedLoad.__dict__.items() if not attr.startswith('_') and value is not None}) data['load_patterns'][key]['frames'][_frame].append(_data) # save displacements if self.displacements: data['displacements'] = {} for key, displacements in self.displacements.items(): data['displacements'][key] = {} for joint, displacement in displacements.items(): data['displacements'][key][joint] = {attr: value for attr, value in displacement.__dict__.items() if not attr.startswith('_') and value is not None} # save reactions if self.reactions: data['reactions'] = {} for key, reactions in self.reactions.items(): data['reactions'][key] = {} for joint, reaction in reactions.items(): data['reactions'][key][joint] = {attr: value for attr, value in reaction.__dict__.items() if not attr.startswith('_') and value is not None} # save end actions if self.end_actions: data['end_actions'] = {} for key, end_actions in self.end_actions.items(): data['end_actions'][key] = {} for frame, end_action in end_actions.items(): data['end_actions'][key][frame] = {attr: value for attr, value in end_action.__dict__.items() if not attr.startswith('_') and value is not None} # save internal forces if self.internal_forces: data['internal_forces'] = {} for key, internal_forces in self.internal_forces.items(): data['internal_forces'][key] = {} for frame, internal_force in internal_forces.items(): data['internal_forces'][key][frame] = {attr: value for attr, value in internal_force.__dict__.items() if not attr.startswith('_') and value is not None} # save internal displacements if self.internal_displacements: data['internal_displacements'] = {} for key, internal_displacements in self.internal_displacements.items(): data['internal_displacements'][key] = {} for frame, internal_displacement in internal_displacements.items(): data['internal_displacements'][key][frame] = {attr: value for attr, value in internal_displacement.__dict__.items() if not attr.startswith('_') and value is not None} with open(filename, 'w') as outfile: json.dump(data, outfile, indent=4) return data
34.437681
190
0.579791
794d78a67716b367497ed9071804b5ea8caef34f
69,945
py
Python
pysnmp-with-texts/NORTEL-WLAN-AP-MIB.py
agustinhenze/mibs.snmplabs.com
1fc5c07860542b89212f4c8ab807057d9a9206c7
[ "Apache-2.0" ]
8
2019-05-09T17:04:00.000Z
2021-06-09T06:50:51.000Z
pysnmp-with-texts/NORTEL-WLAN-AP-MIB.py
agustinhenze/mibs.snmplabs.com
1fc5c07860542b89212f4c8ab807057d9a9206c7
[ "Apache-2.0" ]
4
2019-05-31T16:42:59.000Z
2020-01-31T21:57:17.000Z
pysnmp-with-texts/NORTEL-WLAN-AP-MIB.py
agustinhenze/mibs.snmplabs.com
1fc5c07860542b89212f4c8ab807057d9a9206c7
[ "Apache-2.0" ]
10
2019-04-30T05:51:36.000Z
2022-02-16T03:33:41.000Z
# # PySNMP MIB module NORTEL-WLAN-AP-MIB (http://snmplabs.com/pysmi) # ASN.1 source file:///Users/davwang4/Dev/mibs.snmplabs.com/asn1/NORTEL-WLAN-AP-MIB # Produced by pysmi-0.3.4 at Wed May 1 14:24:05 2019 # On host DAVWANG4-M-1475 platform Darwin version 18.5.0 by user davwang4 # Using Python version 3.7.3 (default, Mar 27 2019, 09:23:15) # Integer, ObjectIdentifier, OctetString = mibBuilder.importSymbols("ASN1", "Integer", "ObjectIdentifier", "OctetString") NamedValues, = mibBuilder.importSymbols("ASN1-ENUMERATION", "NamedValues") ValueRangeConstraint, ConstraintsUnion, ConstraintsIntersection, SingleValueConstraint, ValueSizeConstraint = mibBuilder.importSymbols("ASN1-REFINEMENT", "ValueRangeConstraint", "ConstraintsUnion", "ConstraintsIntersection", "SingleValueConstraint", "ValueSizeConstraint") ifPhysAddress, InterfaceIndex, ifIndex = mibBuilder.importSymbols("IF-MIB", "ifPhysAddress", "InterfaceIndex", "ifIndex") ModuleCompliance, NotificationGroup = mibBuilder.importSymbols("SNMPv2-CONF", "ModuleCompliance", "NotificationGroup") MibIdentifier, Counter32, NotificationType, IpAddress, MibScalar, MibTable, MibTableRow, MibTableColumn, Counter64, Unsigned32, Bits, iso, Gauge32, Integer32, TimeTicks, ModuleIdentity, ObjectIdentity = mibBuilder.importSymbols("SNMPv2-SMI", "MibIdentifier", "Counter32", "NotificationType", "IpAddress", "MibScalar", "MibTable", "MibTableRow", "MibTableColumn", "Counter64", "Unsigned32", "Bits", "iso", "Gauge32", "Integer32", "TimeTicks", "ModuleIdentity", "ObjectIdentity") MacAddress, TextualConvention, TruthValue, RowStatus, PhysAddress, DisplayString = mibBuilder.importSymbols("SNMPv2-TC", "MacAddress", "TextualConvention", "TruthValue", "RowStatus", "PhysAddress", "DisplayString") wlan2200, = mibBuilder.importSymbols("SYNOPTICS-ROOT-MIB", "wlan2200") nortelWlanApMib = ModuleIdentity((1, 3, 6, 1, 4, 1, 45, 1, 11, 1)) nortelWlanApMib.setRevisions(('2003-07-16 00:00', '2003-09-11 00:00', '2004-04-12 00:00',)) if getattr(mibBuilder, 'version', (0, 0, 0)) > (4, 4, 0): if mibBuilder.loadTexts: nortelWlanApMib.setRevisionsDescriptions(('v001: Conversion to SMIv2', 'v005: Added ntWlanDot11TrapTable. Added ntWlanApMuAssocTrap. Added ntWlanApMuDisAssocTrap.', 'v005: Updates for AP version 1.3.: - Added ntWlanApDot1xAuthenticationFail, ntWlanApMuWEPAuthFail, and ntWlanApMuWPAAuthFail NOTIFICATION-TYPEs. - Added ntWlanApMuAssocTrapEnabled and ntWlanApMuDisAssocTrapEnabled NOTIFICATION-TYPEs. - Added ntWlanApLID, ntWlanApRateSupport, ntWlanApSecurity, ntWlanApQoS, ntWlanApVlan, ntWlanApStats groups. - Added NtWlanApDataRate, NtWlanApWEPKey TEXTUAL-CONVENTIONS. - Added ntWlanSwNNDataFileVer. - Changed enumerations of ntWlanTransferType to ftpDownload(1), tftpDownload(2), ftpUpload(3), and tftpUpload(4). - Added ntWlanFileTransferStatus. - Added ntWlanDot11dWorldModeEnabled and ntWlanDot11ClosedSystem. - Added ntWlanDot118021xBcastKeyRefresh, ntWlanDot118021xSessKeyRefresh, and ntWlanDot118021xReAuthRefresh. - Added ntWlanDot11SecondaryAuthServer, ntWlanDot11SecondaryAuthPort, ntWlanDot11SecondaryAuthKey, ntWlanDot11SecondaryAuthRetransmit, and ntWlanDot11SecondaryAuthTimeout.',)) if mibBuilder.loadTexts: nortelWlanApMib.setLastUpdated('200404120000Z') if mibBuilder.loadTexts: nortelWlanApMib.setOrganization('Nortel Networks') if mibBuilder.loadTexts: nortelWlanApMib.setContactInfo('Nortel Networks') if mibBuilder.loadTexts: nortelWlanApMib.setDescription("Nortel Wireless LAN AP MIB Copyright 2003-2004 Nortel Networks, Inc. All rights reserved. This Nortel Networks SNMP Management Information Base Specification (Specification) embodies Nortel Networks' confidential and proprietary intellectual property. Nortel Networks retains all title and ownership in the Specification, including any revisions. This Specification is supplied 'AS IS,' and Nortel Networks makes no warranty, either express or implied, as to the use, operation, condition, or performance of the Specification.") ntWlanApSys = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 1)) ntWlanApLineMgnt = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 2)) ntWlanApPortMgnt = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3)) ntWlanApFileTransferMgt = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 4)) ntWlanApResetMgt = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 5)) ntWlanApIpMgt = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 6)) ntWlanApDot11 = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7)) ntWlanApTraps = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 8)) ntWlanApLID = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 9)) ntWlanApRateSupport = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 10)) ntWlanApSecurity = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11)) ntWlanApQoS = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12)) ntWlanApVlan = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 13)) ntWlanApStats = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 14)) class NtWlanApDataRate(TextualConvention, Integer32): description = 'A data rate, measured in increments of 500Kb/s from 1 Mb/s to 63.5 Mb/s. For example, a value of 2 means a data rate of 1 Mb/s, while a value of 5 means a data rate of 2.5 Mb/s.' status = 'current' subtypeSpec = Integer32.subtypeSpec + ValueRangeConstraint(2, 127) class NtWlanApWEPKey(TextualConvention, OctetString): description = 'A WEP key. Valid sizes are 5, 13, and 16 octets, corresponding to WEP key lengths of 64, 128, and 152 bits, respectively. A length of 0 is also allowed, so that retrieval of objects containing a valid key value can return an empty string so as to avoid key disclosure.' status = 'current' subtypeSpec = OctetString.subtypeSpec + ConstraintsUnion(ValueSizeConstraint(0, 0), ValueSizeConstraint(5, 5), ValueSizeConstraint(13, 13), ValueSizeConstraint(16, 16), ) ntWlanSwHardwareVer = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 1, 1), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 20))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanSwHardwareVer.setStatus('current') if mibBuilder.loadTexts: ntWlanSwHardwareVer.setDescription('Hardware version of the main board.') ntWlanSwBootRomVer = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 1, 2), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 20))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanSwBootRomVer.setStatus('current') if mibBuilder.loadTexts: ntWlanSwBootRomVer.setDescription('Boot ROM code version of the main board.') ntWlanSwOpCodeVer = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 1, 3), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 20))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanSwOpCodeVer.setStatus('current') if mibBuilder.loadTexts: ntWlanSwOpCodeVer.setDescription('Operation code version of the main board.') ntWlanSwCountryCode = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 1, 4), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(2, 2)).setFixedLength(2)).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanSwCountryCode.setStatus('current') if mibBuilder.loadTexts: ntWlanSwCountryCode.setDescription('Country code of the AP. AL-ALBANIA, DZ-ALGERIA, AR-ARGENTINA, AM-ARMENIA, AU-AUSTRALIA, AT-AUSTRIA, AZ-AZERBAIJAN, BH-BAHRAIN, BY-BELARUS, BE-BELGIUM, BZ-BELIZE, BO-BOLVIA, BR-BRAZIL, BN-BRUNEI_DARUSSALAM, BG-BULGARIA, CA-CANADA, CL-CHILE, CN-CHINA, CO-COLOMBIA, CR-COSTA_RICA, HR-CROATIA, CY-CYPRUS, CZ-CZECH_REPUBLIC, DK-DENMARK, DO-DOMINICAN_REPUBLIC, EC-ECUADOR, EG-EGYPT, EE-ESTONIA, FI-FINLAND, FR-FRANCE, GE-GEORGIA, DE-GERMANY, GR-GREECE, GT-GUATEMALA, HK-HONG_KONG, HU-HUNGARY, IS-ICELAND, IN-INDIA, ID-INDONESIA, IR-IRAN, IE-IRELAND, IL-ISRAEL, IT-ITALY, JP-JAPAN, JO-JORDAN, KZ-KAZAKHSTAN, KP-NORTH_KOREA, KR-KOREA_REPUBLIC, KW-KUWAIT, LV-LATVIA, LB-LEBANON, LI-LIECHTENSTEIN, LT-LITHUANIA, LU-LUXEMBOURG, MO-MACAU, MK-MACEDONIA, MY-MALAYSIA, MX-MEXICO, MC-MONACO, MA-MOROCCO, NL-NETHERLANDS, NZ-NEW_ZEALAND, NO-NORWAY, OM-OMAN, PK-PAKISTAN, PA-PANAMA, PE-PERU, PH-PHILIPPINES, PL-POLAND, PT-PORTUGAL, PR-PUERTO_RICO, QA-QATAR, RO-ROMANIA, RU-RUSSIA, SA-SAUDI_ARABIA, SG-SINGAPORE, SK-SLOVAK_REPUBLIC, SI-SLOVENIA, ZA-SOUTH_AFRICA, ES-SPAIN, SE-SWEDEN, CH-SWITZERLAND, SY-SYRIA, TW-TAIWAN, TH-THAILAND, TR-TURKEY, UA-UKRAINE, AE-UNITED_ARAB_EMIRATES, GB-UNITED_KINGDOM, US-UNITED_STATES, UY-URUGUAY, VE-VENEZUELA, VN-VIETNAM') ntWlanSwNNDataFileVer = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 1, 5), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 20))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanSwNNDataFileVer.setStatus('current') if mibBuilder.loadTexts: ntWlanSwNNDataFileVer.setDescription('Nortel Networks Customization Data File Version.') ntWlanLineTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 2, 1), ) if mibBuilder.loadTexts: ntWlanLineTable.setStatus('current') if mibBuilder.loadTexts: ntWlanLineTable.setDescription('Table of descriptive and status information about configuration of each RS-232 line in this system') ntWlanLineEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 2, 1, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanLineIndex")) if mibBuilder.loadTexts: ntWlanLineEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanLineEntry.setDescription('An entry in the table, containing information about configuration in one RS232 line of the Access Point.') ntWlanLineIndex = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 2, 1, 1, 1), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 2147483647))) if mibBuilder.loadTexts: ntWlanLineIndex.setStatus('current') if mibBuilder.loadTexts: ntWlanLineIndex.setDescription('This is defined as RS-232 index.') ntWlanLineDataBits = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 2, 1, 1, 2), Integer32()).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanLineDataBits.setStatus('current') if mibBuilder.loadTexts: ntWlanLineDataBits.setDescription('This is defined as number of data bits for the RS232 interface.') ntWlanLineParity = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 2, 1, 1, 3), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(99, 1, 2))).clone(namedValues=NamedValues(("none", 99), ("odd", 1), ("even", 2)))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanLineParity.setStatus('current') if mibBuilder.loadTexts: ntWlanLineParity.setDescription('This is defined as parity of the RS232 interface.') ntWlanLineSpeed = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 2, 1, 1, 4), Integer32()).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanLineSpeed.setStatus('current') if mibBuilder.loadTexts: ntWlanLineSpeed.setDescription('This is defined as the speed of the RS-232 interface.') ntWlanLineStopBits = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 2, 1, 1, 5), Integer32()).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanLineStopBits.setStatus('current') if mibBuilder.loadTexts: ntWlanLineStopBits.setDescription('This is defined as the number of stop bits for the RS-232 interface.') ntWlanPortTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1), ) if mibBuilder.loadTexts: ntWlanPortTable.setStatus('current') if mibBuilder.loadTexts: ntWlanPortTable.setDescription("Table of descriptive and status information about configuration of each switch port (including expansion slot) in this system. This table also contains information about each trunk (similar to Cisco's EtherChannel).") ntWlanPortEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanPortIndex")) if mibBuilder.loadTexts: ntWlanPortEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanPortEntry.setDescription('An entry in the table, containing information about configuration in one switch port of the switch.') ntWlanPortIndex = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1, 1, 1), InterfaceIndex()) if mibBuilder.loadTexts: ntWlanPortIndex.setStatus('current') if mibBuilder.loadTexts: ntWlanPortIndex.setDescription('This is defined as ifIndex in the IF-MIB.') ntWlanPortName = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1, 1, 2), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 64))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanPortName.setStatus('current') if mibBuilder.loadTexts: ntWlanPortName.setDescription('Indicates the port name. This is same as ifAlias in the IF-MIB (RFC2863 or later).') ntWlanPortType = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1, 1, 3), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3, 4, 5, 6, 7, 8))).clone(namedValues=NamedValues(("other", 1), ("hundredBaseTX", 2), ("hundredBaseFX", 3), ("thousandBaseSX", 4), ("thousandBaseLX", 5), ("thousandBaseT", 6), ("thousandBaseGBIC", 7), ("thousandBaseMiniGBIC", 8)))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanPortType.setStatus('current') if mibBuilder.loadTexts: ntWlanPortType.setDescription('Indicates the port type.') ntWlanPortSpeedDpxCfg = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1, 1, 4), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3, 4, 5, 6, 7))).clone(namedValues=NamedValues(("reserved", 1), ("halfDuplex10", 2), ("fullDuplex10", 3), ("halfDuplex100", 4), ("fullDuplex100", 5), ("halfDuplex1000", 6), ("fullDuplex1000", 7))).clone('halfDuplex10')).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanPortSpeedDpxCfg.setStatus('current') if mibBuilder.loadTexts: ntWlanPortSpeedDpxCfg.setDescription('Set the port speed and duplex mode as follows: halfDuplex10(2) - 10Mbps and half duplex mode fullDuplex10(3) - 10Mbps and full duplex mode halfDuplex100(4) - 100Mbps and half duplex mode fullDuplex100(5) - 100Mbps and full duplex mode halfDuplex1000(6) - 1000Mbps and half duplex mode fullDuplex1000(7) - 1000Mbps and full duplex mode hundredBaseTX port can be set as halfDuplex10(2) fullDuplex10(3) halfDuplex100(4) fullDuplex100(5) hundredBaseFX port can be set as halfDuplex100(4) fullDuplex100(5) thousandBaseSX port can be set as halfDuplex1000(6) fullDuplex1000(7) The actual operating speed and duplex of the port is given by portSpeedDpxStatus.') ntWlanPortFlowCtrlCfg = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1, 1, 5), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3, 4))).clone(namedValues=NamedValues(("enabled", 1), ("disabled", 2), ("backPressure", 3), ("dot3xFlowControl", 4))).clone('enabled')).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanPortFlowCtrlCfg.setStatus('current') if mibBuilder.loadTexts: ntWlanPortFlowCtrlCfg.setDescription('(1) Flow control mechanism is enabled. If the port type is hundredBaseTX or thousandBaseSX: When the port is operating in halfDuplex mode, the port uses backPressure flow control mechanism. When the port is operating in fullDuplex mode, the port uses IEEE 802.3x flow control mechanism. If the port type is hundredBaseFX: When the port is operating in halfDuplex mode, the port uses backPressure flow control mechanism. When the port is operating in fullDuplex mode, Flow control mechanism will not function. (2) Flow control mechanism is disabled. (3) Flow control mechanism is backPressure. when the port is in fullDuplex mode.This flow control mechanism will not function. (4) Flow control mechanism is IEEE 802.3x flow control. when the port is in halfDuplex mode.This flow control mechanism will not function. hundredBaseTX and thousandBaseSX port can be set as: enabled(1), disabled(2), backPressure(3), dot3xFlowControl(4). hundredBaseFX port can be set as: enabled(1), disabled(2), backPressure(3). The actual flow control mechanism is used given by portFlowCtrlStatus.') ntWlanPortCapabilities = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1, 1, 6), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(99, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15))).clone(namedValues=NamedValues(("portCap10half", 99), ("portCap10full", 1), ("portCap100half", 2), ("portCap100full", 3), ("portCap1000half", 4), ("portCap1000full", 5), ("reserved6", 6), ("reserved7", 7), ("reserved8", 8), ("reserved9", 9), ("reserved10", 10), ("reserved11", 11), ("reserved12", 12), ("reserved13", 13), ("portCapSym", 14), ("portCapFlowCtrl", 15)))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanPortCapabilities.setStatus('current') if mibBuilder.loadTexts: ntWlanPortCapabilities.setDescription('Port capabilities.') ntWlanPortAutonegotiation = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1, 1, 7), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2))).clone(namedValues=NamedValues(("enabled", 1), ("disabled", 2)))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanPortAutonegotiation.setStatus('current') if mibBuilder.loadTexts: ntWlanPortAutonegotiation.setDescription('Whether autonegotiation is enabled.') ntWlanPortSpeedDpxStatus = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1, 1, 8), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3, 4, 5, 6, 7))).clone(namedValues=NamedValues(("error", 1), ("halfDuplex10", 2), ("fullDuplex10", 3), ("halfDuplex100", 4), ("fullDuplex100", 5), ("halfDuplex1000", 6), ("fullDuplex1000", 7)))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanPortSpeedDpxStatus.setStatus('current') if mibBuilder.loadTexts: ntWlanPortSpeedDpxStatus.setDescription('The operating speed and duplex mode of the switched port. If this index is a trunk, the speed is the speed of its individual members. If this index is a trunk and the result is inconsistent among its member ports, this value is error(1).') ntWlanPortFlowCtrlStatus = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 3, 1, 1, 9), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3, 4))).clone(namedValues=NamedValues(("error", 1), ("backPressure", 2), ("dot3xFlowControl", 3), ("none", 4)))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanPortFlowCtrlStatus.setStatus('current') if mibBuilder.loadTexts: ntWlanPortFlowCtrlStatus.setDescription('(2) BackPressure flow control machanism is used. (3) IEEE 802.3 flow control machanism is used. (4) Flow control mechanism is disabled. If this index is a trunk and the result is inconsistent among its member ports, this value is error(1).') ntWlanTransferStart = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 4, 1), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2))).clone(namedValues=NamedValues(("go", 1), ("nogo", 2)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanTransferStart.setStatus('current') if mibBuilder.loadTexts: ntWlanTransferStart.setDescription('Set to go(1) to start a transfer.') ntWlanTransferType = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 4, 2), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3, 4))).clone(namedValues=NamedValues(("ftpDownload", 1), ("tftpDownload", 2), ("ftpUpload", 3), ("tftpUpload", 4)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanTransferType.setStatus('current') if mibBuilder.loadTexts: ntWlanTransferType.setDescription('Type of file to transfer.') ntWlanFileType = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 4, 3), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3))).clone(namedValues=NamedValues(("firmware", 1), ("config", 2), ("nortelConfig", 3)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanFileType.setStatus('current') if mibBuilder.loadTexts: ntWlanFileType.setDescription('Type of file to transfer.') ntWlanSrcFile = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 4, 4), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 127))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanSrcFile.setStatus('current') if mibBuilder.loadTexts: ntWlanSrcFile.setDescription('The source file name for TFTP transfer when a transfer is next requested via this MIB. This value is set to the zero length string when no file name has been specified.') ntWlanDestFile = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 4, 5), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 127))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDestFile.setStatus('current') if mibBuilder.loadTexts: ntWlanDestFile.setDescription('The destination file name for TFTP transfer when a transfer is next requested via this MIB. This value is set to the zero length string when no file name has been specified.') ntWlanFileServer = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 4, 6), IpAddress()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanFileServer.setStatus('current') if mibBuilder.loadTexts: ntWlanFileServer.setDescription("The IP address of the TFTP server for transfer when a download is next requested via this MIB. This value is set to `0.0.0.0' when no IP address has been specified.") ntWlanUserName = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 4, 7), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 127))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanUserName.setStatus('current') if mibBuilder.loadTexts: ntWlanUserName.setDescription('The username specified for an FTP Transfer.') ntWlanPassword = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 4, 8), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 127))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanPassword.setStatus('current') if mibBuilder.loadTexts: ntWlanPassword.setDescription('The password specified for an FTP Transfer.') ntWlanFileTransferStatus = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 4, 9), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3, 4, 5, 6, 7, 8, 9))).clone(namedValues=NamedValues(("none", 1), ("inProgress", 2), ("success", 3), ("fail", 4), ("invalidSource", 5), ("invalidDestination", 6), ("outOfMemory", 7), ("outOfSpace", 8), ("fileNotFound", 9)))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanFileTransferStatus.setStatus('current') if mibBuilder.loadTexts: ntWlanFileTransferStatus.setDescription('The status of the current FTP Transfer process.') ntWlanRestartOpCodeFile = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 5, 1), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 127))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanRestartOpCodeFile.setStatus('current') if mibBuilder.loadTexts: ntWlanRestartOpCodeFile.setDescription('Name of op-code file for start-up.') ntWlanRestartControl = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 5, 2), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3))).clone(namedValues=NamedValues(("running", 1), ("warmBoot", 2), ("coldBoot", 3)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanRestartControl.setStatus('current') if mibBuilder.loadTexts: ntWlanRestartControl.setDescription('Setting this object to warmBoot(2) causes the device to restart the application software with current configuration parameters saved in non-volatile memory. Setting this object to coldBoot(3) causes the device to reinitialize configuration parameters in non-volatile memory to default values and restart the application software. When the device is running normally, this variable has a value of running(1).') ntWlanNetConfigIPAddress = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 6, 1), IpAddress()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanNetConfigIPAddress.setStatus('current') if mibBuilder.loadTexts: ntWlanNetConfigIPAddress.setDescription('The IP address of this Net interface. The default value for this object is 0.0.0.0. If either the netConfigIPAddress or netConfigSubnetMask are 0.0.0.0, then when the device boots, it may use DHCP to try to figure out what these values should be. If DHCP fails, before the device can talk on the network, this value must be configured (e.g., through a terminal attached to the device).') ntWlanNetConfigSubnetMask = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 6, 2), IpAddress()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanNetConfigSubnetMask.setStatus('current') if mibBuilder.loadTexts: ntWlanNetConfigSubnetMask.setDescription('The subnet mask of this Net interface. The default value for this object is 0.0.0.0. If either the netConfigIPAddress or netConfigSubnetMask are 0.0.0.0, then when the device boots, it may use DHCP to try to figure out what these values should be. If DHCP fails, before the device can talk on the network, this value must be configured (e.g., through a terminal attached to the device).') ntWlanNetDefaultGateway = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 6, 3), IpAddress()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanNetDefaultGateway.setStatus('current') if mibBuilder.loadTexts: ntWlanNetDefaultGateway.setDescription('The IP Address of the default gateway. If this value is undefined or unknown, it shall have the value 0.0.0.0.') ntWlanIpHttpState = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 6, 4), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2))).clone(namedValues=NamedValues(("enabled", 1), ("disabled", 2)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanIpHttpState.setStatus('current') if mibBuilder.loadTexts: ntWlanIpHttpState.setDescription('Whether HTTP is enabled.') ntWlanIpHttpPort = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 6, 5), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 65535))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanIpHttpPort.setStatus('current') if mibBuilder.loadTexts: ntWlanIpHttpPort.setDescription('The port number for HTTP.') ntWlanDot11Phy = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 4)) ntWlanDot11PhyOperationTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 4, 1), ) if mibBuilder.loadTexts: ntWlanDot11PhyOperationTable.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11PhyOperationTable.setDescription('Table of descriptive and status information about configuration of each radio of the AP.') ntWlanDot11PhyOperationEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 4, 1, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanDot11Index")) if mibBuilder.loadTexts: ntWlanDot11PhyOperationEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11PhyOperationEntry.setDescription('An entry in the table, containing information about configuration in one radio of the AP.') ntWlanDot11Index = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 4, 1, 1, 1), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 2147483647))) if mibBuilder.loadTexts: ntWlanDot11Index.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11Index.setDescription('Radio index of the AP.') ntWlanDot11TurboModeEnabled = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 4, 1, 1, 3), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(99, 1, 2))).clone(namedValues=NamedValues(("none", 99), ("on", 1), ("off", 2)))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanDot11TurboModeEnabled.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11TurboModeEnabled.setDescription("This attribute, when true, shall indicate that the nortel's propietory turbo mode option is enabled. The default value of this attribute shall be false.") ntWlanDot11PreambleLength = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 4, 1, 1, 4), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(99, 1, 2))).clone(namedValues=NamedValues(("twelveSymbols", 99), ("short", 1), ("long", 2)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11PreambleLength.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11PreambleLength.setDescription('This attribute determines whether or not a short or a long preamble is used to delineate 802.11 frames.') ntWlanDot11dWorldModeEnabled = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 4, 1, 1, 5), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11dWorldModeEnabled.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11dWorldModeEnabled.setDescription('This attribute determines whether 802.11d world mode is enabled for this radio.') ntWlanDot11ClosedSystem = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 4, 1, 1, 6), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11ClosedSystem.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11ClosedSystem.setDescription('This attribute determines whether the radio operates as a closed system. When configured as a closed system, the SSID is not included when broadcasting periodic beacon frames.') ntWlanDot11AuthenticationEntry = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 5)) ntWlanDot118021xSupport = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 5, 1), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot118021xSupport.setStatus('current') if mibBuilder.loadTexts: ntWlanDot118021xSupport.setDescription('This attribute, when true(1), indicates that the Enterprise Access Point supports the 802.1x authentication algorithm.') ntWlanDot118021xRequired = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 5, 2), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot118021xRequired.setStatus('current') if mibBuilder.loadTexts: ntWlanDot118021xRequired.setDescription('This attribute, when true(1), indicates that the Enterprise Access Point requires successful 802.1x authentication for any clients accessing the network.') ntWlanDot118021xBcastKeyRefresh = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 5, 3), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 1440))).setUnits('Minutes').setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot118021xBcastKeyRefresh.setStatus('current') if mibBuilder.loadTexts: ntWlanDot118021xBcastKeyRefresh.setDescription('The broadcast key refresh rate.') ntWlanDot118021xSessKeyRefresh = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 5, 4), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 1440))).setUnits('Minutes').setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot118021xSessKeyRefresh.setStatus('current') if mibBuilder.loadTexts: ntWlanDot118021xSessKeyRefresh.setDescription('The session key refresh rate.') ntWlanDot118021xReAuthRefresh = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 5, 5), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 65535))).setUnits('Seconds').setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot118021xReAuthRefresh.setStatus('current') if mibBuilder.loadTexts: ntWlanDot118021xReAuthRefresh.setDescription('The re-authentication refresh rate.') ntWlanDot11AuthenticationServerTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6), ) if mibBuilder.loadTexts: ntWlanDot11AuthenticationServerTable.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11AuthenticationServerTable.setDescription('Table of descriptive and status information about configuration of each authentication server.') ntWlanDot11AuthenticationServerEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanDot11ServerIndex")) if mibBuilder.loadTexts: ntWlanDot11AuthenticationServerEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11AuthenticationServerEntry.setDescription('An entry in the table, containing information about configuration in one radio of the AP.') ntWlanDot11ServerIndex = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 1), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 2147483647))) if mibBuilder.loadTexts: ntWlanDot11ServerIndex.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11ServerIndex.setDescription('Radio index of the AP.') ntWlanDot11AuthenticationServer = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 2), IpAddress()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11AuthenticationServer.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11AuthenticationServer.setDescription('This values indicates the IP address of the authentication server.') ntWlanDot11AuthenticationPort = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 3), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1024, 65535))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11AuthenticationPort.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11AuthenticationPort.setDescription('This values indicates the UDP Port used by the primary authentication server.') ntWlanDot11AuthenticationKey = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 4), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 20))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11AuthenticationKey.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11AuthenticationKey.setDescription('This values indicates the shared key used by the authentication server.') ntWlanDot11AuthenticationRetransmit = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 5), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 30))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11AuthenticationRetransmit.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11AuthenticationRetransmit.setDescription('This values indicates the retransmit timer length used by the authentication server.') ntWlanDot11AuthenticationTimeout = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 6), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 60))).setUnits('Seconds').setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11AuthenticationTimeout.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11AuthenticationTimeout.setDescription('This values indicates the Timeout value(sec) used by the authentication server.') ntWlanDot11SecondaryAuthServer = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 7), IpAddress()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11SecondaryAuthServer.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11SecondaryAuthServer.setDescription('This values indicates the IP address of the secondary authentication server.') ntWlanDot11SecondaryAuthPort = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 8), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1024, 65535))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11SecondaryAuthPort.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11SecondaryAuthPort.setDescription('This values indicates the UDP Port used by the primary secondary authentication server.') ntWlanDot11SecondaryAuthKey = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 9), DisplayString().subtype(subtypeSpec=ValueSizeConstraint(0, 20))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11SecondaryAuthKey.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11SecondaryAuthKey.setDescription('This values indicates the shared key used by the secondary authentication server.') ntWlanDot11SecondaryAuthRetransmit = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 10), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 30))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11SecondaryAuthRetransmit.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11SecondaryAuthRetransmit.setDescription('This values indicates the retransmit timer length used by the secondary authentication server.') ntWlanDot11SecondaryAuthTimeout = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 6, 1, 11), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 60))).setUnits('Seconds').setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11SecondaryAuthTimeout.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11SecondaryAuthTimeout.setDescription('This values indicates the Timeout value(sec) used by the secondary authentication server.') ntWlanDot11FilterTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 7), ) if mibBuilder.loadTexts: ntWlanDot11FilterTable.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11FilterTable.setDescription('Table of status information about each configured MAC Address Filtering Entry.') ntWlanDot11FilterEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 7, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanDot11FilterIndex")) if mibBuilder.loadTexts: ntWlanDot11FilterEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11FilterEntry.setDescription('An entry in the table, containing information about configuration in one radio of the AP.') ntWlanDot11FilterIndex = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 7, 1, 1), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 2147483647))) if mibBuilder.loadTexts: ntWlanDot11FilterIndex.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11FilterIndex.setDescription('Filter index.') ntWlanDot11FilterAddress = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 7, 1, 2), PhysAddress()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11FilterAddress.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11FilterAddress.setDescription('This values indicates the MAC address of the filter entry.') ntWlanDot11FilterStatus = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 7, 1, 3), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(30, 31))).clone(namedValues=NamedValues(("allowed", 30), ("denied", 31)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanDot11FilterStatus.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11FilterStatus.setDescription('This values indicates the Status of the filter entry. Ifallowed, the client is allowed access to the network. If disallowed, the no frames will be forwarded to the network from the client.') ntWlanDot11TrapTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 9), ) if mibBuilder.loadTexts: ntWlanDot11TrapTable.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11TrapTable.setDescription('Table of status information for Association and Disassociation Trap') ntWlanDot11TrapEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 9, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanDot11InterfaceIndex")) if mibBuilder.loadTexts: ntWlanDot11TrapEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11TrapEntry.setDescription('An entry in the table, containing information for the association and disassociation trap.') ntWlanDot11InterfaceIndex = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 9, 1, 1), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 2147483647))) if mibBuilder.loadTexts: ntWlanDot11InterfaceIndex.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11InterfaceIndex.setDescription('Interface index.') ntWlanDot11AssociationStationAddress = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 9, 1, 2), PhysAddress()).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanDot11AssociationStationAddress.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11AssociationStationAddress.setDescription('This values indicates the MAC address of the Mobile Unit associating with this Access Point.') ntWlanDot11DisassociationStationAddress = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 9, 1, 3), PhysAddress()).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanDot11DisassociationStationAddress.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11DisassociationStationAddress.setDescription('This values indicates the MAC address of the Mobile Unit no longer associated with this Access Point.') ntWlanDot11AssociationMU = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 7, 9, 1, 4), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 65535))).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanDot11AssociationMU.setStatus('current') if mibBuilder.loadTexts: ntWlanDot11AssociationMU.setDescription('This values indicates the number of Mobile Units associated on the Interface') ntWlanApTraps0 = MibIdentifier((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 8, 0)) ntWlanApDot1xAuthenticationFail = NotificationType((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 8, 0, 101)).setObjects(("IF-MIB", "ifPhysAddress"), ("NORTEL-WLAN-AP-MIB", "ntWlanDot11AssociationStationAddress")) if mibBuilder.loadTexts: ntWlanApDot1xAuthenticationFail.setStatus('current') if mibBuilder.loadTexts: ntWlanApDot1xAuthenticationFail.setDescription('Indicates that a Mobile Unit failed 802.1x authentication. - ifPhysAddress is the MAC address of the Access Point sending this trap message. - ntWlanDot11AssociationStationAddress is the MAC address of the Mobile Unit.') ntWlanApMuAssocTrap = NotificationType((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 8, 0, 111)).setObjects(("IF-MIB", "ifPhysAddress"), ("NORTEL-WLAN-AP-MIB", "ntWlanDot11AssociationStationAddress"), ("NORTEL-WLAN-AP-MIB", "ntWlanDot11AssociationMU")) if mibBuilder.loadTexts: ntWlanApMuAssocTrap.setStatus('current') if mibBuilder.loadTexts: ntWlanApMuAssocTrap.setDescription('Indicates that there is a Mobile Unit associating with this Access Point. - ifPhysAddress is the MAC address of the Access Point sending this trap message. - ntWlanDot11AssociationStationAddress is the MAC address of the Mobile Unit. Note: If two or more Mobile Units associate with the Access Point at the same time, ntWlanDot11AssociationStationAddress belongs to the last Mobile Unit. - ntWlanDot11AssociationMU is the number of Mobile Units currently associating with this Access Point.') ntWlanApMuDisAssocTrap = NotificationType((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 8, 0, 112)).setObjects(("IF-MIB", "ifPhysAddress"), ("NORTEL-WLAN-AP-MIB", "ntWlanDot11DisassociationStationAddress"), ("NORTEL-WLAN-AP-MIB", "ntWlanDot11AssociationMU")) if mibBuilder.loadTexts: ntWlanApMuDisAssocTrap.setStatus('current') if mibBuilder.loadTexts: ntWlanApMuDisAssocTrap.setDescription('Indicates that there is one Mobile Unit no longer associating with this Access Point anymore. - ifPhysAddress is the MAC address of the Access Point sending this trap message. - ntWlanDot11DisassociationStationAddress is the MAC address of the Mobile Unit. Note: If two or more Mobile Units un-associate with the Access Point at the same time, ntWlanDot11DisassociationStationAddress belongs to the last Mobile Unit. - ntWlanDot11AssociationMU is the number of Mobile Units currently associating with this Access Point.') ntWlanApMuWEPAuthFail = NotificationType((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 8, 0, 113)).setObjects(("IF-MIB", "ifIndex"), ("NORTEL-WLAN-AP-MIB", "ntWlanDot11AssociationStationAddress")) if mibBuilder.loadTexts: ntWlanApMuWEPAuthFail.setStatus('current') if mibBuilder.loadTexts: ntWlanApMuWEPAuthFail.setDescription('Indicates that WEP authentication of a Mobile Unit failed. - ifIndex identifies the radio to which the MU attempted to connect. - ntWlanDot11AssociationStationAddress is the MAC address of the Mobile Unit that attempted to connect.') ntWlanApMuWPAAuthFail = NotificationType((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 8, 0, 114)).setObjects(("IF-MIB", "ifIndex"), ("NORTEL-WLAN-AP-MIB", "ntWlanDot11AssociationStationAddress")) if mibBuilder.loadTexts: ntWlanApMuWPAAuthFail.setStatus('current') if mibBuilder.loadTexts: ntWlanApMuWPAAuthFail.setDescription('Indicates that WPA authentication of a Mobile Unit failed. - ifIndex identifies the radio to which the MU attempted to connect. - ntWlanDot11AssociationStationAddress is the MAC address of the Mobile Unit that attempted to connect.') ntWlanApMuAssocTrapEnabled = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 8, 1), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApMuAssocTrapEnabled.setStatus('current') if mibBuilder.loadTexts: ntWlanApMuAssocTrapEnabled.setDescription('This object controls whether ntWlanApMuAssocTrap notifications will be generated when an MU associates with the AP. If this object is true(1), then the notification will be generated. If it is false, the notification will not be generated.') ntWlanApMuDisAssocTrapEnabled = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 8, 2), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApMuDisAssocTrapEnabled.setStatus('current') if mibBuilder.loadTexts: ntWlanApMuDisAssocTrapEnabled.setDescription('This object controls whether ntWlanApMuDisAssocTrap notifications will be generated when an MU disassociates from the AP. If this object is true(1), then the notification will be generated. If it is false, the notification will not be generated.') ntWlanApLIDCheckEtherLinkEnabled = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 9, 1), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApLIDCheckEtherLinkEnabled.setStatus('current') if mibBuilder.loadTexts: ntWlanApLIDCheckEtherLinkEnabled.setDescription('Indicates whether checks for ethernet link integrity are enabled.') ntWlanApLIDCheckIPLinkEnabled = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 9, 2), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApLIDCheckIPLinkEnabled.setStatus('current') if mibBuilder.loadTexts: ntWlanApLIDCheckIPLinkEnabled.setDescription('Indicates whether checks for IP link integrity are enabled.') ntWlanApLIDCheckIpLinkAddress = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 9, 3), IpAddress()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApLIDCheckIpLinkAddress.setStatus('current') if mibBuilder.loadTexts: ntWlanApLIDCheckIpLinkAddress.setDescription('The IP address to ping to check IP link integrity.') ntWlanApRateSupportTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 10, 1), ) if mibBuilder.loadTexts: ntWlanApRateSupportTable.setStatus('current') if mibBuilder.loadTexts: ntWlanApRateSupportTable.setDescription('Table to configure which hardware-supported rates are actively supported, and which are basic rates required for an MU to connect to an AP. Entries must exist in this table for every hardware-supported rate for each appropriate ifIndex value.') ntWlanApRateSupportEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 10, 1, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanApRateSupportIfIndex"), (0, "NORTEL-WLAN-AP-MIB", "ntWlanApRateSupportSpeed")) if mibBuilder.loadTexts: ntWlanApRateSupportEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanApRateSupportEntry.setDescription('An entry in the ntWlanApRateSupportTable.') ntWlanApRateSupportIfIndex = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 10, 1, 1, 1), InterfaceIndex()) if mibBuilder.loadTexts: ntWlanApRateSupportIfIndex.setStatus('current') if mibBuilder.loadTexts: ntWlanApRateSupportIfIndex.setDescription('The ifIndex value of the radio.') ntWlanApRateSupportSpeed = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 10, 1, 1, 2), NtWlanApDataRate()) if mibBuilder.loadTexts: ntWlanApRateSupportSpeed.setStatus('current') if mibBuilder.loadTexts: ntWlanApRateSupportSpeed.setDescription('A supported data rate.') ntWlanApRateSupportLevel = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 10, 1, 1, 3), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(0, 1, 2))).clone(namedValues=NamedValues(("off", 0), ("supported", 1), ("supportedBasic", 2)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApRateSupportLevel.setStatus('current') if mibBuilder.loadTexts: ntWlanApRateSupportLevel.setDescription('This object indicates whether a specific rate is currently supported, and if so, whether it is a basic rate. off - indicates that this rate is not currently supported. supported - indicates that this rate is currently supported, but is not required to be supported by an MU in order to connect to this radio. supportedBasic - indicates that this rate is currently supported as a basic rate, and an MU must also support this rate in order to connect to this radio.') ntWlanApSecurityTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1), ) if mibBuilder.loadTexts: ntWlanApSecurityTable.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityTable.setDescription('Table to configure WEP and WPA security parameters. Entries are per-radio.') ntWlanApSecurityEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanApSecurityIfIndex")) if mibBuilder.loadTexts: ntWlanApSecurityEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityEntry.setDescription('An entry in the ntWlanApSecurityTable.') ntWlanApSecurityIfIndex = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 1), InterfaceIndex()) if mibBuilder.loadTexts: ntWlanApSecurityIfIndex.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityIfIndex.setDescription('The ifIndex value of the radio.') ntWlanApSecurityEnabled = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 2), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityEnabled.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityEnabled.setDescription('Indicates whether security is enabled for a radio.') ntWlanApSecurityWEPAuthType = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 3), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(0, 1))).clone(namedValues=NamedValues(("openSystem", 0), ("sharedKey", 1)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWEPAuthType.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWEPAuthType.setDescription('The WEP authentication type.') ntWlanApSecurityWEPKeyLength = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 4), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(0, 1, 2))).clone(namedValues=NamedValues(("wepKey64", 0), ("wepKey128", 1), ("wepKey152", 2)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWEPKeyLength.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWEPKeyLength.setDescription('Indicates the WEP key length being used.') ntWlanApSecurityWEPActiveKey = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 5), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 4))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWEPActiveKey.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWEPActiveKey.setDescription('Indicates which WEP key is currently in use.') ntWlanApSecurityWEPKey1 = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 6), NtWlanApWEPKey()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWEPKey1.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWEPKey1.setDescription('The first WEP key. Retrieval of this object will always return an empty string.') ntWlanApSecurityWEPKey2 = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 7), NtWlanApWEPKey()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWEPKey2.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWEPKey2.setDescription('The second WEP key. Retrieval of this object will always return an empty string.') ntWlanApSecurityWEPKey3 = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 8), NtWlanApWEPKey()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWEPKey3.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWEPKey3.setDescription('The third WEP key. Retrieval of this object will always return an empty string.') ntWlanApSecurityWEPKey4 = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 9), NtWlanApWEPKey()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWEPKey4.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWEPKey4.setDescription('The fourth WEP key. Retrieval of this object will always return an empty string.') ntWlanApSecurityWPASupport = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 10), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(1, 2, 3))).clone(namedValues=NamedValues(("supported", 1), ("required", 2), ("wepOnly", 3)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWPASupport.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWPASupport.setDescription('Indicates whether WPA security is supported.') ntWlanApSecurityWPAMode = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 11), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(0, 1))).clone(namedValues=NamedValues(("dynamic", 0), ("preSharedKey", 1)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWPAMode.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWPAMode.setDescription('The WPA mode.') ntWlanApSecurityWPAPreSharedKey = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 12), OctetString().subtype(subtypeSpec=ValueSizeConstraint(63, 63)).setFixedLength(63)).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWPAPreSharedKey.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWPAPreSharedKey.setDescription('The WPA pre-shared key.') ntWlanApSecurityWPAMcastCypherMode = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 11, 1, 1, 13), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(0, 1, 2))).clone(namedValues=NamedValues(("wep", 0), ("tkip", 1), ("aes", 2)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApSecurityWPAMcastCypherMode.setStatus('current') if mibBuilder.loadTexts: ntWlanApSecurityWPAMcastCypherMode.setDescription('The multicast cypher mode.') ntWlanApQoSMode = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 1), Integer32().subtype(subtypeSpec=ConstraintsUnion(SingleValueConstraint(0, 1, 2, 3, 4))).clone(namedValues=NamedValues(("off", 0), ("etherSrc", 1), ("etherDst", 2), ("ethertype", 3), ("directPriorityMap", 4)))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApQoSMode.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSMode.setDescription('This object indicates the QoS mode being used. off - indicates QoS features are disabled. etherSrc - indicates that the ethernet source MAC address of received packets is mapped to an egress queue. etherDst - indicates that the ethernet destination MAC address of received packets is mapped to an egress queue. etherType - indicates that the ethernet type of of received packets is mapped to an egress queue. directPriorityMap - indicates that the priority value of received packets is used to map directly to an egress queue.') ntWlanApQoSEtherTypeToQueueTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 2), ) if mibBuilder.loadTexts: ntWlanApQoSEtherTypeToQueueTable.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSEtherTypeToQueueTable.setDescription('Table to configure mappings from ethernet type to queue number.') ntWlanApQoSEtherTypeToQueueEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 2, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanApQoSEtherTypeToQueueIndex")) if mibBuilder.loadTexts: ntWlanApQoSEtherTypeToQueueEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSEtherTypeToQueueEntry.setDescription('An entry in the ntWlanApQoSTable.') ntWlanApQoSEtherTypeToQueueIndex = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 2, 1, 1), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 65535))) if mibBuilder.loadTexts: ntWlanApQoSEtherTypeToQueueIndex.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSEtherTypeToQueueIndex.setDescription('An ethernet type, to be mapped to a queue.') ntWlanApQoSEtherTypeToQueueNumber = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 2, 1, 2), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 7))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApQoSEtherTypeToQueueNumber.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSEtherTypeToQueueNumber.setDescription('The queue number to which the ether type is mapped.') ntWlanApQoSEtherTypeToQueueRowStatus = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 2, 1, 3), RowStatus()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApQoSEtherTypeToQueueRowStatus.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSEtherTypeToQueueRowStatus.setDescription('This object is used to create and destroy entries in this table.') ntWlanApQoSMACToQueueTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 3), ) if mibBuilder.loadTexts: ntWlanApQoSMACToQueueTable.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSMACToQueueTable.setDescription('Table to configure mappings from ethernet MAC address to queue number.') ntWlanApQoSMACToQueueEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 3, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanApQoSMACToQueueAddress")) if mibBuilder.loadTexts: ntWlanApQoSMACToQueueEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSMACToQueueEntry.setDescription('An entry in the ntWlanApQoSTable.') ntWlanApQoSMACToQueueAddress = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 3, 1, 1), MacAddress()) if mibBuilder.loadTexts: ntWlanApQoSMACToQueueAddress.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSMACToQueueAddress.setDescription('An ethernet MAC address, to be mapped to a queue.') ntWlanApQoSMACToQueueNumber = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 3, 1, 2), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 7))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApQoSMACToQueueNumber.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSMACToQueueNumber.setDescription('The queue number to which the MAC address is mapped.') ntWlanApQoSMACToQueueRowStatus = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 12, 3, 1, 3), RowStatus()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApQoSMACToQueueRowStatus.setStatus('current') if mibBuilder.loadTexts: ntWlanApQoSMACToQueueRowStatus.setDescription('This object is used to create and destroy entries in this table.') ntWlanApVlanEnabled = MibScalar((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 13, 1), TruthValue()).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApVlanEnabled.setStatus('current') if mibBuilder.loadTexts: ntWlanApVlanEnabled.setDescription('This object indicates whether Vlan support is enabled.') ntWlanApVlanTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 13, 2), ) if mibBuilder.loadTexts: ntWlanApVlanTable.setStatus('current') if mibBuilder.loadTexts: ntWlanApVlanTable.setDescription('Table to configure the default VID per-radio.') ntWlanApVlanEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 13, 2, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanApVlanIfIndex")) if mibBuilder.loadTexts: ntWlanApVlanEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanApVlanEntry.setDescription('An entry in the ntWlanApVlanTable.') ntWlanApVlanIfIndex = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 13, 2, 1, 1), InterfaceIndex()) if mibBuilder.loadTexts: ntWlanApVlanIfIndex.setStatus('current') if mibBuilder.loadTexts: ntWlanApVlanIfIndex.setDescription('The ifIndex value of the radio.') ntWlanApVlanDefaultVid = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 13, 2, 1, 2), Integer32().subtype(subtypeSpec=ValueRangeConstraint(0, 4094))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApVlanDefaultVid.setStatus('current') if mibBuilder.loadTexts: ntWlanApVlanDefaultVid.setDescription('The default VID of the radio.') ntWlanApVlanMUMACToVidTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 13, 3), ) if mibBuilder.loadTexts: ntWlanApVlanMUMACToVidTable.setStatus('current') if mibBuilder.loadTexts: ntWlanApVlanMUMACToVidTable.setDescription('Table containing mappings from Mobile Unit MAC addresses to VLAN ID values.') ntWlanApVlanMUMACToVidEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 13, 3, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanApVlanMUMACToVidAddress")) if mibBuilder.loadTexts: ntWlanApVlanMUMACToVidEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanApVlanMUMACToVidEntry.setDescription('An entry in the ntWlanApVlanMUMACToVidTable.') ntWlanApVlanMUMACToVidAddress = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 13, 3, 1, 1), MacAddress()) if mibBuilder.loadTexts: ntWlanApVlanMUMACToVidAddress.setStatus('current') if mibBuilder.loadTexts: ntWlanApVlanMUMACToVidAddress.setDescription('The ethernet MAC address of a Mobile Unit.') ntWlanApVlanMUMACToVidNumber = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 13, 3, 1, 2), Integer32().subtype(subtypeSpec=ValueRangeConstraint(1, 4094))).setMaxAccess("readwrite") if mibBuilder.loadTexts: ntWlanApVlanMUMACToVidNumber.setStatus('current') if mibBuilder.loadTexts: ntWlanApVlanMUMACToVidNumber.setDescription("The VID to which this Mobile Unit's MAC is mapped.") ntWlanApMUStatsTable = MibTable((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 14, 1), ) if mibBuilder.loadTexts: ntWlanApMUStatsTable.setStatus('current') if mibBuilder.loadTexts: ntWlanApMUStatsTable.setDescription('Table of per-Mobile Unit statistics.') ntWlanApMUStatsEntry = MibTableRow((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 14, 1, 1), ).setIndexNames((0, "NORTEL-WLAN-AP-MIB", "ntWlanApMUStatsMUAddress")) if mibBuilder.loadTexts: ntWlanApMUStatsEntry.setStatus('current') if mibBuilder.loadTexts: ntWlanApMUStatsEntry.setDescription('An entry in the ntWlanApMUStatsTable.') ntWlanApMUStatsMUAddress = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 14, 1, 1, 1), MacAddress()) if mibBuilder.loadTexts: ntWlanApMUStatsMUAddress.setStatus('current') if mibBuilder.loadTexts: ntWlanApMUStatsMUAddress.setDescription('The ethernet MAC address of the Mobile Unit.') ntWlanApMUStatsPacketsIn = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 14, 1, 1, 2), Counter32()).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanApMUStatsPacketsIn.setStatus('current') if mibBuilder.loadTexts: ntWlanApMUStatsPacketsIn.setDescription('A count of the packets received from this MU.') ntWlanApMUStatsPacketsOut = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 14, 1, 1, 3), Counter32()).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanApMUStatsPacketsOut.setStatus('current') if mibBuilder.loadTexts: ntWlanApMUStatsPacketsOut.setDescription('A count of the packets sent to this MU.') ntWlanApMUStatsOctetsIn = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 14, 1, 1, 4), Counter32()).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanApMUStatsOctetsIn.setStatus('current') if mibBuilder.loadTexts: ntWlanApMUStatsOctetsIn.setDescription('A count of the octets received from this MU.') ntWlanApMUStatsOctetsOut = MibTableColumn((1, 3, 6, 1, 4, 1, 45, 1, 11, 1, 14, 1, 1, 5), Counter32()).setMaxAccess("readonly") if mibBuilder.loadTexts: ntWlanApMUStatsOctetsOut.setStatus('current') if mibBuilder.loadTexts: ntWlanApMUStatsOctetsOut.setDescription('A count of the octets sent to this MU.') mibBuilder.exportSymbols("NORTEL-WLAN-AP-MIB", ntWlanApRateSupportTable=ntWlanApRateSupportTable, ntWlanApRateSupport=ntWlanApRateSupport, ntWlanDot11Index=ntWlanDot11Index, ntWlanDot11AuthenticationEntry=ntWlanDot11AuthenticationEntry, ntWlanDot11ClosedSystem=ntWlanDot11ClosedSystem, ntWlanRestartControl=ntWlanRestartControl, ntWlanPortFlowCtrlStatus=ntWlanPortFlowCtrlStatus, ntWlanTransferType=ntWlanTransferType, ntWlanLineParity=ntWlanLineParity, ntWlanDot118021xRequired=ntWlanDot118021xRequired, ntWlanDot11ServerIndex=ntWlanDot11ServerIndex, ntWlanApMUStatsEntry=ntWlanApMUStatsEntry, ntWlanDot11InterfaceIndex=ntWlanDot11InterfaceIndex, ntWlanApQoSMACToQueueRowStatus=ntWlanApQoSMACToQueueRowStatus, ntWlanApRateSupportLevel=ntWlanApRateSupportLevel, ntWlanLineStopBits=ntWlanLineStopBits, ntWlanDot11FilterEntry=ntWlanDot11FilterEntry, ntWlanDot11SecondaryAuthPort=ntWlanDot11SecondaryAuthPort, ntWlanSwBootRomVer=ntWlanSwBootRomVer, ntWlanDot11TrapEntry=ntWlanDot11TrapEntry, ntWlanApVlanMUMACToVidNumber=ntWlanApVlanMUMACToVidNumber, ntWlanDot11FilterAddress=ntWlanDot11FilterAddress, ntWlanApQoSMACToQueueNumber=ntWlanApQoSMACToQueueNumber, ntWlanApSecurityEnabled=ntWlanApSecurityEnabled, ntWlanApQoSEtherTypeToQueueNumber=ntWlanApQoSEtherTypeToQueueNumber, ntWlanApMuAssocTrapEnabled=ntWlanApMuAssocTrapEnabled, ntWlanPassword=ntWlanPassword, ntWlanDot11AuthenticationServerEntry=ntWlanDot11AuthenticationServerEntry, ntWlanApPortMgnt=ntWlanApPortMgnt, ntWlanApTraps0=ntWlanApTraps0, ntWlanApRateSupportIfIndex=ntWlanApRateSupportIfIndex, NtWlanApWEPKey=NtWlanApWEPKey, ntWlanPortSpeedDpxCfg=ntWlanPortSpeedDpxCfg, ntWlanPortAutonegotiation=ntWlanPortAutonegotiation, ntWlanApQoSMACToQueueEntry=ntWlanApQoSMACToQueueEntry, ntWlanFileServer=ntWlanFileServer, ntWlanApLineMgnt=ntWlanApLineMgnt, ntWlanNetConfigSubnetMask=ntWlanNetConfigSubnetMask, ntWlanApSecurityWPAMode=ntWlanApSecurityWPAMode, ntWlanDot11FilterStatus=ntWlanDot11FilterStatus, ntWlanLineSpeed=ntWlanLineSpeed, ntWlanApSecurityWEPKeyLength=ntWlanApSecurityWEPKeyLength, ntWlanLineDataBits=ntWlanLineDataBits, ntWlanApTraps=ntWlanApTraps, ntWlanApRateSupportSpeed=ntWlanApRateSupportSpeed, ntWlanApMUStatsOctetsOut=ntWlanApMUStatsOctetsOut, ntWlanApSecurityWEPAuthType=ntWlanApSecurityWEPAuthType, ntWlanPortFlowCtrlCfg=ntWlanPortFlowCtrlCfg, ntWlanApMuWEPAuthFail=ntWlanApMuWEPAuthFail, ntWlanDot118021xSessKeyRefresh=ntWlanDot118021xSessKeyRefresh, ntWlanApMUStatsTable=ntWlanApMUStatsTable, ntWlanIpHttpPort=ntWlanIpHttpPort, ntWlanTransferStart=ntWlanTransferStart, ntWlanApDot1xAuthenticationFail=ntWlanApDot1xAuthenticationFail, ntWlanApVlanMUMACToVidTable=ntWlanApVlanMUMACToVidTable, ntWlanApSecurityIfIndex=ntWlanApSecurityIfIndex, ntWlanApQoS=ntWlanApQoS, ntWlanDot11PhyOperationTable=ntWlanDot11PhyOperationTable, ntWlanFileTransferStatus=ntWlanFileTransferStatus, ntWlanDot118021xSupport=ntWlanDot118021xSupport, ntWlanApMuDisAssocTrap=ntWlanApMuDisAssocTrap, ntWlanDestFile=ntWlanDestFile, ntWlanDot11FilterIndex=ntWlanDot11FilterIndex, ntWlanLineTable=ntWlanLineTable, ntWlanApSecurityWEPActiveKey=ntWlanApSecurityWEPActiveKey, ntWlanSwOpCodeVer=ntWlanSwOpCodeVer, ntWlanApVlanEntry=ntWlanApVlanEntry, ntWlanApVlanEnabled=ntWlanApVlanEnabled, ntWlanSwNNDataFileVer=ntWlanSwNNDataFileVer, ntWlanSwHardwareVer=ntWlanSwHardwareVer, ntWlanDot11TrapTable=ntWlanDot11TrapTable, ntWlanNetConfigIPAddress=ntWlanNetConfigIPAddress, ntWlanFileType=ntWlanFileType, ntWlanDot11SecondaryAuthRetransmit=ntWlanDot11SecondaryAuthRetransmit, ntWlanPortCapabilities=ntWlanPortCapabilities, ntWlanDot11AuthenticationServerTable=ntWlanDot11AuthenticationServerTable, ntWlanApQoSEtherTypeToQueueTable=ntWlanApQoSEtherTypeToQueueTable, ntWlanNetDefaultGateway=ntWlanNetDefaultGateway, ntWlanApMuAssocTrap=ntWlanApMuAssocTrap, ntWlanApQoSMACToQueueTable=ntWlanApQoSMACToQueueTable, NtWlanApDataRate=NtWlanApDataRate, ntWlanApQoSEtherTypeToQueueRowStatus=ntWlanApQoSEtherTypeToQueueRowStatus, ntWlanApMUStatsPacketsIn=ntWlanApMUStatsPacketsIn, nortelWlanApMib=nortelWlanApMib, ntWlanApVlanIfIndex=ntWlanApVlanIfIndex, ntWlanPortType=ntWlanPortType, ntWlanDot11AuthenticationServer=ntWlanDot11AuthenticationServer, ntWlanApSecurityWEPKey4=ntWlanApSecurityWEPKey4, ntWlanApSys=ntWlanApSys, ntWlanApDot11=ntWlanApDot11, ntWlanDot11AssociationMU=ntWlanDot11AssociationMU, ntWlanApVlanTable=ntWlanApVlanTable, ntWlanApSecurityWEPKey1=ntWlanApSecurityWEPKey1, ntWlanApQoSMACToQueueAddress=ntWlanApQoSMACToQueueAddress, ntWlanApQoSMode=ntWlanApQoSMode, ntWlanLineEntry=ntWlanLineEntry, ntWlanSrcFile=ntWlanSrcFile, ntWlanApVlan=ntWlanApVlan, ntWlanApMUStatsPacketsOut=ntWlanApMUStatsPacketsOut, ntWlanApLID=ntWlanApLID, ntWlanDot11AuthenticationTimeout=ntWlanDot11AuthenticationTimeout, ntWlanPortName=ntWlanPortName, ntWlanDot11Phy=ntWlanDot11Phy, ntWlanDot11FilterTable=ntWlanDot11FilterTable, ntWlanDot11PhyOperationEntry=ntWlanDot11PhyOperationEntry, ntWlanDot118021xBcastKeyRefresh=ntWlanDot118021xBcastKeyRefresh, ntWlanDot11SecondaryAuthServer=ntWlanDot11SecondaryAuthServer, ntWlanDot11PreambleLength=ntWlanDot11PreambleLength, ntWlanApFileTransferMgt=ntWlanApFileTransferMgt, ntWlanApQoSEtherTypeToQueueEntry=ntWlanApQoSEtherTypeToQueueEntry, ntWlanApLIDCheckIpLinkAddress=ntWlanApLIDCheckIpLinkAddress, ntWlanApRateSupportEntry=ntWlanApRateSupportEntry, ntWlanApSecurityTable=ntWlanApSecurityTable, ntWlanPortSpeedDpxStatus=ntWlanPortSpeedDpxStatus, ntWlanApMuDisAssocTrapEnabled=ntWlanApMuDisAssocTrapEnabled, ntWlanRestartOpCodeFile=ntWlanRestartOpCodeFile, ntWlanPortEntry=ntWlanPortEntry, ntWlanApVlanMUMACToVidAddress=ntWlanApVlanMUMACToVidAddress, ntWlanApMUStatsMUAddress=ntWlanApMUStatsMUAddress, ntWlanDot11AssociationStationAddress=ntWlanDot11AssociationStationAddress, ntWlanApStats=ntWlanApStats, ntWlanDot11AuthenticationPort=ntWlanDot11AuthenticationPort, ntWlanApSecurityWPAMcastCypherMode=ntWlanApSecurityWPAMcastCypherMode, ntWlanApSecurityWPAPreSharedKey=ntWlanApSecurityWPAPreSharedKey, ntWlanDot11SecondaryAuthKey=ntWlanDot11SecondaryAuthKey, ntWlanDot11DisassociationStationAddress=ntWlanDot11DisassociationStationAddress, ntWlanApLIDCheckEtherLinkEnabled=ntWlanApLIDCheckEtherLinkEnabled, ntWlanDot11TurboModeEnabled=ntWlanDot11TurboModeEnabled, ntWlanApQoSEtherTypeToQueueIndex=ntWlanApQoSEtherTypeToQueueIndex, ntWlanDot11dWorldModeEnabled=ntWlanDot11dWorldModeEnabled, ntWlanApIpMgt=ntWlanApIpMgt, ntWlanApVlanDefaultVid=ntWlanApVlanDefaultVid, ntWlanIpHttpState=ntWlanIpHttpState, ntWlanSwCountryCode=ntWlanSwCountryCode, ntWlanDot11AuthenticationKey=ntWlanDot11AuthenticationKey, ntWlanApSecurity=ntWlanApSecurity, ntWlanDot11AuthenticationRetransmit=ntWlanDot11AuthenticationRetransmit, ntWlanApResetMgt=ntWlanApResetMgt, ntWlanUserName=ntWlanUserName, ntWlanApSecurityWEPKey3=ntWlanApSecurityWEPKey3, ntWlanApSecurityEntry=ntWlanApSecurityEntry, ntWlanApLIDCheckIPLinkEnabled=ntWlanApLIDCheckIPLinkEnabled, ntWlanDot11SecondaryAuthTimeout=ntWlanDot11SecondaryAuthTimeout, ntWlanApSecurityWPASupport=ntWlanApSecurityWPASupport, ntWlanApMUStatsOctetsIn=ntWlanApMUStatsOctetsIn, ntWlanPortTable=ntWlanPortTable, ntWlanApVlanMUMACToVidEntry=ntWlanApVlanMUMACToVidEntry, ntWlanApSecurityWEPKey2=ntWlanApSecurityWEPKey2, ntWlanDot118021xReAuthRefresh=ntWlanDot118021xReAuthRefresh, ntWlanLineIndex=ntWlanLineIndex, ntWlanApMuWPAAuthFail=ntWlanApMuWPAAuthFail, ntWlanPortIndex=ntWlanPortIndex, PYSNMP_MODULE_ID=nortelWlanApMib)
156.127232
7,424
0.792079
794d7a951453fc7d9d1504fa39204bf91f5a6443
6,772
py
Python
src/azure-cli/azure/cli/command_modules/cognitiveservices/tests/latest/test_byox.py
xaliciayang/azure-cli
38c80c875e8a79d08d06a2f42ec82fd54934343e
[ "MIT" ]
7
2020-04-26T09:54:05.000Z
2021-07-22T16:54:41.000Z
src/azure-cli/azure/cli/command_modules/cognitiveservices/tests/latest/test_byox.py
xaliciayang/azure-cli
38c80c875e8a79d08d06a2f42ec82fd54934343e
[ "MIT" ]
2
2017-02-11T21:16:40.000Z
2017-02-11T21:30:54.000Z
src/azure-cli/azure/cli/command_modules/cognitiveservices/tests/latest/test_byox.py
xaliciayang/azure-cli
38c80c875e8a79d08d06a2f42ec82fd54934343e
[ "MIT" ]
13
2020-06-30T16:23:36.000Z
2022-03-29T17:12:05.000Z
# -------------------------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for license information. # -------------------------------------------------------------------------------------------- import unittest import time from azure.cli.testsdk import ScenarioTest, ResourceGroupPreparer from azure.cli.testsdk.decorators import serial_test class CognitiveServicesByoxTests(ScenarioTest): @ResourceGroupPreparer() def test_cognitiveservices_user_owned_storage(self, resource_group): sname = self.create_random_name(prefix='cs_cli_test_', length=16) subscription_id = 'f9b96b36-1f5e-4021-8959-51527e26e6d3' if self.is_live else '00000000-0000-0000-0000-000000000000' self.kwargs.update({ 'sname': sname, 'kind': 'SpeechServices', 'sku': 'S0', 'location': 'centraluseuap', 'storageIds': '[{\\\"resourceId\\\":\\\"/subscriptions/' + subscription_id + '/resourceGroups/felixwa-01/providers/Microsoft.Storage/storageAccounts/felixwatest\\\"}]' }) # test to create cognitive services account self.cmd('az cognitiveservices account create -n {sname} -g {rg} --kind {kind} --sku {sku} -l {location} ' '--assign-identity --storage {storageIds} --yes', checks=[self.check('name', '{sname}'), self.check('location', '{location}'), self.check('sku.name', '{sku}'), self.check('properties.provisioningState', 'Succeeded')]) account = self.cmd('az cognitiveservices account show -n {sname} -g {rg}').get_output_in_json() self.assertEqual(account['identity']['type'], 'SystemAssigned') self.assertEqual(len(account['properties']['userOwnedStorage']), 1) self.assertEqual(account['properties']['userOwnedStorage'][0]['resourceId'], '/subscriptions/{}/resourceGroups/felixwa-01/providers/Microsoft.Storage/storageAccounts/felixwatest'.format(subscription_id)) # delete the cognitive services account ret = self.cmd('az cognitiveservices account delete -n {sname} -g {rg}') self.assertEqual(ret.exit_code, 0) # test to create cognitive services account self.cmd('az cognitiveservices account create -n {sname} -g {rg} --kind {kind} --sku {sku} -l {location} ' '--assign-identity --yes', checks=[self.check('name', '{sname}'), self.check('location', '{location}'), self.check('sku.name', '{sku}'), self.check('properties.provisioningState', 'Succeeded')]) self.cmd('az cognitiveservices account update -n {sname} -g {rg} --storage {storageIds}').get_output_in_json() account = self.cmd('az cognitiveservices account show -n {sname} -g {rg}').get_output_in_json() self.assertEqual(account['identity']['type'], 'SystemAssigned') self.assertEqual(len(account['properties']['userOwnedStorage']), 1) self.assertEqual(account['properties']['userOwnedStorage'][0]['resourceId'], '/subscriptions/{}/resourceGroups/felixwa-01/providers/Microsoft.Storage/storageAccounts/felixwatest'.format(subscription_id)) # delete the cognitive services account ret = self.cmd('az cognitiveservices account delete -n {sname} -g {rg}') self.assertEqual(ret.exit_code, 0) @serial_test() @ResourceGroupPreparer() def test_cognitiveservices_encryption(self, resource_group): sname = self.create_random_name(prefix='cs_cli_test_', length=16) self.kwargs.update({ 'sname': sname, 'kind': 'Face', 'sku': 'E0', 'location': 'centraluseuap', 'encryption': '{\\\"keySource\\\":\\\"Microsoft.CognitiveServices\\\"}' }) # test to create cognitive services account self.cmd('az cognitiveservices account create -n {sname} -g {rg} --kind {kind} --sku {sku} -l {location} ' '--assign-identity --encryption {encryption} --yes', checks=[self.check('name', '{sname}'), self.check('location', '{location}'), self.check('sku.name', '{sku}'), self.check('properties.provisioningState', 'Creating')]) for i in range(10): time.sleep(15) account = self.cmd('az cognitiveservices account show -n {sname} -g {rg}').get_output_in_json() if 'Creating' != account['properties']['provisioningState']: break self.assertEqual(account['identity']['type'], 'SystemAssigned') self.assertTrue(account['properties']['encryption'] is not None) self.assertEqual(account['properties']['encryption']['keySource'], 'Microsoft.CognitiveServices') # delete the cognitive services account ret = self.cmd('az cognitiveservices account delete -n {sname} -g {rg}') self.assertEqual(ret.exit_code, 0) self.kwargs.update({ 'sname': self.create_random_name(prefix='cs_cli_test_', length=16) }) # test to create cognitive services account self.cmd('az cognitiveservices account create -n {sname} -g {rg} --kind {kind} --sku {sku} -l {location} ' '--assign-identity --yes', checks=[self.check('name', '{sname}'), self.check('location', '{location}'), self.check('sku.name', '{sku}'), self.check('properties.provisioningState', 'Creating')]) for i in range(10): time.sleep(15) # sleep() is mocked in replay mode, so just use a large value. account = self.cmd('az cognitiveservices account show -n {sname} -g {rg}').get_output_in_json() if 'Creating' != account['properties']['provisioningState']: break self.cmd('az cognitiveservices account update -n {sname} -g {rg} --encryption {encryption}') account = self.cmd('az cognitiveservices account show -n {sname} -g {rg}').get_output_in_json() self.assertEqual(account['identity']['type'], 'SystemAssigned') self.assertTrue(account['properties']['encryption'] is not None) self.assertEqual(account['properties']['encryption']['keySource'], 'Microsoft.CognitiveServices') # delete the cognitive services account ret = self.cmd('az cognitiveservices account delete -n {sname} -g {rg}') self.assertEqual(ret.exit_code, 0) if __name__ == '__main__': unittest.main()
52.496124
211
0.601447
794d7b0267c94177341a972ec5c739d8b6703e9e
4,626
py
Python
gromozeka/concurrency/scheduler.py
moelius/gromozeka
6f7f684064ea0414c0bf346e9d3f8804ab4f399b
[ "MIT" ]
1
2020-06-12T17:48:43.000Z
2020-06-12T17:48:43.000Z
gromozeka/concurrency/scheduler.py
moelius/gromozeka
6f7f684064ea0414c0bf346e9d3f8804ab4f399b
[ "MIT" ]
2
2018-03-14T16:21:48.000Z
2018-06-08T14:41:17.000Z
gromozeka/concurrency/scheduler.py
moelius/gromozeka
6f7f684064ea0414c0bf346e9d3f8804ab4f399b
[ "MIT" ]
null
null
null
"""Module for schedule tasks """ import logging import queue import sched import threading import time from datetime import datetime, timedelta import gromozeka.app class Scheduler(threading.Thread): """Scheduler class Attributes: logger(:obj:`logging.Logger`): Class logger _stop_event(:obj:`threading.Event`): Stop event _shed(:obj:`sched.scheduler`) : Scheduler instance _event_queue(:obj:`queue.Queue`): Queue for scheduler events """ __slots__ = ['logger', '_shed', '_event_queue', '_stop_event'] def __init__(self): self.logger = logging.getLogger('%s.scheduler' % gromozeka.app.get_app().config.app_id) self._shed = sched.scheduler(time.time, time.sleep) self._event_queue = queue.Queue() self._stop_event = threading.Event() super().__init__() def run(self): """Run scheduler in thread """ self.logger.info("start") self.listen_cmd() def add(self, wait_time, func, args=None, kwargs=None, priority=1): """Add event to scheduler queue Args: wait_time(float): Delay to run (seconds) func: Callable function to run args: Func arguments kwargs: Func keyword arguments priority: Event priority """ self._event_queue.put((wait_time, priority, func, args or [], kwargs or {})) def stop(self): """Stop scheduler """ self._stop_event.set() def listen_cmd(self): """Listen to commands """ while not self._stop_event.is_set(): self._shed.run(blocking=False) try: event = self._event_queue.get(True, 0.05) except queue.Empty: continue self._shed.enter(delay=event[0], priority=event[1], action=event[2], argument=event[3], kwargs=event[4]) self.logger.info("stop") def delay_from_eta(eta): """Make new delay from eta Examples: .. code-block:: python delay_from_eta(countdown='2018-02-20 13:38:33.0') # This event will run 20.02.18 at 13:38:33 once Args: eta(:obj:`str`): If `exceptions.Retry` occurs other retry will run at this time. Other arguments ignored. Returns: :obj:`int` or :obj:`None`: Delay """ try: return time.mktime(time.strptime(eta, "%Y-%m-%d %H:%M:%S.%f")) - time.time() except ValueError: return time.mktime(time.strptime(eta, "%Y-%m-%d %H:%M:%S")) - time.time() def new_eta(seconds): """Make new eta from summary of current time and `seconds` Args: seconds(float): Seconds Returns: str: Datetime in format "%Y-%m-%d %H:%M:%S.%f" """ return (datetime.now() + timedelta(seconds=seconds)).strftime("%Y-%m-%d %H:%M:%S.%f") def delay(last_run=None, every=None, interval=None, at=None): """Make new delay from arguments Examples: .. code-block:: python delay(every='days', interval=2) # This event will run every 2 days delay(every='weeks', interval=1, at='15:40') # This event will run every week at 15:40 Args: last_run(:obj:`str`, optional): Last run time every(:obj:`str`, optional): May be seconds, minutes, hours, days, weeks. interval(:obj:`int`, optional): Uses with `every`. at(:obj:`str`, optional): Time formatted as `hour:minute` or `hour:minute:second`. Returns: :obj:`int` or False: new event delay. Returns False when not need more event to run - last run. """ if every: if last_run: last_run_timestamp = time.mktime(time.strptime(last_run, "%Y-%m-%d %H:%M:%S.%f")) interval_seconds = timedelta(**{every: (interval or 1)}).total_seconds() next_timestamp = last_run_timestamp + interval_seconds next_run_date = datetime.fromtimestamp(next_timestamp) else: next_run_date = datetime.now() if at and every in ('days', 'weeks'): time_list = at.split(":") hour, minute, second = map(lambda x: int(x), (time_list[0], time_list[1], time_list[2]) if len(time_list) == 3 else ( time_list[0], time_list[1], 0)) next_run_date = next_run_date.replace(hour=hour, minute=minute, second=second) while next_run_date < datetime.now() or next_run_date.timestamp() < 0: next_run_date += timedelta(days=1) return (next_run_date - datetime.now()).total_seconds()
32.34965
116
0.593385
794d7c0bdb7cf85e9a79653c30295205d2671049
8,814
py
Python
subtle_data_crimes/crime_2_jpeg/Fig7/DL/Test_MoDL_R4_forFig6.py
mikgroup/subtle_data_crimes
210025d9cb8f92583f5f983be15af06b57cfea36
[ "MIT" ]
8
2021-12-12T02:32:30.000Z
2021-12-17T02:26:26.000Z
subtle_data_crimes/crime_2_jpeg/Fig7/DL/Test_MoDL_R4_forFig6.py
mikgroup/data_crimes
210025d9cb8f92583f5f983be15af06b57cfea36
[ "MIT" ]
null
null
null
subtle_data_crimes/crime_2_jpeg/Fig7/DL/Test_MoDL_R4_forFig6.py
mikgroup/data_crimes
210025d9cb8f92583f5f983be15af06b57cfea36
[ "MIT" ]
null
null
null
''' This code is used for testing MoDL on JPEG-compressed data, for the results shown in figures 6, 7 and 8c in the paper. Before running this script you should update the following: basic_data_folder - it should be the same as the output folder defined in the script /crime_2_jpeg/data_prep/jpeg_data_prep.py (c) Efrat Shimron, UC Berkeley, 2021 ''' import logging import os import matplotlib.pyplot as plt import numpy as np import torch from MoDL_single import UnrolledModel from subtle_data_crimes.functions.error_funcs import error_metrics from utils import complex_utils as cplx from utils.datasets import create_data_loaders logging.basicConfig(level=logging.INFO) logger = logging.getLogger(__name__) device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu') # create a folder for the test figures if not os.path.exists('test_figs'): os.makedirs('test_figs') ##################### create test loader ########################### class Namespace: def __init__(self, **kwargs): self.__dict__.update(kwargs) # Hyper parameters params = Namespace() params.batch_size = 1 # image dimensions params.NX = 640 params.NY = 372 # calib is assumed to be 12 for NX=640 calib_x = int(12) calib_y = int(12 * params.NY / params.NX) params.calib = np.array([calib_x, calib_y]) params.shuffle_flag = False # should be True for training, False for testing. Notice that this is not a string, semicolons aren't necessary. params.sampling_flag = 'var_dens_2D' params.var_dens_flag = 'strong' # 'weak' / 'strong' checkpoint_num = int(69) # load saved model (trained network) q_vec = np.array([20, 50, 75, 999]) R_vec = np.array([4]) N_examples_4display = 15 # number of examples to display N_examples_stats = 15 # number of examples over which the mean and STD will be computed NRMSE_av_vs_q_and_R = np.zeros((R_vec.shape[0], q_vec.shape[0])) NRMSE_std_vs_q_and_R = np.zeros((R_vec.shape[0], q_vec.shape[0])) SSIM_av_vs_q_and_R = np.zeros((R_vec.shape[0], q_vec.shape[0])) SSIM_std_vs_q_and_R = np.zeros((R_vec.shape[0], q_vec.shape[0])) N_calc_err = 200 NRMSE_examples_4display = np.zeros((R_vec.shape[0], q_vec.shape[0], N_calc_err)) SSIM_examples_4display = np.zeros((R_vec.shape[0], q_vec.shape[0], N_calc_err)) small_dataset_flag = 0 for r in range(R_vec.shape[0]): R = R_vec[r] print('================================================== ') print(' R={} '.format(R)) print('================================================== ') # Important - here we update R in the params in order to create masks with appropriate sampling # The mask is created in the DataTransform (utils/datasets params.R = R for qi in range(q_vec.shape[0]): q = q_vec[qi] params.q = q # update the next path to YOUR path basic_data_folder = "/mikQNAP/NYU_knee_data/multicoil_efrat/5_JPEG_compressed_data/" data_type = 'test' im_type_str = 'full_im' # training & validation is done on blocks (to accelerate training). Test is done on full-size images. params.data_path = basic_data_folder + data_type + "/q" + str(params.q) + "/" + im_type_str + "/" test_loader = create_data_loaders(params) N_test_batches = len(test_loader.dataset) print('N_test_batches =', N_test_batches) checkpoint_file = 'R{}_q{}/checkpoints/model_{}.pt'.format(R, q, checkpoint_num) checkpoint = torch.load(checkpoint_file, map_location=device) # load the parameters of the trained network params_loaded = checkpoint["params"] single_MoDL = UnrolledModel(params_loaded).to(device) single_MoDL.load_state_dict(checkpoint['model']) single_MoDL.eval() NRMSE_test_list = [] SSIM_test_list = [] cnt = 0 with torch.no_grad(): for iter, data in enumerate(test_loader): if iter % 10 == 0: print('loading test batch ', iter) # input_batch, target_batch, mask_batch, target_no_JPEG_batch = data input_batch, target_batch, mask_batch = data # display the mask (before converting it to torch tensor) if (iter == 0): # print('mask_batch shape:',mask_batch.shape) mask_squeezed = mask_batch[0, :, :, 0].squeeze() # fig = plt.figure() # plt.imshow(mask_squeezed, cmap="gray") # plt.title(params.sampling_flag + ' epoch 0, iter {}'.format(iter)) # plt.show() # fig.savefig('mask_iter{}.png'.format(iter)) # move data to GPU input_batch = input_batch.to(device) target_batch = target_batch.to(device) mask_batch = mask_batch.to(device) # forward pass - for the full batch out_batch = single_MoDL(input_batch.float(), mask=mask_batch) for i in range(params.batch_size): cnt += 1 # counts the number of test images print('cnt={}'.format(cnt)) im_input = cplx.to_numpy(input_batch.cpu())[i, :, :] im_target = cplx.to_numpy(target_batch.cpu())[i, :, :] im_out = cplx.to_numpy(out_batch.cpu())[i, :, :] MoDL_err = error_metrics(np.abs(im_target), np.abs(im_out)) MoDL_err.calc_NRMSE() MoDL_err.calc_SSIM() NRMSE_test_list.append(MoDL_err.NRMSE) SSIM_test_list.append(MoDL_err.SSIM) if cnt < N_calc_err: NRMSE_examples_4display[r, qi, cnt - 1] = MoDL_err.NRMSE SSIM_examples_4display[r, qi, cnt - 1] = MoDL_err.SSIM if cnt <= N_examples_4display: target_im_rotated = np.rot90(np.abs(im_target), 2) im_out_rotated = np.rot90(np.abs(im_out), 2) NX = im_out_rotated.shape[0] NY = im_out_rotated.shape[1] if (r == 0) & (qi == 0) & (iter == 0): TARGETS = np.zeros((NX, NY, q_vec.shape[0], N_examples_4display)) RECS = np.zeros((NX, NY, R_vec.shape[0], q_vec.shape[0], N_examples_4display)) TARGETS[:, :, qi, iter] = target_im_rotated RECS[:, :, r, qi, iter] = im_out_rotated # if iter==0: fig = plt.figure() plt.imshow(target_im_rotated, cmap="gray") plt.colorbar(shrink=0.5) plt.axis('off') plt.title('target - iter={} - R{} q{}'.format(iter, R, q)) plt.show() figname = 'check3_target_R{}_q{}_iter{}'.format(R, q, iter) fig.savefig(figname) if iter >= N_examples_stats: break # NRMSE - calc av & std NRMSE_test_array = np.asarray(NRMSE_test_list) NRMSE_av = np.mean(NRMSE_test_array[0:N_examples_stats].squeeze()) NRMSE_std = np.std(NRMSE_test_array[0:N_examples_stats].squeeze()) NRMSE_av_vs_q_and_R[r, qi] = NRMSE_av NRMSE_std_vs_q_and_R[r, qi] = NRMSE_std # SSIM - calc av & std SSIM_test_array = np.asarray(SSIM_test_list) SSIM_av = np.mean(SSIM_test_array[0:N_examples_stats].squeeze()) SSIM_std = np.std(SSIM_test_array[0:N_examples_stats].squeeze()) SSIM_av_vs_q_and_R[r, qi] = SSIM_av SSIM_std_vs_q_and_R[r, qi] = SSIM_std print('q={} NRMSE_av = {}, SSIM_av = {}'.format(q, NRMSE_av, SSIM_av)) # save NRMSE_av & SSIM print('saving results') results_filename = 'Res_for_Fig6.npz' np.savez(results_filename, R_vec=R_vec, q_vec=q_vec, params=params, checkpoint_num=checkpoint_num, NRMSE_av_vs_q_and_R=NRMSE_av_vs_q_and_R, NRMSE_std_vs_q_and_R=NRMSE_std_vs_q_and_R, SSIM_av_vs_q_and_R=SSIM_av_vs_q_and_R, SSIM_std_vs_q_and_R=SSIM_std_vs_q_and_R, NRMSE_examples_4display=NRMSE_examples_4display, SSIM_examples_4display=SSIM_examples_4display, N_examples_stats=N_examples_stats, N_examples_4display=N_examples_4display, TARGETS=TARGETS, RECS=RECS, )
40.063636
142
0.579192
794d7e6a626920f1c5497016c664224909eab9f2
3,448
py
Python
util/generate_summary.py
cmltaWt0/edx-load-tests
c25547a3ee241b3aa323dc6931e161a356ed5314
[ "Apache-2.0" ]
18
2016-01-31T13:29:56.000Z
2019-02-08T18:08:49.000Z
util/generate_summary.py
cmltaWt0/edx-load-tests
c25547a3ee241b3aa323dc6931e161a356ed5314
[ "Apache-2.0" ]
92
2015-07-31T20:16:51.000Z
2019-08-09T14:32:12.000Z
util/generate_summary.py
cmltaWt0/edx-load-tests
c25547a3ee241b3aa323dc6931e161a356ed5314
[ "Apache-2.0" ]
15
2015-08-19T15:23:58.000Z
2018-02-01T19:47:38.000Z
# -*- coding: utf-8 -*- """ Generate a summary of a previous loadtest run in this environment. See for usage example in a jenkins job dsl: https://github.com/edx/jenkins-job-dsl/blob/master/testeng/jobs/loadtestDriver.groovy Prerequisites: A logfile produced by util/run-loadtest.sh should be present in its standard location. Output: Produces summary on standard output in YAML format. The structure is as follows: * monitoring_links: * list of link text/url pairs pointing to monitoring dashboards. * timeline: * begin: ISO 8601 date for when the test began. * end: ISO 8601 date for when the test ended. """ from datetime import timedelta import yaml import helpers.markers from util.app_monitors_config import MONITORS # Refer to util/run-loadtest.sh in case this file path changes. STANDARD_LOGFILE_PATH = "results/log.txt" def parse_logfile_events(logfile): """ Parse the logfile for events Parameters: logfile (file): the file containing locust logs for a single load test Returns: iterator of (datetime.datetime, str) tuples: the parsed events in the order they are encountered. """ for line in logfile: data = helpers.markers.parse_logfile_event_marker(line) if data is not None: yield (data['time'], data['event']) def get_time_bounds(logfile): """ Determine when the load test started and stopped. Parameters: logfile (file): the file containing locust logs for a single load test Returns: two-tuple of datetime.datetime: the time bounds of the load test """ begin_time = end_time = None relevant_events = ['locust_start_hatching', 'edx_heartbeat', 'quitting'] relevant_times = [ time for time, event in parse_logfile_events(logfile) if event in relevant_events ] begin_time, end_time = (min(relevant_times), max(relevant_times)) return (begin_time, end_time) def main(): """ Generate a summary of a previous load test run. This script assumes "results/log.txt" is the logfile in question. """ with open(STANDARD_LOGFILE_PATH) as logfile: loadtest_begin_time, loadtest_end_time = get_time_bounds(logfile) monitoring_links = [] for monitor in MONITORS: monitoring_links.append({ 'url': monitor.url( begin_time=loadtest_begin_time, end_time=loadtest_end_time, ), 'text': u'{}: {} ({} — {})'.format( monitor.monitoring_service_name, monitor.app_name, # We use naive datetimes (i.e. no attached tz) and just # assume UTC all along. Tacking on the "Z" implies UTC. loadtest_begin_time.strftime('%Y-%m-%dT%H:%M:%SZ'), loadtest_end_time.strftime('%Y-%m-%dT%H:%M:%SZ'), ), }) print(yaml.dump( { 'timeline': { 'begin': loadtest_begin_time.strftime('%Y-%m-%dT%H:%M:%SZ'), 'end': loadtest_end_time.strftime('%Y-%m-%dT%H:%M:%SZ'), }, 'monitoring_links': monitoring_links }, default_flow_style=False, # Represent objects using indented blocks # rather than inline enclosures. allow_unicode=True, )) if __name__ == "__main__": main()
31.063063
85
0.62645
794d7f3e6ee25d1c31cdecf9f2a29faa2864b1de
106,718
py
Python
env/lib/python3.9/site-packages/treebeard/tests/test_treebeard.py
afrokyss/DIT-ILLUMNI-WAGTAIL-APP
c1b22efa6b5b6c0ed5afb30c721b630ec8e37968
[ "MIT" ]
null
null
null
env/lib/python3.9/site-packages/treebeard/tests/test_treebeard.py
afrokyss/DIT-ILLUMNI-WAGTAIL-APP
c1b22efa6b5b6c0ed5afb30c721b630ec8e37968
[ "MIT" ]
null
null
null
env/lib/python3.9/site-packages/treebeard/tests/test_treebeard.py
afrokyss/DIT-ILLUMNI-WAGTAIL-APP
c1b22efa6b5b6c0ed5afb30c721b630ec8e37968
[ "MIT" ]
null
null
null
"""Unit/Functional tests""" import datetime import os from django.contrib.admin.sites import AdminSite from django.contrib.admin.views.main import ChangeList from django.contrib.auth.models import User, AnonymousUser from django.contrib.messages.storage.fallback import FallbackStorage from django.db.models import Q from django.db.models.signals import post_save from django.dispatch import receiver from django.template import Template, Context from django.test import TestCase from django.test.client import RequestFactory from django.templatetags.static import static from django.contrib.admin.options import TO_FIELD_VAR import pytest from treebeard import numconv from treebeard.admin import admin_factory from treebeard.exceptions import ( InvalidPosition, InvalidMoveToDescendant, PathOverflow, MissingNodeOrderBy, NodeAlreadySaved, ) from treebeard.forms import movenodeform_factory from treebeard.tests import models from treebeard.tests.admin import register_all as admin_register_all admin_register_all() BASE_DATA = [ {"data": {"desc": "1"}}, { "data": {"desc": "2"}, "children": [ {"data": {"desc": "21"}}, {"data": {"desc": "22"}}, { "data": {"desc": "23"}, "children": [ {"data": {"desc": "231"}}, ], }, {"data": {"desc": "24"}}, ], }, {"data": {"desc": "3"}}, { "data": {"desc": "4"}, "children": [ {"data": {"desc": "41"}}, ], }, ] UNCHANGED = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] @pytest.fixture(scope="function", params=models.BASE_MODELS + models.PROXY_MODELS) def model(request): request.param.load_bulk(BASE_DATA) return request.param @pytest.fixture(scope="function", params=models.BASE_MODELS + models.PROXY_MODELS) def model_without_data(request): return request.param @pytest.fixture(scope="function", params=models.BASE_MODELS) def model_without_proxy(request): request.param.load_bulk(BASE_DATA) return request.param @pytest.fixture(scope="function", params=models.UNICODE_MODELS) def model_with_unicode(request): return request.param @pytest.fixture(scope="function", params=models.SORTED_MODELS) def sorted_model(request): return request.param @pytest.fixture(scope="function", params=models.RELATED_MODELS) def related_model(request): return request.param @pytest.fixture(scope="function", params=models.INHERITED_MODELS) def inherited_model(request): return request.param @pytest.fixture(scope="function", params=models.MP_SHORTPATH_MODELS) def mpshort_model(request): return request.param @pytest.fixture(scope="function", params=[models.MP_TestNodeShortPath]) def mpshortnotsorted_model(request): return request.param @pytest.fixture(scope="function", params=[models.MP_TestNodeAlphabet]) def mpalphabet_model(request): return request.param @pytest.fixture(scope="function", params=[models.MP_TestNodeSortedAutoNow]) def mpsortedautonow_model(request): return request.param @pytest.fixture(scope="function", params=[models.MP_TestNodeSmallStep]) def mpsmallstep_model(request): return request.param @pytest.fixture(scope="function", params=[models.MP_TestManyToManyWithUser]) def mpm2muser_model(request): return request.param class TestTreeBase(object): def got(self, model): if model in [models.NS_TestNode, models.NS_TestNode_Proxy]: # this slows down nested sets tests quite a bit, but it has the # advantage that we'll check the node edges are correct d = {} for tree_id, lft, rgt in model.objects.values_list("tree_id", "lft", "rgt"): d.setdefault(tree_id, []).extend([lft, rgt]) for tree_id, got_edges in d.items(): assert len(got_edges) == max(got_edges) good_edges = list(range(1, len(got_edges) + 1)) assert sorted(got_edges) == good_edges return [ (o.desc, o.get_depth(), o.get_children_count()) for o in model.get_tree() ] def _assert_get_annotated_list(self, model, expected, parent=None): results = model.get_annotated_list(parent) got = [ (obj[0].desc, obj[1]["open"], obj[1]["close"], obj[1]["level"]) for obj in results ] assert expected == got assert all([type(obj[0]) == model for obj in results]) @pytest.mark.django_db class TestEmptyTree(TestTreeBase): def test_load_bulk_empty(self, model_without_data): ids = model_without_data.load_bulk(BASE_DATA) got_descs = [obj.desc for obj in model_without_data.objects.filter(pk__in=ids)] expected_descs = [x[0] for x in UNCHANGED] assert sorted(got_descs) == sorted(expected_descs) assert self.got(model_without_data) == UNCHANGED def test_dump_bulk_empty(self, model_without_data): assert model_without_data.dump_bulk() == [] def test_add_root_empty(self, model_without_data): model_without_data.add_root(desc="1") expected = [("1", 1, 0)] assert self.got(model_without_data) == expected def test_get_root_nodes_empty(self, model_without_data): got = model_without_data.get_root_nodes() expected = [] assert [node.desc for node in got] == expected def test_get_first_root_node_empty(self, model_without_data): got = model_without_data.get_first_root_node() assert got is None def test_get_last_root_node_empty(self, model_without_data): got = model_without_data.get_last_root_node() assert got is None def test_get_tree(self, model_without_data): got = list(model_without_data.get_tree()) assert got == [] def test_get_annotated_list(self, model_without_data): expected = [] self._assert_get_annotated_list(model_without_data, expected) def test_add_multiple_root_nodes_adds_sibling_leaves(self, model_without_data): model_without_data.add_root(desc="1") model_without_data.add_root(desc="2") model_without_data.add_root(desc="3") model_without_data.add_root(desc="4") # these are all sibling root nodes (depth=1), and leaf nodes (children=0) expected = [("1", 1, 0), ("2", 1, 0), ("3", 1, 0), ("4", 1, 0)] assert self.got(model_without_data) == expected class TestNonEmptyTree(TestTreeBase): pass @pytest.mark.django_db class TestClassMethods(TestNonEmptyTree): def test_load_bulk_existing(self, model): # inserting on an existing node node = model.objects.get(desc="231") ids = model.load_bulk(BASE_DATA, node) expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 4), ("1", 4, 0), ("2", 4, 4), ("21", 5, 0), ("22", 5, 0), ("23", 5, 1), ("231", 6, 0), ("24", 5, 0), ("3", 4, 0), ("4", 4, 1), ("41", 5, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] expected_descs = ["1", "2", "21", "22", "23", "231", "24", "3", "4", "41"] got_descs = [obj.desc for obj in model.objects.filter(pk__in=ids)] assert sorted(got_descs) == sorted(expected_descs) assert self.got(model) == expected def test_get_tree_all(self, model): nodes = model.get_tree() got = [(o.desc, o.get_depth(), o.get_children_count()) for o in nodes] assert got == UNCHANGED assert all([type(o) == model for o in nodes]) def test_dump_bulk_all(self, model): assert model.dump_bulk(keep_ids=False) == BASE_DATA def test_get_tree_node(self, model): node = model.objects.get(desc="231") model.load_bulk(BASE_DATA, node) # the tree was modified by load_bulk, so we reload our node object node = model.objects.get(pk=node.pk) nodes = model.get_tree(node) got = [(o.desc, o.get_depth(), o.get_children_count()) for o in nodes] expected = [ ("231", 3, 4), ("1", 4, 0), ("2", 4, 4), ("21", 5, 0), ("22", 5, 0), ("23", 5, 1), ("231", 6, 0), ("24", 5, 0), ("3", 4, 0), ("4", 4, 1), ("41", 5, 0), ] assert got == expected assert all([type(o) == model for o in nodes]) def test_get_tree_leaf(self, model): node = model.objects.get(desc="1") assert 0 == node.get_children_count() nodes = model.get_tree(node) got = [(o.desc, o.get_depth(), o.get_children_count()) for o in nodes] expected = [("1", 1, 0)] assert got == expected assert all([type(o) == model for o in nodes]) def test_get_annotated_list_all(self, model): expected = [ ("1", True, [], 0), ("2", False, [], 0), ("21", True, [], 1), ("22", False, [], 1), ("23", False, [], 1), ("231", True, [0], 2), ("24", False, [0], 1), ("3", False, [], 0), ("4", False, [], 0), ("41", True, [0, 1], 1), ] self._assert_get_annotated_list(model, expected) def test_get_annotated_list_node(self, model): node = model.objects.get(desc="2") expected = [ ("2", True, [], 0), ("21", True, [], 1), ("22", False, [], 1), ("23", False, [], 1), ("231", True, [0], 2), ("24", False, [0, 1], 1), ] self._assert_get_annotated_list(model, expected, node) def test_get_annotated_list_leaf(self, model): node = model.objects.get(desc="1") expected = [("1", True, [0], 0)] self._assert_get_annotated_list(model, expected, node) def test_dump_bulk_node(self, model): node = model.objects.get(desc="231") model.load_bulk(BASE_DATA, node) # the tree was modified by load_bulk, so we reload our node object node = model.objects.get(pk=node.pk) got = model.dump_bulk(node, False) expected = [{"data": {"desc": "231"}, "children": BASE_DATA}] assert got == expected def test_load_and_dump_bulk_keeping_ids(self, model): exp = model.dump_bulk(keep_ids=True) model.objects.all().delete() model.load_bulk(exp, None, True) got = model.dump_bulk(keep_ids=True) assert got == exp # do we really have an unchaged tree after the dump/delete/load? got = [ (o.desc, o.get_depth(), o.get_children_count()) for o in model.get_tree() ] assert got == UNCHANGED def test_load_and_dump_bulk_with_fk(self, related_model): # https://bitbucket.org/tabo/django-treebeard/issue/48/ related_model.objects.all().delete() related, created = models.RelatedModel.objects.get_or_create( desc="Test %s" % related_model.__name__ ) related_data = [ {"data": {"desc": "1", "related": related.pk}}, { "data": {"desc": "2", "related": related.pk}, "children": [ {"data": {"desc": "21", "related": related.pk}}, {"data": {"desc": "22", "related": related.pk}}, { "data": {"desc": "23", "related": related.pk}, "children": [ {"data": {"desc": "231", "related": related.pk}}, ], }, {"data": {"desc": "24", "related": related.pk}}, ], }, {"data": {"desc": "3", "related": related.pk}}, { "data": {"desc": "4", "related": related.pk}, "children": [ {"data": {"desc": "41", "related": related.pk}}, ], }, ] related_model.load_bulk(related_data) got = related_model.dump_bulk(keep_ids=False) assert got == related_data def test_get_root_nodes(self, model): got = model.get_root_nodes() expected = ["1", "2", "3", "4"] assert [node.desc for node in got] == expected assert all([type(node) == model for node in got]) def test_get_first_root_node(self, model): got = model.get_first_root_node() assert got.desc == "1" assert type(got) == model def test_get_last_root_node(self, model): got = model.get_last_root_node() assert got.desc == "4" assert type(got) == model def test_add_root(self, model): obj = model.add_root(desc="5") assert obj.get_depth() == 1 got = model.get_last_root_node() assert got.desc == "5" assert type(got) == model def test_add_root_with_passed_instance(self, model): obj = model(desc="5") result = model.add_root(instance=obj) assert result == obj got = model.get_last_root_node() assert got.desc == "5" assert type(got) == model def test_add_root_with_already_saved_instance(self, model): obj = model.objects.get(desc="4") with pytest.raises(NodeAlreadySaved): model.add_root(instance=obj) @pytest.mark.django_db class TestSimpleNodeMethods(TestNonEmptyTree): def test_is_root(self, model): data = [ ("2", True), ("1", True), ("4", True), ("21", False), ("24", False), ("22", False), ("231", False), ] for desc, expected in data: got = model.objects.get(desc=desc).is_root() assert got == expected def test_is_leaf(self, model): data = [ ("2", False), ("23", False), ("231", True), ] for desc, expected in data: got = model.objects.get(desc=desc).is_leaf() assert got == expected def test_get_root(self, model): data = [ ("2", "2"), ("1", "1"), ("4", "4"), ("21", "2"), ("24", "2"), ("22", "2"), ("231", "2"), ] for desc, expected in data: node = model.objects.get(desc=desc).get_root() assert node.desc == expected assert type(node) == model def test_get_parent(self, model): data = [ ("2", None), ("1", None), ("4", None), ("21", "2"), ("24", "2"), ("22", "2"), ("231", "23"), ] data = dict(data) objs = {} for desc, expected in data.items(): node = model.objects.get(desc=desc) parent = node.get_parent() if expected: assert parent.desc == expected assert type(parent) == model else: assert parent is None objs[desc] = node # corrupt the objects' parent cache node._parent_obj = "CORRUPTED!!!" for desc, expected in data.items(): node = objs[desc] # asking get_parent to not use the parent cache (since we # corrupted it in the previous loop) parent = node.get_parent(True) if expected: assert parent.desc == expected assert type(parent) == model else: assert parent is None def test_get_children(self, model): data = [ ("2", ["21", "22", "23", "24"]), ("23", ["231"]), ("231", []), ] for desc, expected in data: children = model.objects.get(desc=desc).get_children() assert [node.desc for node in children] == expected assert all([type(node) == model for node in children]) def test_get_children_count(self, model): data = [ ("2", 4), ("23", 1), ("231", 0), ] for desc, expected in data: got = model.objects.get(desc=desc).get_children_count() assert got == expected def test_get_siblings(self, model): data = [ ("2", ["1", "2", "3", "4"]), ("21", ["21", "22", "23", "24"]), ("231", ["231"]), ] for desc, expected in data: siblings = model.objects.get(desc=desc).get_siblings() assert [node.desc for node in siblings] == expected assert all([type(node) == model for node in siblings]) def test_get_first_sibling(self, model): data = [ ("2", "1"), ("1", "1"), ("4", "1"), ("21", "21"), ("24", "21"), ("22", "21"), ("231", "231"), ] for desc, expected in data: node = model.objects.get(desc=desc).get_first_sibling() assert node.desc == expected assert type(node) == model def test_get_prev_sibling(self, model): data = [ ("2", "1"), ("1", None), ("4", "3"), ("21", None), ("24", "23"), ("22", "21"), ("231", None), ] for desc, expected in data: node = model.objects.get(desc=desc).get_prev_sibling() if expected is None: assert node is None else: assert node.desc == expected assert type(node) == model def test_get_next_sibling(self, model): data = [ ("2", "3"), ("1", "2"), ("4", None), ("21", "22"), ("24", None), ("22", "23"), ("231", None), ] for desc, expected in data: node = model.objects.get(desc=desc).get_next_sibling() if expected is None: assert node is None else: assert node.desc == expected assert type(node) == model def test_get_last_sibling(self, model): data = [ ("2", "4"), ("1", "4"), ("4", "4"), ("21", "24"), ("24", "24"), ("22", "24"), ("231", "231"), ] for desc, expected in data: node = model.objects.get(desc=desc).get_last_sibling() assert node.desc == expected assert type(node) == model def test_get_first_child(self, model): data = [ ("2", "21"), ("21", None), ("23", "231"), ("231", None), ] for desc, expected in data: node = model.objects.get(desc=desc).get_first_child() if expected is None: assert node is None else: assert node.desc == expected assert type(node) == model def test_get_last_child(self, model): data = [ ("2", "24"), ("21", None), ("23", "231"), ("231", None), ] for desc, expected in data: node = model.objects.get(desc=desc).get_last_child() if expected is None: assert node is None else: assert node.desc == expected assert type(node) == model def test_get_ancestors(self, model): data = [ ("2", []), ("21", ["2"]), ("231", ["2", "23"]), ] for desc, expected in data: nodes = model.objects.get(desc=desc).get_ancestors() assert [node.desc for node in nodes] == expected assert all([type(node) == model for node in nodes]) def test_get_descendants(self, model): data = [ ("2", ["21", "22", "23", "231", "24"]), ("23", ["231"]), ("231", []), ("1", []), ("4", ["41"]), ] for desc, expected in data: nodes = model.objects.get(desc=desc).get_descendants() assert [node.desc for node in nodes] == expected assert all([type(node) == model for node in nodes]) def test_get_descendant_count(self, model): data = [ ("2", 5), ("23", 1), ("231", 0), ("1", 0), ("4", 1), ] for desc, expected in data: got = model.objects.get(desc=desc).get_descendant_count() assert got == expected def test_is_sibling_of(self, model): data = [ ("2", "2", True), ("2", "1", True), ("21", "2", False), ("231", "2", False), ("22", "23", True), ("231", "23", False), ("231", "231", True), ] for desc1, desc2, expected in data: node1 = model.objects.get(desc=desc1) node2 = model.objects.get(desc=desc2) assert node1.is_sibling_of(node2) == expected def test_is_child_of(self, model): data = [ ("2", "2", False), ("2", "1", False), ("21", "2", True), ("231", "2", False), ("231", "23", True), ("231", "231", False), ] for desc1, desc2, expected in data: node1 = model.objects.get(desc=desc1) node2 = model.objects.get(desc=desc2) assert node1.is_child_of(node2) == expected def test_is_descendant_of(self, model): data = [ ("2", "2", False), ("2", "1", False), ("21", "2", True), ("231", "2", True), ("231", "23", True), ("231", "231", False), ] for desc1, desc2, expected in data: node1 = model.objects.get(desc=desc1) node2 = model.objects.get(desc=desc2) assert node1.is_descendant_of(node2) == expected @pytest.mark.django_db class TestAddChild(TestNonEmptyTree): def test_add_child_to_leaf(self, model): model.objects.get(desc="231").add_child(desc="2311") expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 1), ("2311", 4, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_child_to_node(self, model): model.objects.get(desc="2").add_child(desc="25") expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("25", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_child_with_passed_instance(self, model): child = model(desc="2311") result = model.objects.get(desc="231").add_child(instance=child) assert result == child expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 1), ("2311", 4, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_child_with_already_saved_instance(self, model): child = model.objects.get(desc="21") with pytest.raises(NodeAlreadySaved): model.objects.get(desc="2").add_child(instance=child) def test_add_child_with_pk_set(self, model): """ If the model is using a natural primary key then it will be already set when the instance is inserted. """ child = model(pk=999999, desc="natural key") result = model.objects.get(desc="2").add_child(instance=child) assert result == child def test_add_child_post_save(self, model): try: @receiver(post_save, dispatch_uid="test_add_child_post_save") def on_post_save(instance, **kwargs): parent = instance.get_parent() parent.refresh_from_db() assert parent.get_descendant_count() == 1 # It's important that we're testing a leaf node parent = model.objects.get(desc="231") assert parent.is_leaf() parent.add_child(desc="2311") finally: post_save.disconnect(dispatch_uid="test_add_child_post_save") @pytest.mark.django_db class TestAddSibling(TestNonEmptyTree): def test_add_sibling_invalid_pos(self, model): with pytest.raises(InvalidPosition): model.objects.get(desc="231").add_sibling("invalid_pos") def test_add_sibling_missing_nodeorderby(self, model): node_wchildren = model.objects.get(desc="2") with pytest.raises(MissingNodeOrderBy): node_wchildren.add_sibling("sorted-sibling", desc="aaa") def test_add_sibling_last_root(self, model): node_wchildren = model.objects.get(desc="2") obj = node_wchildren.add_sibling("last-sibling", desc="5") assert obj.get_depth() == 1 assert node_wchildren.get_last_sibling().desc == "5" def test_add_sibling_last(self, model): node = model.objects.get(desc="231") obj = node.add_sibling("last-sibling", desc="232") assert obj.get_depth() == 3 assert node.get_last_sibling().desc == "232" def test_add_sibling_first_root(self, model): node_wchildren = model.objects.get(desc="2") obj = node_wchildren.add_sibling("first-sibling", desc="new") assert obj.get_depth() == 1 expected = [ ("new", 1, 0), ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_sibling_first(self, model): node_wchildren = model.objects.get(desc="23") obj = node_wchildren.add_sibling("first-sibling", desc="new") assert obj.get_depth() == 2 expected = [ ("1", 1, 0), ("2", 1, 5), ("new", 2, 0), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_sibling_left_root(self, model): node_wchildren = model.objects.get(desc="2") obj = node_wchildren.add_sibling("left", desc="new") assert obj.get_depth() == 1 expected = [ ("1", 1, 0), ("new", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_sibling_left(self, model): node_wchildren = model.objects.get(desc="23") obj = node_wchildren.add_sibling("left", desc="new") assert obj.get_depth() == 2 expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("new", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_sibling_left_noleft_root(self, model): node = model.objects.get(desc="1") obj = node.add_sibling("left", desc="new") assert obj.get_depth() == 1 expected = [ ("new", 1, 0), ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_sibling_left_noleft(self, model): node = model.objects.get(desc="231") obj = node.add_sibling("left", desc="new") assert obj.get_depth() == 3 expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 2), ("new", 3, 0), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_sibling_right_root(self, model): node_wchildren = model.objects.get(desc="2") obj = node_wchildren.add_sibling("right", desc="new") assert obj.get_depth() == 1 expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("new", 1, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_sibling_right(self, model): node_wchildren = model.objects.get(desc="23") obj = node_wchildren.add_sibling("right", desc="new") assert obj.get_depth() == 2 expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("new", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_sibling_right_noright_root(self, model): node = model.objects.get(desc="4") obj = node.add_sibling("right", desc="new") assert obj.get_depth() == 1 expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ("new", 1, 0), ] assert self.got(model) == expected def test_add_sibling_right_noright(self, model): node = model.objects.get(desc="231") obj = node.add_sibling("right", desc="new") assert obj.get_depth() == 3 expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 2), ("231", 3, 0), ("new", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_add_sibling_with_passed_instance(self, model): node_wchildren = model.objects.get(desc="2") obj = model(desc="5") result = node_wchildren.add_sibling("last-sibling", instance=obj) assert result == obj assert obj.get_depth() == 1 assert node_wchildren.get_last_sibling().desc == "5" def test_add_sibling_already_saved_instance(self, model): node_wchildren = model.objects.get(desc="2") existing_node = model.objects.get(desc="4") with pytest.raises(NodeAlreadySaved): node_wchildren.add_sibling("last-sibling", instance=existing_node) def test_add_child_with_pk_set(self, model): """ If the model is using a natural primary key then it will be already set when the instance is inserted. """ child = model(pk=999999, desc="natural key") result = model.objects.get(desc="2").add_child(instance=child) assert result == child @pytest.mark.django_db class TestDelete(TestTreeBase): @staticmethod @pytest.fixture( scope="function", params=zip(models.BASE_MODELS, models.DEP_MODELS), ids=lambda fv: f"base={fv[0].__name__} dep={fv[1].__name__}", ) def delete_model(request): base_model, dep_model = request.param base_model.load_bulk(BASE_DATA) for node in base_model.objects.all(): dep_model(node=node).save() return base_model def test_delete_leaf(self, delete_model): delete_model.objects.get(desc="231").delete() expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(delete_model) == expected def test_delete_node(self, delete_model): delete_model.objects.get(desc="23").delete() expected = [ ("1", 1, 0), ("2", 1, 3), ("21", 2, 0), ("22", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(delete_model) == expected def test_delete_root(self, delete_model): delete_model.objects.get(desc="2").delete() expected = [("1", 1, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0)] assert self.got(delete_model) == expected def test_delete_filter_root_nodes(self, delete_model): delete_model.objects.filter(desc__in=("2", "3")).delete() expected = [("1", 1, 0), ("4", 1, 1), ("41", 2, 0)] assert self.got(delete_model) == expected def test_delete_filter_children(self, delete_model): delete_model.objects.filter(desc__in=("2", "23", "231")).delete() expected = [("1", 1, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0)] assert self.got(delete_model) == expected def test_delete_nonexistant_nodes(self, delete_model): delete_model.objects.filter(desc__in=("ZZZ", "XXX")).delete() assert self.got(delete_model) == UNCHANGED def test_delete_same_node_twice(self, delete_model): delete_model.objects.filter(desc__in=("2", "2")).delete() expected = [("1", 1, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0)] assert self.got(delete_model) == expected def test_delete_all_root_nodes(self, delete_model): delete_model.get_root_nodes().delete() count = delete_model.objects.count() assert count == 0 def test_delete_all_nodes(self, delete_model): delete_model.objects.all().delete() count = delete_model.objects.count() assert count == 0 @pytest.mark.django_db class TestMoveErrors(TestNonEmptyTree): def test_move_invalid_pos(self, model): node = model.objects.get(desc="231") with pytest.raises(InvalidPosition): node.move(node, "invalid_pos") def test_move_to_descendant(self, model): node = model.objects.get(desc="2") target = model.objects.get(desc="231") with pytest.raises(InvalidMoveToDescendant): node.move(target, "first-sibling") def test_move_missing_nodeorderby(self, model): node = model.objects.get(desc="231") with pytest.raises(MissingNodeOrderBy): node.move(node, "sorted-child") with pytest.raises(MissingNodeOrderBy): node.move(node, "sorted-sibling") @pytest.mark.django_db class TestMoveSortedErrors(TestTreeBase): def test_nonsorted_move_in_sorted(self, sorted_model): node = sorted_model.add_root(val1=3, val2=3, desc="zxy") with pytest.raises(InvalidPosition): node.move(node, "left") @pytest.mark.django_db class TestMoveLeafRoot(TestNonEmptyTree): def test_move_leaf_last_sibling_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="231").move(target, "last-sibling") expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ("231", 1, 0), ] assert self.got(model) == expected def test_move_leaf_first_sibling_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="231").move(target, "first-sibling") expected = [ ("231", 1, 0), ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_leaf_left_sibling_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="231").move(target, "left") expected = [ ("1", 1, 0), ("231", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_leaf_right_sibling_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="231").move(target, "right") expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("231", 1, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_leaf_last_child_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="231").move(target, "last-child") expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("231", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_leaf_first_child_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="231").move(target, "first-child") expected = [ ("1", 1, 0), ("2", 1, 5), ("231", 2, 0), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected @pytest.mark.django_db class TestMoveLeaf(TestNonEmptyTree): def test_move_leaf_last_sibling(self, model): target = model.objects.get(desc="22") model.objects.get(desc="231").move(target, "last-sibling") expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("231", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_leaf_first_sibling(self, model): target = model.objects.get(desc="22") model.objects.get(desc="231").move(target, "first-sibling") expected = [ ("1", 1, 0), ("2", 1, 5), ("231", 2, 0), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_leaf_left_sibling(self, model): target = model.objects.get(desc="22") model.objects.get(desc="231").move(target, "left") expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("231", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_leaf_right_sibling(self, model): target = model.objects.get(desc="22") model.objects.get(desc="231").move(target, "right") expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("231", 2, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_leaf_left_sibling_itself(self, model): target = model.objects.get(desc="231") model.objects.get(desc="231").move(target, "left") assert self.got(model) == UNCHANGED def test_move_leaf_last_child(self, model): target = model.objects.get(desc="22") model.objects.get(desc="231").move(target, "last-child") expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 1), ("231", 3, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_leaf_first_child(self, model): target = model.objects.get(desc="22") model.objects.get(desc="231").move(target, "first-child") expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 1), ("231", 3, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected @pytest.mark.django_db class TestMoveBranchRoot(TestNonEmptyTree): def test_move_branch_first_sibling_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="4").move(target, "first-sibling") expected = [ ("4", 1, 1), ("41", 2, 0), ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_last_sibling_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="4").move(target, "last-sibling") expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_branch_left_sibling_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="4").move(target, "left") expected = [ ("1", 1, 0), ("4", 1, 1), ("41", 2, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_right_sibling_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="4").move(target, "right") expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("4", 1, 1), ("41", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_left_noleft_sibling_root(self, model): target = model.objects.get(desc="2").get_first_sibling() model.objects.get(desc="4").move(target, "left") expected = [ ("4", 1, 1), ("41", 2, 0), ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_right_noright_sibling_root(self, model): target = model.objects.get(desc="2").get_last_sibling() model.objects.get(desc="4").move(target, "right") expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_branch_first_child_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="4").move(target, "first-child") expected = [ ("1", 1, 0), ("2", 1, 5), ("4", 2, 1), ("41", 3, 0), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_last_child_root(self, model): target = model.objects.get(desc="2") model.objects.get(desc="4").move(target, "last-child") expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("4", 2, 1), ("41", 3, 0), ("3", 1, 0), ] assert self.got(model) == expected @pytest.mark.django_db class TestMoveBranch(TestNonEmptyTree): def test_move_branch_first_sibling(self, model): target = model.objects.get(desc="23") model.objects.get(desc="4").move(target, "first-sibling") expected = [ ("1", 1, 0), ("2", 1, 5), ("4", 2, 1), ("41", 3, 0), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_last_sibling(self, model): target = model.objects.get(desc="23") model.objects.get(desc="4").move(target, "last-sibling") expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("4", 2, 1), ("41", 3, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_left_sibling(self, model): target = model.objects.get(desc="23") model.objects.get(desc="4").move(target, "left") expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("4", 2, 1), ("41", 3, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_right_sibling(self, model): target = model.objects.get(desc="23") model.objects.get(desc="4").move(target, "right") expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("4", 2, 1), ("41", 3, 0), ("24", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_left_noleft_sibling(self, model): target = model.objects.get(desc="23").get_first_sibling() model.objects.get(desc="4").move(target, "left") expected = [ ("1", 1, 0), ("2", 1, 5), ("4", 2, 1), ("41", 3, 0), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_right_noright_sibling(self, model): target = model.objects.get(desc="23").get_last_sibling() model.objects.get(desc="4").move(target, "right") expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("23", 2, 1), ("231", 3, 0), ("24", 2, 0), ("4", 2, 1), ("41", 3, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_left_itself_sibling(self, model): target = model.objects.get(desc="4") model.objects.get(desc="4").move(target, "left") assert self.got(model) == UNCHANGED def test_move_branch_first_child(self, model): target = model.objects.get(desc="23") model.objects.get(desc="4").move(target, "first-child") expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 2), ("4", 3, 1), ("41", 4, 0), ("231", 3, 0), ("24", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected def test_move_branch_last_child(self, model): target = model.objects.get(desc="23") model.objects.get(desc="4").move(target, "last-child") expected = [ ("1", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 2), ("231", 3, 0), ("4", 3, 1), ("41", 4, 0), ("24", 2, 0), ("3", 1, 0), ] assert self.got(model) == expected @pytest.mark.django_db class TestTreeSorted(TestTreeBase): def got(self, sorted_model): return [ (o.val1, o.val2, o.desc, o.get_depth(), o.get_children_count()) for o in sorted_model.get_tree() ] def test_add_root_sorted(self, sorted_model): sorted_model.add_root(val1=3, val2=3, desc="zxy") sorted_model.add_root(val1=1, val2=4, desc="bcd") sorted_model.add_root(val1=2, val2=5, desc="zxy") sorted_model.add_root(val1=3, val2=3, desc="abc") sorted_model.add_root(val1=4, val2=1, desc="fgh") sorted_model.add_root(val1=3, val2=3, desc="abc") sorted_model.add_root(val1=2, val2=2, desc="qwe") sorted_model.add_root(val1=3, val2=2, desc="vcx") expected = [ (1, 4, "bcd", 1, 0), (2, 2, "qwe", 1, 0), (2, 5, "zxy", 1, 0), (3, 2, "vcx", 1, 0), (3, 3, "abc", 1, 0), (3, 3, "abc", 1, 0), (3, 3, "zxy", 1, 0), (4, 1, "fgh", 1, 0), ] assert self.got(sorted_model) == expected def test_add_child_root_sorted(self, sorted_model): root = sorted_model.add_root(val1=0, val2=0, desc="aaa") root.add_child(val1=3, val2=3, desc="zxy") root.add_child(val1=1, val2=4, desc="bcd") root.add_child(val1=2, val2=5, desc="zxy") root.add_child(val1=3, val2=3, desc="abc") root.add_child(val1=4, val2=1, desc="fgh") root.add_child(val1=3, val2=3, desc="abc") root.add_child(val1=2, val2=2, desc="qwe") root.add_child(val1=3, val2=2, desc="vcx") expected = [ (0, 0, "aaa", 1, 8), (1, 4, "bcd", 2, 0), (2, 2, "qwe", 2, 0), (2, 5, "zxy", 2, 0), (3, 2, "vcx", 2, 0), (3, 3, "abc", 2, 0), (3, 3, "abc", 2, 0), (3, 3, "zxy", 2, 0), (4, 1, "fgh", 2, 0), ] assert self.got(sorted_model) == expected def test_add_child_nonroot_sorted(self, sorted_model): get_node = lambda node_id: sorted_model.objects.get(pk=node_id) root_id = sorted_model.add_root(val1=0, val2=0, desc="a").pk node_id = get_node(root_id).add_child(val1=0, val2=0, desc="ac").pk get_node(root_id).add_child(val1=0, val2=0, desc="aa") get_node(root_id).add_child(val1=0, val2=0, desc="av") get_node(node_id).add_child(val1=0, val2=0, desc="aca") get_node(node_id).add_child(val1=0, val2=0, desc="acc") get_node(node_id).add_child(val1=0, val2=0, desc="acb") expected = [ (0, 0, "a", 1, 3), (0, 0, "aa", 2, 0), (0, 0, "ac", 2, 3), (0, 0, "aca", 3, 0), (0, 0, "acb", 3, 0), (0, 0, "acc", 3, 0), (0, 0, "av", 2, 0), ] assert self.got(sorted_model) == expected def test_move_sorted(self, sorted_model): sorted_model.add_root(val1=3, val2=3, desc="zxy") sorted_model.add_root(val1=1, val2=4, desc="bcd") sorted_model.add_root(val1=2, val2=5, desc="zxy") sorted_model.add_root(val1=3, val2=3, desc="abc") sorted_model.add_root(val1=4, val2=1, desc="fgh") sorted_model.add_root(val1=3, val2=3, desc="abc") sorted_model.add_root(val1=2, val2=2, desc="qwe") sorted_model.add_root(val1=3, val2=2, desc="vcx") root_nodes = sorted_model.get_root_nodes() target = root_nodes[0] for node in root_nodes[1:]: # because raw queries don't update django objects node = sorted_model.objects.get(pk=node.pk) target = sorted_model.objects.get(pk=target.pk) node.move(target, "sorted-child") expected = [ (1, 4, "bcd", 1, 7), (2, 2, "qwe", 2, 0), (2, 5, "zxy", 2, 0), (3, 2, "vcx", 2, 0), (3, 3, "abc", 2, 0), (3, 3, "abc", 2, 0), (3, 3, "zxy", 2, 0), (4, 1, "fgh", 2, 0), ] assert self.got(sorted_model) == expected def test_move_sortedsibling(self, sorted_model): # https://bitbucket.org/tabo/django-treebeard/issue/27 sorted_model.add_root(val1=3, val2=3, desc="zxy") sorted_model.add_root(val1=1, val2=4, desc="bcd") sorted_model.add_root(val1=2, val2=5, desc="zxy") sorted_model.add_root(val1=3, val2=3, desc="abc") sorted_model.add_root(val1=4, val2=1, desc="fgh") sorted_model.add_root(val1=3, val2=3, desc="abc") sorted_model.add_root(val1=2, val2=2, desc="qwe") sorted_model.add_root(val1=3, val2=2, desc="vcx") root_nodes = sorted_model.get_root_nodes() target = root_nodes[0] for node in root_nodes[1:]: # because raw queries don't update django objects node = sorted_model.objects.get(pk=node.pk) target = sorted_model.objects.get(pk=target.pk) node.val1 -= 2 node.save() node.move(target, "sorted-sibling") expected = [ (0, 2, "qwe", 1, 0), (0, 5, "zxy", 1, 0), (1, 2, "vcx", 1, 0), (1, 3, "abc", 1, 0), (1, 3, "abc", 1, 0), (1, 3, "zxy", 1, 0), (1, 4, "bcd", 1, 0), (2, 1, "fgh", 1, 0), ] assert self.got(sorted_model) == expected @pytest.mark.django_db class TestInheritedModels(TestTreeBase): @staticmethod @pytest.fixture( scope="function", params=zip(models.BASE_MODELS, models.INHERITED_MODELS), ids=lambda fv: f"base={fv[0].__name__} inherited={fv[1].__name__}", ) def inherited_model(request): base_model, inherited_model = request.param base_model.add_root(desc="1") base_model.add_root(desc="2") node21 = inherited_model(desc="21") base_model.objects.get(desc="2").add_child(instance=node21) base_model.objects.get(desc="21").add_child(desc="211") base_model.objects.get(desc="21").add_child(desc="212") base_model.objects.get(desc="2").add_child(desc="22") node3 = inherited_model(desc="3") base_model.add_root(instance=node3) return inherited_model def test_get_tree_all(self, inherited_model): got = [ (o.desc, o.get_depth(), o.get_children_count()) for o in inherited_model.get_tree() ] expected = [ ("1", 1, 0), ("2", 1, 2), ("21", 2, 2), ("211", 3, 0), ("212", 3, 0), ("22", 2, 0), ("3", 1, 0), ] assert got == expected def test_get_tree_node(self, inherited_model): node = inherited_model.objects.get(desc="21") got = [ (o.desc, o.get_depth(), o.get_children_count()) for o in inherited_model.get_tree(node) ] expected = [ ("21", 2, 2), ("211", 3, 0), ("212", 3, 0), ] assert got == expected def test_get_root_nodes(self, inherited_model): got = inherited_model.get_root_nodes() expected = ["1", "2", "3"] assert [node.desc for node in got] == expected def test_get_first_root_node(self, inherited_model): got = inherited_model.get_first_root_node() assert got.desc == "1" def test_get_last_root_node(self, inherited_model): got = inherited_model.get_last_root_node() assert got.desc == "3" def test_is_root(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.is_root() is False assert node3.is_root() is True def test_is_leaf(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.is_leaf() is False assert node3.is_leaf() is True def test_get_root(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.get_root().desc == "2" assert node3.get_root().desc == "3" def test_get_parent(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.get_parent().desc == "2" assert node3.get_parent() is None def test_get_children(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert [node.desc for node in node21.get_children()] == ["211", "212"] assert [node.desc for node in node3.get_children()] == [] def test_get_children_count(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.get_children_count() == 2 assert node3.get_children_count() == 0 def test_get_siblings(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert [node.desc for node in node21.get_siblings()] == ["21", "22"] assert [node.desc for node in node3.get_siblings()] == ["1", "2", "3"] def test_get_first_sibling(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.get_first_sibling().desc == "21" assert node3.get_first_sibling().desc == "1" def test_get_prev_sibling(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.get_prev_sibling() is None assert node3.get_prev_sibling().desc == "2" def test_get_next_sibling(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.get_next_sibling().desc == "22" assert node3.get_next_sibling() is None def test_get_last_sibling(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.get_last_sibling().desc == "22" assert node3.get_last_sibling().desc == "3" def test_get_first_child(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.get_first_child().desc == "211" assert node3.get_first_child() is None def test_get_last_child(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.get_last_child().desc == "212" assert node3.get_last_child() is None def test_get_ancestors(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert [node.desc for node in node21.get_ancestors()] == ["2"] assert [node.desc for node in node3.get_ancestors()] == [] def test_get_descendants(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert [node.desc for node in node21.get_descendants()] == ["211", "212"] assert [node.desc for node in node3.get_descendants()] == [] def test_get_descendant_count(self, inherited_model): node21 = inherited_model.objects.get(desc="21") node3 = inherited_model.objects.get(desc="3") assert node21.get_descendant_count() == 2 assert node3.get_descendant_count() == 0 def test_cascading_deletion(self, inherited_model): # Deleting a node by calling delete() on the inherited_model class # should delete descendants, even if those descendants are not # instances of inherited_model base_model = inherited_model.__bases__[0] node21 = inherited_model.objects.get(desc="21") node21.delete() node2 = base_model.objects.get(desc="2") for desc in ["21", "211", "212"]: assert not base_model.objects.filter(desc=desc).exists() assert [node.desc for node in node2.get_descendants()] == ["22"] class TestMP_TreeAlphabet(TestTreeBase): @pytest.mark.skipif( not os.getenv("TREEBEARD_TEST_ALPHABET", False), reason="TREEBEARD_TEST_ALPHABET env variable not set.", ) def test_alphabet(self, mpalphabet_model): """This isn't actually a test, it's an informational routine.""" basealpha = numconv.BASE85 got_err = False last_good = None for alphabetlen in range(3, len(basealpha) + 1): alphabet = basealpha[0:alphabetlen] assert len(alphabet) >= 3 expected = [alphabet[0] + char for char in alphabet[1:]] expected.extend([alphabet[1] + char for char in alphabet]) expected.append(alphabet[2] + alphabet[0]) # remove all nodes mpalphabet_model.objects.all().delete() # change the model's alphabet mpalphabet_model.alphabet = alphabet mpalphabet_model.numconv_obj_ = None # insert root nodes for pos in range(len(alphabet) * 2): try: mpalphabet_model.add_root(numval=pos) except: got_err = True break if got_err: break got = [obj.path for obj in mpalphabet_model.objects.all()] if got != expected: break last_good = alphabet assert False, "Best BASE85 based alphabet for your setup: {} (base {})".format( last_good, len(last_good) ) @pytest.mark.django_db class TestHelpers(TestTreeBase): @staticmethod @pytest.fixture(scope="function", params=models.BASE_MODELS + models.PROXY_MODELS) def helpers_model(request): model = request.param model.load_bulk(BASE_DATA) for node in model.get_root_nodes(): model.load_bulk(BASE_DATA, node) model.add_root(desc="5") return model def test_descendants_group_count_root(self, helpers_model): expected = [ (o.desc, o.get_descendant_count()) for o in helpers_model.get_root_nodes() ] got = [ (o.desc, o.descendants_count) for o in helpers_model.get_descendants_group_count() ] assert got == expected def test_descendants_group_count_node(self, helpers_model): parent = helpers_model.get_root_nodes().get(desc="2") expected = [(o.desc, o.get_descendant_count()) for o in parent.get_children()] got = [ (o.desc, o.descendants_count) for o in helpers_model.get_descendants_group_count(parent) ] assert got == expected @pytest.mark.django_db class TestMP_TreeSortedAutoNow(TestTreeBase): """ The sorting mechanism used by treebeard when adding a node can fail if the ordering is using an "auto_now" field """ def test_sorted_by_autonow_workaround(self, mpsortedautonow_model): # workaround for i in range(1, 5): mpsortedautonow_model.add_root( desc="node%d" % (i,), created=datetime.datetime.now() ) def test_sorted_by_autonow_FAIL(self, mpsortedautonow_model): """ This test asserts that we have a problem. fix this, somehow """ mpsortedautonow_model.add_root(desc="node1") with pytest.raises(ValueError): mpsortedautonow_model.add_root(desc="node2") @pytest.mark.django_db class TestMP_TreeStepOverflow(TestTreeBase): def test_add_root(self, mpsmallstep_model): method = mpsmallstep_model.add_root for i in range(1, 10): method() with pytest.raises(PathOverflow): method() def test_add_child(self, mpsmallstep_model): root = mpsmallstep_model.add_root() method = root.add_child for i in range(1, 10): method() with pytest.raises(PathOverflow): method() def test_add_sibling(self, mpsmallstep_model): root = mpsmallstep_model.add_root() for i in range(1, 10): root.add_child() positions = ("first-sibling", "left", "right", "last-sibling") for pos in positions: with pytest.raises(PathOverflow): root.get_last_child().add_sibling(pos) def test_move(self, mpsmallstep_model): root = mpsmallstep_model.add_root() for i in range(1, 10): root.add_child() newroot = mpsmallstep_model.add_root() targets = [ (root, ["first-child", "last-child"]), ( root.get_first_child(), ["first-sibling", "left", "right", "last-sibling"], ), ] for target, positions in targets: for pos in positions: with pytest.raises(PathOverflow): newroot.move(target, pos) @pytest.mark.django_db class TestMP_TreeShortPath(TestTreeBase): """Test a tree with a very small path field (max_length=4) and a steplen of 1 """ def test_short_path(self, mpshortnotsorted_model): obj = mpshortnotsorted_model.add_root() obj = obj.add_child().add_child().add_child() with pytest.raises(PathOverflow): obj.add_child() @pytest.mark.django_db class TestMP_TreeFindProblems(TestTreeBase): def test_find_problems(self, mpalphabet_model): mpalphabet_model.alphabet = "01234" mpalphabet_model(path="01", depth=1, numchild=0, numval=0).save() mpalphabet_model(path="1", depth=1, numchild=0, numval=0).save() mpalphabet_model(path="111", depth=1, numchild=0, numval=0).save() mpalphabet_model(path="abcd", depth=1, numchild=0, numval=0).save() mpalphabet_model(path="qa#$%!", depth=1, numchild=0, numval=0).save() mpalphabet_model(path="0201", depth=2, numchild=0, numval=0).save() mpalphabet_model(path="020201", depth=3, numchild=0, numval=0).save() mpalphabet_model(path="03", depth=1, numchild=2, numval=0).save() mpalphabet_model(path="0301", depth=2, numchild=0, numval=0).save() mpalphabet_model(path="030102", depth=3, numchild=10, numval=0).save() mpalphabet_model(path="04", depth=10, numchild=1, numval=0).save() mpalphabet_model(path="0401", depth=20, numchild=0, numval=0).save() def got(ids): return [o.path for o in mpalphabet_model.objects.filter(pk__in=ids)] ( evil_chars, bad_steplen, orphans, wrong_depth, wrong_numchild, ) = mpalphabet_model.find_problems() assert ["abcd", "qa#$%!"] == got(evil_chars) assert ["1", "111"] == got(bad_steplen) assert ["0201", "020201"] == got(orphans) assert ["03", "0301", "030102"] == got(wrong_numchild) assert ["04", "0401"] == got(wrong_depth) @pytest.mark.django_db class TestMP_TreeFix(TestTreeBase): expected_no_holes = { models.MP_TestNodeShortPath: [ ("1", "b", 1, 2), ("11", "u", 2, 1), ("111", "i", 3, 1), ("1111", "e", 4, 0), ("12", "o", 2, 0), ("2", "d", 1, 0), ("3", "g", 1, 0), ("4", "a", 1, 4), ("41", "a", 2, 0), ("42", "a", 2, 0), ("43", "u", 2, 1), ("431", "i", 3, 1), ("4311", "e", 4, 0), ("44", "o", 2, 0), ], models.MP_TestSortedNodeShortPath: [ ("1", "a", 1, 4), ("11", "a", 2, 0), ("12", "a", 2, 0), ("13", "o", 2, 0), ("14", "u", 2, 1), ("141", "i", 3, 1), ("1411", "e", 4, 0), ("2", "b", 1, 2), ("21", "o", 2, 0), ("22", "u", 2, 1), ("221", "i", 3, 1), ("2211", "e", 4, 0), ("3", "d", 1, 0), ("4", "g", 1, 0), ], } expected_with_holes = { models.MP_TestNodeShortPath: [ ("1", "b", 1, 2), ("13", "u", 2, 1), ("134", "i", 3, 1), ("1343", "e", 4, 0), ("14", "o", 2, 0), ("2", "d", 1, 0), ("3", "g", 1, 0), ("4", "a", 1, 4), ("41", "a", 2, 0), ("42", "a", 2, 0), ("43", "u", 2, 1), ("434", "i", 3, 1), ("4343", "e", 4, 0), ("44", "o", 2, 0), ], models.MP_TestSortedNodeShortPath: [ ("1", "b", 1, 2), ("13", "u", 2, 1), ("134", "i", 3, 1), ("1343", "e", 4, 0), ("14", "o", 2, 0), ("2", "d", 1, 0), ("3", "g", 1, 0), ("4", "a", 1, 4), ("41", "a", 2, 0), ("42", "a", 2, 0), ("43", "u", 2, 1), ("434", "i", 3, 1), ("4343", "e", 4, 0), ("44", "o", 2, 0), ], } def got(self, model): return [ (o.path, o.desc, o.get_depth(), o.get_children_count()) for o in model.get_tree() ] def add_broken_test_data(self, model): model(path="4", depth=2, numchild=2, desc="a").save() model(path="13", depth=1000, numchild=0, desc="u").save() model(path="14", depth=4, numchild=500, desc="o").save() model(path="134", depth=321, numchild=543, desc="i").save() model(path="1343", depth=321, numchild=543, desc="e").save() model(path="42", depth=1, numchild=1, desc="a").save() model(path="43", depth=1000, numchild=0, desc="u").save() model(path="44", depth=4, numchild=500, desc="o").save() model(path="434", depth=321, numchild=543, desc="i").save() model(path="4343", depth=321, numchild=543, desc="e").save() model(path="41", depth=1, numchild=1, desc="a").save() model(path="3", depth=221, numchild=322, desc="g").save() model(path="1", depth=10, numchild=3, desc="b").save() model(path="2", depth=10, numchild=3, desc="d").save() def test_fix_tree_non_destructive(self, mpshort_model): self.add_broken_test_data(mpshort_model) mpshort_model.fix_tree(destructive=False) got = self.got(mpshort_model) expected = self.expected_with_holes[mpshort_model] assert got == expected mpshort_model.find_problems() def test_fix_tree_destructive(self, mpshort_model): self.add_broken_test_data(mpshort_model) mpshort_model.fix_tree(destructive=True) got = self.got(mpshort_model) expected = self.expected_no_holes[mpshort_model] assert got == expected mpshort_model.find_problems() def test_fix_tree_with_fix_paths(self, mpshort_model): self.add_broken_test_data(mpshort_model) mpshort_model.fix_tree(fix_paths=True) got = self.got(mpshort_model) expected = self.expected_no_holes[mpshort_model] assert got == expected mpshort_model.find_problems() @pytest.mark.django_db class TestIssues(TestTreeBase): # test for http://code.google.com/p/django-treebeard/issues/detail?id=14 def test_many_to_many_django_user_anonymous(self, mpm2muser_model): # Using AnonymousUser() in the querysets will expose non-treebeard # related problems in Django 1.0 # # Postgres: # ProgrammingError: can't adapt # SQLite: # InterfaceError: Error binding parameter 4 - probably unsupported # type. # MySQL compared a string to an integer field: # `treebeard_mp_testissue14_users`.`user_id` = 'AnonymousUser' # # Using a None field instead works (will be translated to IS NULL). # # anonuserobj = AnonymousUser() anonuserobj = None def qs_check(qs, expected): assert [o.name for o in qs] == expected def qs_check_first_or_user(expected, root, user): qs_check( root.get_children().filter(Q(name="first") | Q(users=user)), expected ) user = User.objects.create_user("test_user", "test@example.com", "testpasswd") user.save() root = mpm2muser_model.add_root(name="the root node") root.add_child(name="first") second = root.add_child(name="second") qs_check(root.get_children(), ["first", "second"]) qs_check(root.get_children().filter(Q(name="first")), ["first"]) qs_check(root.get_children().filter(Q(users=user)), []) qs_check_first_or_user(["first"], root, user) qs_check_first_or_user(["first", "second"], root, anonuserobj) user = User.objects.get(username="test_user") second.users.add(user) qs_check_first_or_user(["first", "second"], root, user) qs_check_first_or_user(["first"], root, anonuserobj) @pytest.mark.django_db class TestMoveNodeForm(TestNonEmptyTree): def _get_nodes_list(self, nodes): return [ (pk, "%s%s" % ("&nbsp;" * 4 * (depth - 1), str)) for pk, str, depth in nodes ] def _assert_nodes_in_choices(self, form, nodes): choices = form.fields["_ref_node_id"].choices assert choices.pop(0)[0] is None assert nodes == [(choice[0], choice[1]) for choice in choices] def _move_node_helper(self, node, safe_parent_nodes): form_class = movenodeform_factory(type(node)) form = form_class(instance=node) assert ["desc", "_position", "_ref_node_id"] == list(form.base_fields.keys()) got = [choice[0] for choice in form.fields["_position"].choices] assert ["first-child", "left", "right"] == got nodes = self._get_nodes_list(safe_parent_nodes) self._assert_nodes_in_choices(form, nodes) def _get_node_ids_strs_and_depths(self, nodes): return [(node.pk, str(node), node.get_depth()) for node in nodes] def test_form_root_node(self, model): nodes = list(model.get_tree()) node = nodes.pop(0) safe_parent_nodes = self._get_node_ids_strs_and_depths(nodes) self._move_node_helper(node, safe_parent_nodes) def test_form_leaf_node(self, model): nodes = list(model.get_tree()) safe_parent_nodes = self._get_node_ids_strs_and_depths(nodes) node = nodes.pop() self._move_node_helper(node, safe_parent_nodes) def test_form_admin(self, model): request = None nodes = list(model.get_tree()) safe_parent_nodes = self._get_node_ids_strs_and_depths(nodes) for node in model.objects.all(): site = AdminSite() form_class = movenodeform_factory(model) admin_class = admin_factory(form_class) ma = admin_class(model, site) got = list(ma.get_form(request).base_fields.keys()) desc_pos_refnodeid = ["desc", "_position", "_ref_node_id"] assert desc_pos_refnodeid == got got = ma.get_fieldsets(request) expected = [(None, {"fields": desc_pos_refnodeid})] assert got == expected got = ma.get_fieldsets(request, node) assert got == expected form = ma.get_form(request)() nodes = self._get_nodes_list(safe_parent_nodes) self._assert_nodes_in_choices(form, nodes) @pytest.mark.django_db class TestModelAdmin(TestNonEmptyTree): def test_default_fields(self, model): site = AdminSite() form_class = movenodeform_factory(model) admin_class = admin_factory(form_class) ma = admin_class(model, site) assert list(ma.get_form(None).base_fields.keys()) == [ "desc", "_position", "_ref_node_id", ] @pytest.mark.django_db class TestSortedForm(TestTreeSorted): def test_sorted_form(self, sorted_model): sorted_model.add_root(val1=3, val2=3, desc="zxy") sorted_model.add_root(val1=1, val2=4, desc="bcd") sorted_model.add_root(val1=2, val2=5, desc="zxy") sorted_model.add_root(val1=3, val2=3, desc="abc") sorted_model.add_root(val1=4, val2=1, desc="fgh") sorted_model.add_root(val1=3, val2=3, desc="abc") sorted_model.add_root(val1=2, val2=2, desc="qwe") sorted_model.add_root(val1=3, val2=2, desc="vcx") form_class = movenodeform_factory(sorted_model) form = form_class() assert list(form.fields.keys()) == [ "val1", "val2", "desc", "_position", "_ref_node_id", ] form = form_class(instance=sorted_model.objects.get(desc="bcd")) assert list(form.fields.keys()) == [ "val1", "val2", "desc", "_position", "_ref_node_id", ] assert "id__position" in str(form) assert "id__ref_node_id" in str(form) @pytest.mark.django_db class TestForm(TestNonEmptyTree): def test_form(self, model): form_class = movenodeform_factory(model) form = form_class() assert list(form.fields.keys()) == ["desc", "_position", "_ref_node_id"] form = form_class(instance=model.objects.get(desc="1")) assert list(form.fields.keys()) == ["desc", "_position", "_ref_node_id"] assert "id__position" in str(form) assert "id__ref_node_id" in str(form) def test_move_node_form(self, model): form_class = movenodeform_factory(model) bad_node = model.objects.get(desc="1").add_child( desc='Benign<script>alert("Compromised");</script>' ) form = form_class(instance=bad_node) rendered_html = form.as_p() assert "Benign" in rendered_html assert "<script>" not in rendered_html assert "&lt;script&gt;" in rendered_html def test_get_position_ref_node(self, model): form_class = movenodeform_factory(model) instance_parent = model.objects.get(desc="1") form = form_class(instance=instance_parent) assert form._get_position_ref_node(instance_parent) == { "_position": "first-child", "_ref_node_id": "", } instance_child = model.objects.get(desc="21") form = form_class(instance=instance_child) assert form._get_position_ref_node(instance_child) == { "_position": "first-child", "_ref_node_id": model.objects.get(desc="2").pk, } instance_grandchild = model.objects.get(desc="22") form = form_class(instance=instance_grandchild) assert form._get_position_ref_node(instance_grandchild) == { "_position": "right", "_ref_node_id": model.objects.get(desc="21").pk, } instance_grandchild = model.objects.get(desc="231") form = form_class(instance=instance_grandchild) assert form._get_position_ref_node(instance_grandchild) == { "_position": "first-child", "_ref_node_id": model.objects.get(desc="23").pk, } def test_clean_cleaned_data(self, model): instance_parent = model.objects.get(desc="1") _position = "first-child" _ref_node_id = "" form_class = movenodeform_factory(model) form = form_class( instance=instance_parent, data={ "_position": _position, "_ref_node_id": _ref_node_id, "desc": instance_parent.desc, }, ) assert form.is_valid() assert form._clean_cleaned_data() == (_position, _ref_node_id) def test_save_edit(self, model): instance_parent = model.objects.get(desc="1") original_count = len(model.objects.all()) form_class = movenodeform_factory(model) form = form_class( instance=instance_parent, data={ "_position": "first-child", "_ref_node_id": model.objects.get(desc="2").pk, "desc": instance_parent.desc, }, ) assert form.is_valid() saved_instance = form.save() assert original_count == model.objects.all().count() assert saved_instance.get_children_count() == 0 assert saved_instance.get_depth() == 2 assert not saved_instance.is_root() assert saved_instance.is_leaf() # Return to original state form_class = movenodeform_factory(model) form = form_class( instance=saved_instance, data={ "_position": "first-child", "_ref_node_id": "", "desc": saved_instance.desc, }, ) assert form.is_valid() restored_instance = form.save() assert original_count == model.objects.all().count() assert restored_instance.get_children_count() == 0 assert restored_instance.get_depth() == 1 assert restored_instance.is_root() assert restored_instance.is_leaf() def test_save_new(self, model): original_count = model.objects.all().count() assert original_count == 10 _position = "first-child" form_class = movenodeform_factory(model) form = form_class(data={"_position": _position, "desc": "New Form Test"}) assert form.is_valid() assert form.save() is not None assert original_count < model.objects.all().count() def test_save_new_with_pk_set(self, model): """ If the model is using a natural primary key then it will be already set when the instance is inserted. """ original_count = model.objects.all().count() assert original_count == 10 _position = "first-child" form_class = movenodeform_factory(model) form = form_class( data={"_position": _position, "id": 999999, "desc": "New Form Test"} ) assert form.is_valid() # Fake a natural key by updating the instance directly, because # the model form will have removed the id from cleaned data because # it thinks it is an AutoField. form.instance.id = 999999 assert form.save() is not None assert original_count < model.objects.all().count() class TestAdminTreeTemplateTags(TestCase): def test_treebeard_css(self): template = Template("{% load admin_tree %}{% treebeard_css %}") context = Context() rendered = template.render(context) expected = ( '<link rel="stylesheet" type="text/css" ' 'href="' + static("treebeard/treebeard-admin.css") + '"/>' ) assert expected == rendered def test_treebeard_js(self): template = Template("{% load admin_tree %}{% treebeard_js %}") context = Context() rendered = template.render(context) expected = ( '<script type="text/javascript" src="jsi18n"></script>' '<script type="text/javascript" ' 'src="' + static("treebeard/treebeard-admin.js") + '"></script>' "<script>(function($){" "jQuery = $.noConflict(true);" "})(django.jQuery);</script>" '<script type="text/javascript" ' 'src="' + static("treebeard/jquery-ui-1.8.5.custom.min.js") + '"></script>' ) assert expected == rendered @pytest.mark.django_db class TestAdminTree(TestNonEmptyTree): template = Template( "{% load admin_tree %}{% spaceless %}" "{% result_tree cl request %}{% endspaceless %}" ) def test_result_tree(self, model_without_proxy): """ Verifies that inclusion tag result_list generates a table when with default ModelAdmin settings. """ model = model_without_proxy request = RequestFactory().get("/admin/tree/") request.user = AnonymousUser() site = AdminSite() form_class = movenodeform_factory(model) admin_class = admin_factory(form_class) m = admin_class(model, site) list_display = m.get_list_display(request) list_display_links = m.get_list_display_links(request, list_display) cl = ChangeList( request, model, list_display, list_display_links, m.list_filter, m.date_hierarchy, m.search_fields, m.list_select_related, m.list_per_page, m.list_max_show_all, m.list_editable, m, None, ) cl.formset = None context = Context({"cl": cl, "request": request}) table_output = self.template.render(context) # We have the same amount of drag handlers as objects drag_handler = '<td class="drag-handler"><span>&nbsp;</span></td>' assert table_output.count(drag_handler) == model.objects.count() # All nodes are in the result tree for object in model.objects.all(): url = cl.url_for_result(object) node = '<a href="%s">%s</a>' % (url, str(object)) assert node in table_output # Unfiltered assert '<input type="hidden" id="has-filters" value="0"/>' in table_output def test_unicode_result_tree(self, model_with_unicode): """ Verifies that inclusion tag result_list generates a table when with default ModelAdmin settings. """ model = model_with_unicode # Add a unicode description model.add_root(desc="áéîøü") request = RequestFactory().get("/admin/tree/") request.user = AnonymousUser() site = AdminSite() form_class = movenodeform_factory(model) ModelAdmin = admin_factory(form_class) class UnicodeModelAdmin(ModelAdmin): list_display = ("__str__", "desc") m = UnicodeModelAdmin(model, site) list_display = m.get_list_display(request) list_display_links = m.get_list_display_links(request, list_display) cl = ChangeList( request, model, list_display, list_display_links, m.list_filter, m.date_hierarchy, m.search_fields, m.list_select_related, m.list_per_page, m.list_max_show_all, m.list_editable, m, None, ) cl.formset = None context = Context({"cl": cl, "request": request}) table_output = self.template.render(context) # We have the same amount of drag handlers as objects drag_handler = '<td class="drag-handler"><span>&nbsp;</span></td>' assert table_output.count(drag_handler) == model.objects.count() # All nodes are in the result tree for object in model.objects.all(): url = cl.url_for_result(object) node = '<a href="%s">%s</a>' % (url, object.desc) assert node in table_output # Unfiltered assert '<input type="hidden" id="has-filters" value="0"/>' in table_output def test_result_filtered(self, model_without_proxy): """Test template changes with filters or pagination.""" model = model_without_proxy # Filtered GET request = RequestFactory().get("/admin/tree/?desc=1") request.user = AnonymousUser() site = AdminSite() form_class = movenodeform_factory(model) admin_class = admin_factory(form_class) m = admin_class(model, site) list_display = m.get_list_display(request) list_display_links = m.get_list_display_links(request, list_display) cl = ChangeList( request, model, list_display, list_display_links, m.list_filter, m.date_hierarchy, m.search_fields, m.list_select_related, m.list_per_page, m.list_max_show_all, m.list_editable, m, None, ) cl.formset = None context = Context({"cl": cl, "request": request}) table_output = self.template.render(context) # Filtered assert '<input type="hidden" id="has-filters" value="1"/>' in table_output # Not Filtered GET, it should ignore pagination request = RequestFactory().get("/admin/tree/?p=1") request.user = AnonymousUser() list_display = m.get_list_display(request) list_display_links = m.get_list_display_links(request, list_display) cl = ChangeList( request, model, list_display, list_display_links, m.list_filter, m.date_hierarchy, m.search_fields, m.list_select_related, m.list_per_page, m.list_max_show_all, m.list_editable, m, None, ) cl.formset = None context = Context({"cl": cl, "request": request}) table_output = self.template.render(context) # Not Filtered assert '<input type="hidden" id="has-filters" value="0"/>' in table_output # Not Filtered GET, it should ignore all request = RequestFactory().get("/admin/tree/?all=1") request.user = AnonymousUser() list_display = m.get_list_display(request) list_display_links = m.get_list_display_links(request, list_display) cl = ChangeList( request, model, list_display, list_display_links, m.list_filter, m.date_hierarchy, m.search_fields, m.list_select_related, m.list_per_page, m.list_max_show_all, m.list_editable, m, None, ) cl.formset = None context = Context({"cl": cl, "request": request}) table_output = self.template.render(context) # Not Filtered assert '<input type="hidden" id="has-filters" value="0"/>' in table_output @pytest.mark.django_db class TestAdminTreeList(TestNonEmptyTree): template = Template( "{% load admin_tree_list %}{% spaceless %}" "{% result_tree cl request %}{% endspaceless %}" ) def test_result_tree_list(self, model_without_proxy): """ Verifies that inclusion tag result_list generates a table when with default ModelAdmin settings. """ model = model_without_proxy request = RequestFactory().get("/admin/tree/") request.user = AnonymousUser() site = AdminSite() form_class = movenodeform_factory(model) admin_class = admin_factory(form_class) m = admin_class(model, site) list_display = m.get_list_display(request) list_display_links = m.get_list_display_links(request, list_display) cl = ChangeList( request, model, list_display, list_display_links, m.list_filter, m.date_hierarchy, m.search_fields, m.list_select_related, m.list_per_page, m.list_max_show_all, m.list_editable, m, None, ) cl.formset = None context = Context({"cl": cl, "request": request}) table_output = self.template.render(context) output_template = '<li><a href="%s/" >%s</a>' for object in model.objects.all(): expected_output = output_template % (object.pk, str(object)) assert expected_output in table_output def test_result_tree_list_with_action(self, model_without_proxy): model = model_without_proxy request = RequestFactory().get("/admin/tree/") request.user = AnonymousUser() site = AdminSite() form_class = movenodeform_factory(model) admin_class = admin_factory(form_class) m = admin_class(model, site) list_display = m.get_list_display(request) list_display_links = m.get_list_display_links(request, list_display) cl = ChangeList( request, model, list_display, list_display_links, m.list_filter, m.date_hierarchy, m.search_fields, m.list_select_related, m.list_per_page, m.list_max_show_all, m.list_editable, m, None, ) cl.formset = None context = Context({"cl": cl, "request": request, "action_form": True}) table_output = self.template.render(context) output_template = ( '<input type="checkbox" class="action-select" ' 'value="%s" name="_selected_action" />' '<a href="%s/" >%s</a>' ) for object in model.objects.all(): expected_output = output_template % (object.pk, object.pk, str(object)) assert expected_output in table_output def test_result_tree_list_with_get(self, model_without_proxy): model = model_without_proxy pk_field = model._meta.pk.attname # Test t GET parameter with value id request = RequestFactory().get( "/admin/tree/?{0}={1}".format(TO_FIELD_VAR, pk_field) ) request.user = AnonymousUser() site = AdminSite() admin_register_all(site) form_class = movenodeform_factory(model) admin_class = admin_factory(form_class) m = admin_class(model, site) list_display = m.get_list_display(request) list_display_links = m.get_list_display_links(request, list_display) cl = ChangeList( request, model, list_display, list_display_links, m.list_filter, m.date_hierarchy, m.search_fields, m.list_select_related, m.list_per_page, m.list_max_show_all, m.list_editable, m, None, ) cl.formset = None context = Context({"cl": cl, "request": request}) table_output = self.template.render(context) output_template = "opener.dismissRelatedLookupPopup(window, '%s');" for object in model.objects.all(): expected_output = output_template % object.pk assert expected_output in table_output @pytest.mark.django_db class TestTreeAdmin(TestNonEmptyTree): site = AdminSite() def _create_superuser(self, username): return User.objects.create(username=username, is_superuser=True) def _mocked_authenticated_request(self, url, user): request_factory = RequestFactory() request = request_factory.get(url) request.user = user return request def _mocked_request(self, data): request_factory = RequestFactory() request = request_factory.post("/", data=data) setattr(request, "session", "session") messages = FallbackStorage(request) setattr(request, "_messages", messages) return request def _get_admin_obj(self, model_class): form_class = movenodeform_factory(model_class) admin_class = admin_factory(form_class) return admin_class(model_class, self.site) def test_changelist_view(self): tmp_user = self._create_superuser("changelist_tmp") request = self._mocked_authenticated_request("/", tmp_user) admin_obj = self._get_admin_obj(models.AL_TestNode) admin_obj.changelist_view(request) assert admin_obj.change_list_template == "admin/tree_list.html" admin_obj = self._get_admin_obj(models.MP_TestNode) admin_obj.changelist_view(request) assert admin_obj.change_list_template != "admin/tree_list.html" def test_get_node(self, model): admin_obj = self._get_admin_obj(model) target = model.objects.get(desc="2") assert admin_obj.get_node(target.pk) == target def test_move_node_validate_keyerror(self, model): admin_obj = self._get_admin_obj(model) request = self._mocked_request(data={}) response = admin_obj.move_node(request) assert response.status_code == 400 request = self._mocked_request(data={"node_id": 1}) response = admin_obj.move_node(request) assert response.status_code == 400 def test_move_node_validate_valueerror(self, model): admin_obj = self._get_admin_obj(model) request = self._mocked_request( data={"node_id": 1, "sibling_id": 2, "as_child": "invalid"} ) response = admin_obj.move_node(request) assert response.status_code == 400 def test_move_validate_missing_nodeorderby(self, model): node = model.objects.get(desc="231") admin_obj = self._get_admin_obj(model) request = self._mocked_request(data={}) response = admin_obj.try_to_move_node( True, node, "sorted-child", request, target=node ) assert response.status_code == 400 response = admin_obj.try_to_move_node( True, node, "sorted-sibling", request, target=node ) assert response.status_code == 400 def test_move_validate_invalid_pos(self, model): node = model.objects.get(desc="231") admin_obj = self._get_admin_obj(model) request = self._mocked_request(data={}) response = admin_obj.try_to_move_node( True, node, "invalid_pos", request, target=node ) assert response.status_code == 400 def test_move_validate_to_descendant(self, model): node = model.objects.get(desc="2") target = model.objects.get(desc="231") admin_obj = self._get_admin_obj(model) request = self._mocked_request(data={}) response = admin_obj.try_to_move_node( True, node, "first-sibling", request, target ) assert response.status_code == 400 def test_move_left(self, model): node = model.objects.get(desc="231") target = model.objects.get(desc="2") admin_obj = self._get_admin_obj(model) request = self._mocked_request( data={"node_id": node.pk, "sibling_id": target.pk, "as_child": 0} ) response = admin_obj.move_node(request) assert response.status_code == 200 expected = [ ("1", 1, 0), ("231", 1, 0), ("2", 1, 4), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected def test_move_last_child(self, model): node = model.objects.get(desc="231") target = model.objects.get(desc="2") admin_obj = self._get_admin_obj(model) request = self._mocked_request( data={"node_id": node.pk, "sibling_id": target.pk, "as_child": 1} ) response = admin_obj.move_node(request) assert response.status_code == 200 expected = [ ("1", 1, 0), ("2", 1, 5), ("21", 2, 0), ("22", 2, 0), ("23", 2, 0), ("24", 2, 0), ("231", 2, 0), ("3", 1, 0), ("4", 1, 1), ("41", 2, 0), ] assert self.got(model) == expected @pytest.mark.django_db class TestMPFormPerformance(object): def test_form_add_subtree_no_of_queries(self, django_assert_num_queries): model = models.MP_TestNode model.load_bulk(BASE_DATA) form_class = movenodeform_factory(model) form = form_class() with django_assert_num_queries(len(model.get_root_nodes()) + 1): form.mk_dropdown_tree(model)
35.093061
89
0.512903
794d7fba919a87336fbca3c2e7acbd8dcb18c680
1,047
py
Python
manage.py
valerymelou/opentimesheet-server
0da97ebb3c3e59962132d1bc5e83e1d727f7331b
[ "MIT" ]
null
null
null
manage.py
valerymelou/opentimesheet-server
0da97ebb3c3e59962132d1bc5e83e1d727f7331b
[ "MIT" ]
95
2021-02-20T21:53:29.000Z
2022-01-14T17:24:50.000Z
manage.py
valerymelou/opentimesheet-server
0da97ebb3c3e59962132d1bc5e83e1d727f7331b
[ "MIT" ]
null
null
null
#!/usr/bin/env python import os import sys from pathlib import Path if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "config.settings.local") try: from django.core.management import execute_from_command_line except ImportError: # The above import may fail for some other reason. Ensure that the # issue is really that Django is missing to avoid masking other # exceptions on Python 2. try: import django # noqa except ImportError: raise ImportError( "Couldn't import Django. Are you sure it's installed and " "available on your PYTHONPATH environment variable? Did you " "forget to activate a virtual environment?" ) raise # This allows easy placement of apps within the interior # opentimesheet directory. current_path = Path(__file__).parent.resolve() sys.path.append(str(current_path / "opentimesheet")) execute_from_command_line(sys.argv)
32.71875
77
0.660936
794d809366b828f9bc7210f343e24d674a17042c
1,240
py
Python
pydash/Dash/DashLint/__init__.py
ensomniac/dash
5a5cabd1a1d057015dd1446b6b1000af1e521355
[ "MIT" ]
null
null
null
pydash/Dash/DashLint/__init__.py
ensomniac/dash
5a5cabd1a1d057015dd1446b6b1000af1e521355
[ "MIT" ]
null
null
null
pydash/Dash/DashLint/__init__.py
ensomniac/dash
5a5cabd1a1d057015dd1446b6b1000af1e521355
[ "MIT" ]
null
null
null
# Ensomniac 2022 Ryan Martin, ryan@ensomniac.com # Andrew Stet, stetandrew@gmail.com """ Stub for a module to run whenever a python or javascript file is saved. Unresolved Notes / Thoughts for Python: + Time the compilation / read time for each file and flag time-consuming imports + Create a format for comment / description (like this) and force it to exist - for docs? + Consider a formatting ability to specify that the name of a function had changed so that we can auto migrate other code + For js lint: Find 'print' and change to 'console.log'! """ import os import sys from .js_lint import JSLinter from .py_lint import PyLinter from .lint_utils import LintUtils class _DashLint: def __init__(self): pass def Process(self, is_client, dash_context, code_path): lint_succeeded = True msg = "(DashLint) Did not run PyLinter or JSLinter. What's the file's extension?" if code_path.endswith(".py"): lint_succeeded, msg = PyLinter.Process(is_client, dash_context, code_path) elif code_path.endswith(".js"): lint_succeeded, msg = JSLinter.Process(is_client, dash_context, code_path) return lint_succeeded, msg DashLint = _DashLint()
31
121
0.71129
794d81230b1a8cffc44a73f8fa8ec90d51ec1ebd
47,921
py
Python
mne/coreg.py
mohdsherif/mne-python
affc6854168e32e73a075a0104e7af8bddd7eefe
[ "BSD-3-Clause" ]
null
null
null
mne/coreg.py
mohdsherif/mne-python
affc6854168e32e73a075a0104e7af8bddd7eefe
[ "BSD-3-Clause" ]
null
null
null
mne/coreg.py
mohdsherif/mne-python
affc6854168e32e73a075a0104e7af8bddd7eefe
[ "BSD-3-Clause" ]
null
null
null
# -*- coding: utf-8 -*- """Coregistration between different coordinate frames.""" # Authors: Christian Brodbeck <christianbrodbeck@nyu.edu> # # License: BSD (3-clause) import configparser import fnmatch from glob import glob, iglob import os import os.path as op import stat import sys import re import shutil from functools import reduce import numpy as np from .io import read_fiducials, write_fiducials, read_info from .io.constants import FIFF from .label import read_label, Label from .source_space import (add_source_space_distances, read_source_spaces, write_source_spaces, _get_mri_header, _read_talxfm) from .surface import read_surface, write_surface, _normalize_vectors from .bem import read_bem_surfaces, write_bem_surfaces from .transforms import (rotation, rotation3d, scaling, translation, Transform, _read_fs_xfm, _write_fs_xfm, invert_transform, combine_transforms, apply_trans) from .utils import (get_config, get_subjects_dir, logger, pformat, verbose, warn, has_nibabel) from .viz._3d import _fiducial_coords # some path templates trans_fname = os.path.join('{raw_dir}', '{subject}-trans.fif') subject_dirname = os.path.join('{subjects_dir}', '{subject}') bem_dirname = os.path.join(subject_dirname, 'bem') mri_dirname = os.path.join(subject_dirname, 'mri') mri_transforms_dirname = os.path.join(subject_dirname, 'mri', 'transforms') surf_dirname = os.path.join(subject_dirname, 'surf') bem_fname = os.path.join(bem_dirname, "{subject}-{name}.fif") head_bem_fname = pformat(bem_fname, name='head') fid_fname = pformat(bem_fname, name='fiducials') fid_fname_general = os.path.join(bem_dirname, "{head}-fiducials.fif") src_fname = os.path.join(bem_dirname, '{subject}-{spacing}-src.fif') _head_fnames = (os.path.join(bem_dirname, 'outer_skin.surf'), head_bem_fname) _high_res_head_fnames = (os.path.join(bem_dirname, '{subject}-head-dense.fif'), os.path.join(surf_dirname, 'lh.seghead'), os.path.join(surf_dirname, 'lh.smseghead')) def _make_writable(fname): """Make a file writable.""" os.chmod(fname, stat.S_IMODE(os.lstat(fname)[stat.ST_MODE]) | 128) # write def _make_writable_recursive(path): """Recursively set writable.""" if sys.platform.startswith('win'): return # can't safely set perms for root, dirs, files in os.walk(path, topdown=False): for f in dirs + files: _make_writable(os.path.join(root, f)) def _find_head_bem(subject, subjects_dir, high_res=False): """Find a high resolution head.""" # XXX this should be refactored with mne.surface.get_head_surf ... fnames = _high_res_head_fnames if high_res else _head_fnames for fname in fnames: path = fname.format(subjects_dir=subjects_dir, subject=subject) if os.path.exists(path): return path def coregister_fiducials(info, fiducials, tol=0.01): """Create a head-MRI transform by aligning 3 fiducial points. Parameters ---------- info : Info Measurement info object with fiducials in head coordinate space. fiducials : str | list of dict Fiducials in MRI coordinate space (either path to a ``*-fiducials.fif`` file or list of fiducials as returned by :func:`read_fiducials`. Returns ------- trans : Transform The device-MRI transform. """ if isinstance(info, str): info = read_info(info) if isinstance(fiducials, str): fiducials, coord_frame_to = read_fiducials(fiducials) else: coord_frame_to = FIFF.FIFFV_COORD_MRI frames_from = {d['coord_frame'] for d in info['dig']} if len(frames_from) > 1: raise ValueError("info contains fiducials from different coordinate " "frames") else: coord_frame_from = frames_from.pop() coords_from = _fiducial_coords(info['dig']) coords_to = _fiducial_coords(fiducials, coord_frame_to) trans = fit_matched_points(coords_from, coords_to, tol=tol) return Transform(coord_frame_from, coord_frame_to, trans) @verbose def create_default_subject(fs_home=None, update=False, subjects_dir=None, verbose=None): """Create an average brain subject for subjects without structural MRI. Create a copy of fsaverage from the Freesurfer directory in subjects_dir and add auxiliary files from the mne package. Parameters ---------- fs_home : None | str The freesurfer home directory (only needed if FREESURFER_HOME is not specified as environment variable). update : bool In cases where a copy of the fsaverage brain already exists in the subjects_dir, this option allows to only copy files that don't already exist in the fsaverage directory. subjects_dir : None | str Override the SUBJECTS_DIR environment variable (os.environ['SUBJECTS_DIR']) as destination for the new subject. %(verbose)s Notes ----- When no structural MRI is available for a subject, an average brain can be substituted. Freesurfer comes with such an average brain model, and MNE comes with some auxiliary files which make coregistration easier. :py:func:`create_default_subject` copies the relevant files from Freesurfer into the current subjects_dir, and also adds the auxiliary files provided by MNE. """ subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) if fs_home is None: fs_home = get_config('FREESURFER_HOME', fs_home) if fs_home is None: raise ValueError( "FREESURFER_HOME environment variable not found. Please " "specify the fs_home parameter in your call to " "create_default_subject().") # make sure freesurfer files exist fs_src = os.path.join(fs_home, 'subjects', 'fsaverage') if not os.path.exists(fs_src): raise IOError('fsaverage not found at %r. Is fs_home specified ' 'correctly?' % fs_src) for name in ('label', 'mri', 'surf'): dirname = os.path.join(fs_src, name) if not os.path.isdir(dirname): raise IOError("Freesurfer fsaverage seems to be incomplete: No " "directory named %s found in %s" % (name, fs_src)) # make sure destination does not already exist dest = os.path.join(subjects_dir, 'fsaverage') if dest == fs_src: raise IOError( "Your subjects_dir points to the freesurfer subjects_dir (%r). " "The default subject can not be created in the freesurfer " "installation directory; please specify a different " "subjects_dir." % subjects_dir) elif (not update) and os.path.exists(dest): raise IOError( "Can not create fsaverage because %r already exists in " "subjects_dir %r. Delete or rename the existing fsaverage " "subject folder." % ('fsaverage', subjects_dir)) # copy fsaverage from freesurfer logger.info("Copying fsaverage subject from freesurfer directory...") if (not update) or not os.path.exists(dest): shutil.copytree(fs_src, dest) _make_writable_recursive(dest) # copy files from mne source_fname = os.path.join(os.path.dirname(__file__), 'data', 'fsaverage', 'fsaverage-%s.fif') dest_bem = os.path.join(dest, 'bem') if not os.path.exists(dest_bem): os.mkdir(dest_bem) logger.info("Copying auxiliary fsaverage files from mne...") dest_fname = os.path.join(dest_bem, 'fsaverage-%s.fif') _make_writable_recursive(dest_bem) for name in ('fiducials', 'head', 'inner_skull-bem', 'trans'): if not os.path.exists(dest_fname % name): shutil.copy(source_fname % name, dest_bem) def _decimate_points(pts, res=10): """Decimate the number of points using a voxel grid. Create a voxel grid with a specified resolution and retain at most one point per voxel. For each voxel, the point closest to its center is retained. Parameters ---------- pts : array, shape (n_points, 3) The points making up the head shape. res : scalar The resolution of the voxel space (side length of each voxel). Returns ------- pts : array, shape = (n_points, 3) The decimated points. """ from scipy.spatial.distance import cdist pts = np.asarray(pts) # find the bin edges for the voxel space xmin, ymin, zmin = pts.min(0) - res / 2. xmax, ymax, zmax = pts.max(0) + res xax = np.arange(xmin, xmax, res) yax = np.arange(ymin, ymax, res) zax = np.arange(zmin, zmax, res) # find voxels containing one or more point H, _ = np.histogramdd(pts, bins=(xax, yax, zax), normed=False) # for each voxel, select one point X, Y, Z = pts.T out = np.empty((np.sum(H > 0), 3)) for i, (xbin, ybin, zbin) in enumerate(zip(*np.nonzero(H))): x = xax[xbin] y = yax[ybin] z = zax[zbin] xi = np.logical_and(X >= x, X < x + res) yi = np.logical_and(Y >= y, Y < y + res) zi = np.logical_and(Z >= z, Z < z + res) idx = np.logical_and(zi, np.logical_and(yi, xi)) ipts = pts[idx] mid = np.array([x, y, z]) + res / 2. dist = cdist(ipts, [mid]) i_min = np.argmin(dist) ipt = ipts[i_min] out[i] = ipt return out def _trans_from_params(param_info, params): """Convert transformation parameters into a transformation matrix. Parameters ---------- param_info : tuple, len = 3 Tuple describing the parameters in x (do_translate, do_rotate, do_scale). params : tuple The transformation parameters. Returns ------- trans : array, shape = (4, 4) Transformation matrix. """ do_rotate, do_translate, do_scale = param_info i = 0 trans = [] if do_rotate: x, y, z = params[:3] trans.append(rotation(x, y, z)) i += 3 if do_translate: x, y, z = params[i:i + 3] trans.insert(0, translation(x, y, z)) i += 3 if do_scale == 1: s = params[i] trans.append(scaling(s, s, s)) elif do_scale == 3: x, y, z = params[i:i + 3] trans.append(scaling(x, y, z)) trans = reduce(np.dot, trans) return trans # XXX this function should be moved out of coreg as used elsewhere def fit_matched_points(src_pts, tgt_pts, rotate=True, translate=True, scale=False, tol=None, x0=None, out='trans', weights=None): """Find a transform between matched sets of points. This minimizes the squared distance between two matching sets of points. Uses :func:`scipy.optimize.leastsq` to find a transformation involving a combination of rotation, translation, and scaling (in that order). Parameters ---------- src_pts : array, shape = (n, 3) Points to which the transform should be applied. tgt_pts : array, shape = (n, 3) Points to which src_pts should be fitted. Each point in tgt_pts should correspond to the point in src_pts with the same index. rotate : bool Allow rotation of the ``src_pts``. translate : bool Allow translation of the ``src_pts``. scale : bool Number of scaling parameters. With False, points are not scaled. With True, points are scaled by the same factor along all axes. tol : scalar | None The error tolerance. If the distance between any of the matched points exceeds this value in the solution, a RuntimeError is raised. With None, no error check is performed. x0 : None | tuple Initial values for the fit parameters. out : 'params' | 'trans' In what format to return the estimate: 'params' returns a tuple with the fit parameters; 'trans' returns a transformation matrix of shape (4, 4). Returns ------- trans : array, shape (4, 4) Transformation that, if applied to src_pts, minimizes the squared distance to tgt_pts. Only returned if out=='trans'. params : array, shape (n_params, ) A single tuple containing the rotation, translation, and scaling parameters in that order (as applicable). """ # XXX eventually this should be refactored with the cHPI fitting code, # which use fmin_cobyla with constraints from scipy.optimize import leastsq src_pts = np.atleast_2d(src_pts) tgt_pts = np.atleast_2d(tgt_pts) if src_pts.shape != tgt_pts.shape: raise ValueError("src_pts and tgt_pts must have same shape (got " "{}, {})".format(src_pts.shape, tgt_pts.shape)) if weights is not None: weights = np.array(weights, float) if weights.ndim != 1 or weights.size not in (src_pts.shape[0], 1): raise ValueError("weights (shape=%s) must be None or have shape " "(%s,)" % (weights.shape, src_pts.shape[0],)) weights = weights[:, np.newaxis] rotate = bool(rotate) translate = bool(translate) scale = int(scale) if translate: src_pts = np.hstack((src_pts, np.ones((len(src_pts), 1)))) param_info = (rotate, translate, scale) if param_info == (True, False, 0): def error(x): rx, ry, rz = x trans = rotation3d(rx, ry, rz) est = np.dot(src_pts, trans.T) d = tgt_pts - est if weights is not None: d *= weights return d.ravel() if x0 is None: x0 = (0, 0, 0) elif param_info == (True, True, 0): def error(x): rx, ry, rz, tx, ty, tz = x trans = np.dot(translation(tx, ty, tz), rotation(rx, ry, rz)) est = np.dot(src_pts, trans.T)[:, :3] d = tgt_pts - est if weights is not None: d *= weights return d.ravel() if x0 is None: x0 = (0, 0, 0, 0, 0, 0) elif param_info == (True, True, 1): def error(x): rx, ry, rz, tx, ty, tz, s = x trans = reduce(np.dot, (translation(tx, ty, tz), rotation(rx, ry, rz), scaling(s, s, s))) est = np.dot(src_pts, trans.T)[:, :3] d = tgt_pts - est if weights is not None: d *= weights return d.ravel() if x0 is None: x0 = (0, 0, 0, 0, 0, 0, 1) elif param_info == (True, True, 3): def error(x): rx, ry, rz, tx, ty, tz, sx, sy, sz = x trans = reduce(np.dot, (translation(tx, ty, tz), rotation(rx, ry, rz), scaling(sx, sy, sz))) est = np.dot(src_pts, trans.T)[:, :3] d = tgt_pts - est if weights is not None: d *= weights return d.ravel() if x0 is None: x0 = (0, 0, 0, 0, 0, 0, 1, 1, 1) else: raise NotImplementedError( "The specified parameter combination is not implemented: " "rotate=%r, translate=%r, scale=%r" % param_info) x, _, _, _, _ = leastsq(error, x0, full_output=True) # re-create the final transformation matrix if (tol is not None) or (out == 'trans'): trans = _trans_from_params(param_info, x) # assess the error of the solution if tol is not None: if not translate: src_pts = np.hstack((src_pts, np.ones((len(src_pts), 1)))) est_pts = np.dot(src_pts, trans.T)[:, :3] err = np.sqrt(np.sum((est_pts - tgt_pts) ** 2, axis=1)) if np.any(err > tol): raise RuntimeError("Error exceeds tolerance. Error = %r" % err) if out == 'params': return x elif out == 'trans': return trans else: raise ValueError("Invalid out parameter: %r. Needs to be 'params' or " "'trans'." % out) def _find_label_paths(subject='fsaverage', pattern=None, subjects_dir=None): """Find paths to label files in a subject's label directory. Parameters ---------- subject : str Name of the mri subject. pattern : str | None Pattern for finding the labels relative to the label directory in the MRI subject directory (e.g., "aparc/*.label" will find all labels in the "subject/label/aparc" directory). With None, find all labels. subjects_dir : None | str Override the SUBJECTS_DIR environment variable (sys.environ['SUBJECTS_DIR']) Returns ------- paths : list List of paths relative to the subject's label directory """ subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) subject_dir = os.path.join(subjects_dir, subject) lbl_dir = os.path.join(subject_dir, 'label') if pattern is None: paths = [] for dirpath, _, filenames in os.walk(lbl_dir): rel_dir = os.path.relpath(dirpath, lbl_dir) for filename in fnmatch.filter(filenames, '*.label'): path = os.path.join(rel_dir, filename) paths.append(path) else: paths = [os.path.relpath(path, lbl_dir) for path in iglob(pattern)] return paths def _find_mri_paths(subject, skip_fiducials, subjects_dir): """Find all files of an mri relevant for source transformation. Parameters ---------- subject : str Name of the mri subject. skip_fiducials : bool Do not scale the MRI fiducials. If False, an IOError will be raised if no fiducials file can be found. subjects_dir : None | str Override the SUBJECTS_DIR environment variable (sys.environ['SUBJECTS_DIR']) Returns ------- paths : dict Dictionary whose keys are relevant file type names (str), and whose values are lists of paths. """ subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) paths = {} # directories to create paths['dirs'] = [bem_dirname, surf_dirname] # surf/ files paths['surf'] = [] surf_fname = os.path.join(surf_dirname, '{name}') surf_names = ('inflated', 'white', 'orig', 'orig_avg', 'inflated_avg', 'inflated_pre', 'pial', 'pial_avg', 'smoothwm', 'white_avg', 'seghead', 'smseghead') if os.getenv('_MNE_FEW_SURFACES', '') == 'true': # for testing surf_names = surf_names[:4] for surf_name in surf_names: for hemi in ('lh.', 'rh.'): name = hemi + surf_name path = surf_fname.format(subjects_dir=subjects_dir, subject=subject, name=name) if os.path.exists(path): paths['surf'].append(pformat(surf_fname, name=name)) surf_fname = os.path.join(bem_dirname, '{name}') surf_names = ('inner_skull.surf', 'outer_skull.surf', 'outer_skin.surf') for surf_name in surf_names: path = surf_fname.format(subjects_dir=subjects_dir, subject=subject, name=surf_name) if os.path.exists(path): paths['surf'].append(pformat(surf_fname, name=surf_name)) del surf_names, surf_name, path, hemi # BEM files paths['bem'] = bem = [] path = head_bem_fname.format(subjects_dir=subjects_dir, subject=subject) if os.path.exists(path): bem.append('head') bem_pattern = pformat(bem_fname, subjects_dir=subjects_dir, subject=subject, name='*-bem') re_pattern = pformat(bem_fname, subjects_dir=subjects_dir, subject=subject, name='(.+)').replace('\\', '\\\\') for path in iglob(bem_pattern): match = re.match(re_pattern, path) name = match.group(1) bem.append(name) del bem, path, bem_pattern, re_pattern # fiducials if skip_fiducials: paths['fid'] = [] else: paths['fid'] = _find_fiducials_files(subject, subjects_dir) # check that we found at least one if len(paths['fid']) == 0: raise IOError("No fiducials file found for %s. The fiducials " "file should be named " "{subject}/bem/{subject}-fiducials.fif. In " "order to scale an MRI without fiducials set " "skip_fiducials=True." % subject) # duplicate files (curvature and some surfaces) paths['duplicate'] = [] path = os.path.join(surf_dirname, '{name}') surf_fname = os.path.join(surf_dirname, '{name}') surf_dup_names = ('curv', 'sphere', 'sphere.reg', 'sphere.reg.avg') for surf_dup_name in surf_dup_names: for hemi in ('lh.', 'rh.'): name = hemi + surf_dup_name path = surf_fname.format(subjects_dir=subjects_dir, subject=subject, name=name) if os.path.exists(path): paths['duplicate'].append(pformat(surf_fname, name=name)) del surf_dup_name, name, path, hemi # transform files (talairach) paths['transforms'] = [] transform_fname = os.path.join(mri_transforms_dirname, 'talairach.xfm') path = transform_fname.format(subjects_dir=subjects_dir, subject=subject) if os.path.exists(path): paths['transforms'].append(transform_fname) del transform_fname, path # find source space files paths['src'] = src = [] bem_dir = bem_dirname.format(subjects_dir=subjects_dir, subject=subject) fnames = fnmatch.filter(os.listdir(bem_dir), '*-src.fif') prefix = subject + '-' for fname in fnames: if fname.startswith(prefix): fname = "{subject}-%s" % fname[len(prefix):] path = os.path.join(bem_dirname, fname) src.append(path) # find MRIs mri_dir = mri_dirname.format(subjects_dir=subjects_dir, subject=subject) fnames = fnmatch.filter(os.listdir(mri_dir), '*.mgz') paths['mri'] = [os.path.join(mri_dir, f) for f in fnames] return paths def _find_fiducials_files(subject, subjects_dir): """Find fiducial files.""" fid = [] # standard fiducials if os.path.exists(fid_fname.format(subjects_dir=subjects_dir, subject=subject)): fid.append(fid_fname) # fiducials with subject name pattern = pformat(fid_fname_general, subjects_dir=subjects_dir, subject=subject, head='*') regex = pformat(fid_fname_general, subjects_dir=subjects_dir, subject=subject, head='(.+)').replace('\\', '\\\\') for path in iglob(pattern): match = re.match(regex, path) head = match.group(1).replace(subject, '{subject}') fid.append(pformat(fid_fname_general, head=head)) return fid def _is_mri_subject(subject, subjects_dir=None): """Check whether a directory in subjects_dir is an mri subject directory. Parameters ---------- subject : str Name of the potential subject/directory. subjects_dir : None | str Override the SUBJECTS_DIR environment variable. Returns ------- is_mri_subject : bool Whether ``subject`` is an mri subject. """ subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) return bool(_find_head_bem(subject, subjects_dir) or _find_head_bem(subject, subjects_dir, high_res=True)) def _is_scaled_mri_subject(subject, subjects_dir=None): """Check whether a directory in subjects_dir is a scaled mri subject. Parameters ---------- subject : str Name of the potential subject/directory. subjects_dir : None | str Override the SUBJECTS_DIR environment variable. Returns ------- is_scaled_mri_subject : bool Whether ``subject`` is a scaled mri subject. """ subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) if not _is_mri_subject(subject, subjects_dir): return False fname = os.path.join(subjects_dir, subject, 'MRI scaling parameters.cfg') return os.path.exists(fname) def _mri_subject_has_bem(subject, subjects_dir=None): """Check whether an mri subject has a file matching the bem pattern. Parameters ---------- subject : str Name of the subject. subjects_dir : None | str Override the SUBJECTS_DIR environment variable. Returns ------- has_bem_file : bool Whether ``subject`` has a bem file. """ subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) pattern = bem_fname.format(subjects_dir=subjects_dir, subject=subject, name='*-bem') fnames = glob(pattern) return bool(len(fnames)) def read_mri_cfg(subject, subjects_dir=None): """Read information from the cfg file of a scaled MRI brain. Parameters ---------- subject : str Name of the scaled MRI subject. subjects_dir : None | str Override the SUBJECTS_DIR environment variable. Returns ------- cfg : dict Dictionary with entries from the MRI's cfg file. """ subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) fname = os.path.join(subjects_dir, subject, 'MRI scaling parameters.cfg') if not os.path.exists(fname): raise IOError("%r does not seem to be a scaled mri subject: %r does " "not exist." % (subject, fname)) logger.info("Reading MRI cfg file %s" % fname) config = configparser.RawConfigParser() config.read(fname) n_params = config.getint("MRI Scaling", 'n_params') if n_params == 1: scale = config.getfloat("MRI Scaling", 'scale') elif n_params == 3: scale_str = config.get("MRI Scaling", 'scale') scale = np.array([float(s) for s in scale_str.split()]) else: raise ValueError("Invalid n_params value in MRI cfg: %i" % n_params) out = {'subject_from': config.get("MRI Scaling", 'subject_from'), 'n_params': n_params, 'scale': scale} return out def _write_mri_config(fname, subject_from, subject_to, scale): """Write the cfg file describing a scaled MRI subject. Parameters ---------- fname : str Target file. subject_from : str Name of the source MRI subject. subject_to : str Name of the scaled MRI subject. scale : float | array_like, shape = (3,) The scaling parameter. """ scale = np.asarray(scale) if np.isscalar(scale) or scale.shape == (): n_params = 1 else: n_params = 3 config = configparser.RawConfigParser() config.add_section("MRI Scaling") config.set("MRI Scaling", 'subject_from', subject_from) config.set("MRI Scaling", 'subject_to', subject_to) config.set("MRI Scaling", 'n_params', str(n_params)) if n_params == 1: config.set("MRI Scaling", 'scale', str(scale)) else: config.set("MRI Scaling", 'scale', ' '.join([str(s) for s in scale])) config.set("MRI Scaling", 'version', '1') with open(fname, 'w') as fid: config.write(fid) def _scale_params(subject_to, subject_from, scale, subjects_dir): """Assemble parameters for scaling. Returns ------- subjects_dir : str Subjects directory. subject_from : str Name of the source subject. scale : array Scaling factor, either shape=() for uniform scaling or shape=(3,) for non-uniform scaling. uniform : bool Whether scaling is uniform. """ subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) if (subject_from is None) != (scale is None): raise TypeError("Need to provide either both subject_from and scale " "parameters, or neither.") if subject_from is None: cfg = read_mri_cfg(subject_to, subjects_dir) subject_from = cfg['subject_from'] n_params = cfg['n_params'] assert n_params in (1, 3) scale = cfg['scale'] scale = np.atleast_1d(scale) if scale.ndim != 1 or scale.shape[0] not in (1, 3): raise ValueError("Invalid shape for scale parameer. Need scalar " "or array of length 3. Got shape %s." % (scale.shape,)) n_params = len(scale) return subjects_dir, subject_from, scale, n_params == 1 @verbose def scale_bem(subject_to, bem_name, subject_from=None, scale=None, subjects_dir=None, verbose=None): """Scale a bem file. Parameters ---------- subject_to : str Name of the scaled MRI subject (the destination mri subject). bem_name : str Name of the bem file. For example, to scale ``fsaverage-inner_skull-bem.fif``, the bem_name would be "inner_skull-bem". subject_from : None | str The subject from which to read the source space. If None, subject_from is read from subject_to's config file. scale : None | float | array, shape = (3,) Scaling factor. Has to be specified if subjects_from is specified, otherwise it is read from subject_to's config file. subjects_dir : None | str Override the SUBJECTS_DIR environment variable. %(verbose)s """ subjects_dir, subject_from, scale, uniform = \ _scale_params(subject_to, subject_from, scale, subjects_dir) src = bem_fname.format(subjects_dir=subjects_dir, subject=subject_from, name=bem_name) dst = bem_fname.format(subjects_dir=subjects_dir, subject=subject_to, name=bem_name) if os.path.exists(dst): raise IOError("File already exists: %s" % dst) surfs = read_bem_surfaces(src) for surf in surfs: surf['rr'] *= scale if not uniform: assert len(surf['nn']) > 0 surf['nn'] /= scale _normalize_vectors(surf['nn']) write_bem_surfaces(dst, surfs) def scale_labels(subject_to, pattern=None, overwrite=False, subject_from=None, scale=None, subjects_dir=None): r"""Scale labels to match a brain that was previously created by scaling. Parameters ---------- subject_to : str Name of the scaled MRI subject (the destination brain). pattern : str | None Pattern for finding the labels relative to the label directory in the MRI subject directory (e.g., "lh.BA3a.label" will scale "fsaverage/label/lh.BA3a.label"; "aparc/\*.label" will find all labels in the "fsaverage/label/aparc" directory). With None, scale all labels. overwrite : bool Overwrite any label file that already exists for subject_to (otherwise existing labels are skipped). subject_from : None | str Name of the original MRI subject (the brain that was scaled to create subject_to). If None, the value is read from subject_to's cfg file. scale : None | float | array_like, shape = (3,) Scaling parameter. If None, the value is read from subject_to's cfg file. subjects_dir : None | str Override the SUBJECTS_DIR environment variable. """ subjects_dir, subject_from, scale, _ = _scale_params( subject_to, subject_from, scale, subjects_dir) # find labels paths = _find_label_paths(subject_from, pattern, subjects_dir) if not paths: return subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) src_root = os.path.join(subjects_dir, subject_from, 'label') dst_root = os.path.join(subjects_dir, subject_to, 'label') # scale labels for fname in paths: dst = os.path.join(dst_root, fname) if not overwrite and os.path.exists(dst): continue dirname = os.path.dirname(dst) if not os.path.exists(dirname): os.makedirs(dirname) src = os.path.join(src_root, fname) l_old = read_label(src) pos = l_old.pos * scale l_new = Label(l_old.vertices, pos, l_old.values, l_old.hemi, l_old.comment, subject=subject_to) l_new.save(dst) @verbose def scale_mri(subject_from, subject_to, scale, overwrite=False, subjects_dir=None, skip_fiducials=False, labels=True, annot=False, verbose=None): """Create a scaled copy of an MRI subject. Parameters ---------- subject_from : str Name of the subject providing the MRI. subject_to : str New subject name for which to save the scaled MRI. scale : float | array_like, shape = (3,) The scaling factor (one or 3 parameters). overwrite : bool If an MRI already exists for subject_to, overwrite it. subjects_dir : None | str Override the SUBJECTS_DIR environment variable. skip_fiducials : bool Do not scale the MRI fiducials. If False (default), an IOError will be raised if no fiducials file can be found. labels : bool Also scale all labels (default True). annot : bool Copy ``*.annot`` files to the new location (default False). %(verbose)s See Also -------- scale_labels : add labels to a scaled MRI scale_source_space : add a source space to a scaled MRI """ subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) paths = _find_mri_paths(subject_from, skip_fiducials, subjects_dir) scale = np.atleast_1d(scale) if scale.shape == (3,): if np.isclose(scale[1], scale[0]) and np.isclose(scale[2], scale[0]): scale = scale[0] # speed up scaling conditionals using a singleton elif scale.shape != (1,): raise ValueError('scale must have shape (3,) or (1,), got %s' % (scale.shape,)) # make sure we have an empty target directory dest = subject_dirname.format(subject=subject_to, subjects_dir=subjects_dir) if os.path.exists(dest): if not overwrite: raise IOError("Subject directory for %s already exists: %r" % (subject_to, dest)) shutil.rmtree(dest) logger.debug('create empty directory structure') for dirname in paths['dirs']: dir_ = dirname.format(subject=subject_to, subjects_dir=subjects_dir) os.makedirs(dir_) logger.debug('save MRI scaling parameters') fname = os.path.join(dest, 'MRI scaling parameters.cfg') _write_mri_config(fname, subject_from, subject_to, scale) logger.debug('surf files [in mm]') for fname in paths['surf']: src = fname.format(subject=subject_from, subjects_dir=subjects_dir) src = os.path.realpath(src) dest = fname.format(subject=subject_to, subjects_dir=subjects_dir) pts, tri = read_surface(src) write_surface(dest, pts * scale, tri) logger.debug('BEM files [in m]') for bem_name in paths['bem']: scale_bem(subject_to, bem_name, subject_from, scale, subjects_dir, verbose=False) logger.debug('fiducials [in m]') for fname in paths['fid']: src = fname.format(subject=subject_from, subjects_dir=subjects_dir) src = os.path.realpath(src) pts, cframe = read_fiducials(src, verbose=False) for pt in pts: pt['r'] = pt['r'] * scale dest = fname.format(subject=subject_to, subjects_dir=subjects_dir) write_fiducials(dest, pts, cframe, verbose=False) logger.debug('MRIs [nibabel]') os.mkdir(mri_dirname.format(subjects_dir=subjects_dir, subject=subject_to)) for fname in paths['mri']: mri_name = os.path.basename(fname) _scale_mri(subject_to, mri_name, subject_from, scale, subjects_dir) logger.debug('Transforms') for mri_name in paths['mri']: if mri_name.endswith('T1.mgz'): os.mkdir(mri_transforms_dirname.format(subjects_dir=subjects_dir, subject=subject_to)) for fname in paths['transforms']: xfm_name = os.path.basename(fname) _scale_xfm(subject_to, xfm_name, mri_name, subject_from, scale, subjects_dir) break logger.debug('duplicate files') for fname in paths['duplicate']: src = fname.format(subject=subject_from, subjects_dir=subjects_dir) dest = fname.format(subject=subject_to, subjects_dir=subjects_dir) shutil.copyfile(src, dest) logger.debug('source spaces') for fname in paths['src']: src_name = os.path.basename(fname) scale_source_space(subject_to, src_name, subject_from, scale, subjects_dir, verbose=False) logger.debug('labels [in m]') os.mkdir(os.path.join(subjects_dir, subject_to, 'label')) if labels: scale_labels(subject_to, subject_from=subject_from, scale=scale, subjects_dir=subjects_dir) logger.debug('copy *.annot files') # they don't contain scale-dependent information if annot: src_pattern = os.path.join(subjects_dir, subject_from, 'label', '*.annot') dst_dir = os.path.join(subjects_dir, subject_to, 'label') for src_file in iglob(src_pattern): shutil.copy(src_file, dst_dir) @verbose def scale_source_space(subject_to, src_name, subject_from=None, scale=None, subjects_dir=None, n_jobs=1, verbose=None): """Scale a source space for an mri created with scale_mri(). Parameters ---------- subject_to : str Name of the scaled MRI subject (the destination mri subject). src_name : str Source space name. Can be a spacing parameter (e.g., ``'7'``, ``'ico4'``, ``'oct6'``) or a file name of a source space file relative to the bem directory; if the file name contains the subject name, it should be indicated as "{subject}" in ``src_name`` (e.g., ``"{subject}-my_source_space-src.fif"``). subject_from : None | str The subject from which to read the source space. If None, subject_from is read from subject_to's config file. scale : None | float | array, shape = (3,) Scaling factor. Has to be specified if subjects_from is specified, otherwise it is read from subject_to's config file. subjects_dir : None | str Override the SUBJECTS_DIR environment variable. n_jobs : int Number of jobs to run in parallel if recomputing distances (only applies if scale is an array of length 3, and will not use more cores than there are source spaces). %(verbose)s Notes ----- When scaling volume source spaces, the source (vertex) locations are scaled, but the reference to the MRI volume is left unchanged. Transforms are updated so that source estimates can be plotted on the original MRI volume. """ subjects_dir, subject_from, scale, uniform = \ _scale_params(subject_to, subject_from, scale, subjects_dir) # if n_params==1 scale is a scalar; if n_params==3 scale is a (3,) array # find the source space file names if src_name.isdigit(): spacing = src_name # spacing in mm src_pattern = src_fname else: match = re.match(r"(oct|ico|vol)-?(\d+)$", src_name) if match: spacing = '-'.join(match.groups()) src_pattern = src_fname else: spacing = None src_pattern = os.path.join(bem_dirname, src_name) src = src_pattern.format(subjects_dir=subjects_dir, subject=subject_from, spacing=spacing) dst = src_pattern.format(subjects_dir=subjects_dir, subject=subject_to, spacing=spacing) # read and scale the source space [in m] sss = read_source_spaces(src) logger.info("scaling source space %s: %s -> %s", spacing, subject_from, subject_to) logger.info("Scale factor: %s", scale) add_dist = False for ss in sss: ss['subject_his_id'] = subject_to ss['rr'] *= scale # additional tags for volume source spaces if 'vox_mri_t' in ss: # maintain transform to original MRI volume ss['mri_volume_name'] ss['vox_mri_t']['trans'][:3, :3] /= scale ss['src_mri_t']['trans'][:3, :3] /= scale # distances and patch info if uniform: if ss['dist'] is not None: ss['dist'] *= scale[0] # Sometimes this is read-only due to how it's read ss['nearest_dist'] = ss['nearest_dist'] * scale ss['dist_limit'] = ss['dist_limit'] * scale else: # non-uniform scaling ss['nn'] /= scale _normalize_vectors(ss['nn']) if ss['dist'] is not None: add_dist = True if add_dist: logger.info("Recomputing distances, this might take a while") dist_limit = float(np.abs(sss[0]['dist_limit'])) add_source_space_distances(sss, dist_limit, n_jobs) write_source_spaces(dst, sss) def _scale_mri(subject_to, mri_fname, subject_from, scale, subjects_dir): """Scale an MRI by setting its affine.""" subjects_dir, subject_from, scale, _ = _scale_params( subject_to, subject_from, scale, subjects_dir) if not has_nibabel(): warn('Skipping MRI scaling for %s, please install nibabel') return import nibabel fname_from = op.join(mri_dirname.format( subjects_dir=subjects_dir, subject=subject_from), mri_fname) fname_to = op.join(mri_dirname.format( subjects_dir=subjects_dir, subject=subject_to), mri_fname) img = nibabel.load(fname_from) zooms = np.array(img.header.get_zooms()) zooms[[0, 2, 1]] *= scale img.header.set_zooms(zooms) # Hack to fix nibabel problems, see # https://github.com/nipy/nibabel/issues/619 img._affine = img.header.get_affine() # or could use None nibabel.save(img, fname_to) def _scale_xfm(subject_to, xfm_fname, mri_name, subject_from, scale, subjects_dir): """Scale a transform.""" subjects_dir, subject_from, scale, _ = _scale_params( subject_to, subject_from, scale, subjects_dir) # The nibabel warning should already be there in MRI step, if applicable, # as we only get here if T1.mgz is present (and thus a scaling was # attempted) so we can silently return here. if not has_nibabel(): return fname_from = os.path.join( mri_transforms_dirname.format( subjects_dir=subjects_dir, subject=subject_from), xfm_fname) fname_to = op.join( mri_transforms_dirname.format( subjects_dir=subjects_dir, subject=subject_to), xfm_fname) assert op.isfile(fname_from), fname_from assert op.isdir(op.dirname(fname_to)), op.dirname(fname_to) # The "talairach.xfm" file stores the ras_mni transform. # # For "from" subj F, "to" subj T, F->T scaling S, some equivalent vertex # positions F_x and T_x in MRI (Freesurfer RAS) coords, knowing that # we have T_x = S @ F_x, we want to have the same MNI coords computed # for these vertices: # # T_mri_mni @ T_x = F_mri_mni @ F_x # # We need to find the correct T_ras_mni (talaraich.xfm file) that yields # this. So we derive (where † indicates inversion): # # T_mri_mni @ S @ F_x = F_mri_mni @ F_x # T_mri_mni @ S = F_mri_mni # T_ras_mni @ T_mri_ras @ S = F_ras_mni @ F_mri_ras # T_ras_mni @ T_mri_ras = F_ras_mni @ F_mri_ras @ S⁻¹ # T_ras_mni = F_ras_mni @ F_mri_ras @ S⁻¹ @ T_ras_mri # # prepare the scale (S) transform scale = np.atleast_1d(scale) scale = np.tile(scale, 3) if len(scale) == 1 else scale S = Transform('mri', 'mri', scaling(*scale)) # F_mri->T_mri # # Get the necessary transforms of the "from" subject # xfm, kind = _read_fs_xfm(fname_from) assert kind == 'MNI Transform File', kind F_ras_mni = Transform('ras', 'mni_tal', xfm) hdr = _get_mri_header(mri_name) F_vox_ras = Transform('mri_voxel', 'ras', hdr.get_vox2ras()) F_vox_mri = Transform('mri_voxel', 'mri', hdr.get_vox2ras_tkr()) F_mri_ras = combine_transforms( invert_transform(F_vox_mri), F_vox_ras, 'mri', 'ras') del F_vox_ras, F_vox_mri, hdr, xfm # # Get the necessary transforms of the "to" subject # mri_name = op.join(mri_dirname.format( subjects_dir=subjects_dir, subject=subject_to), op.basename(mri_name)) hdr = _get_mri_header(mri_name) T_vox_ras = Transform('mri_voxel', 'ras', hdr.get_vox2ras()) T_vox_mri = Transform('mri_voxel', 'mri', hdr.get_vox2ras_tkr()) T_ras_mri = combine_transforms( invert_transform(T_vox_ras), T_vox_mri, 'ras', 'mri') del mri_name, hdr, T_vox_ras, T_vox_mri # Finally we construct as above: # # T_ras_mni = F_ras_mni @ F_mri_ras @ S⁻¹ @ T_ras_mri # # By moving right to left through the equation. T_ras_mni = \ combine_transforms( combine_transforms( combine_transforms( T_ras_mri, invert_transform(S), 'ras', 'mri'), F_mri_ras, 'ras', 'ras'), F_ras_mni, 'ras', 'mni_tal') _write_fs_xfm(fname_to, T_ras_mni['trans'], kind) @verbose def get_mni_fiducials(subject, subjects_dir=None, verbose=None): """Estimate fiducials for a subject. Parameters ---------- subject : str Name of the mri subject subjects_dir : None | str Override the SUBJECTS_DIR environment variable (sys.environ['SUBJECTS_DIR']) %(verbose)s Returns ------- fids_mri : list List of estimated fiducials (each point in a dict) Notes ----- This takes the ``fsaverage-fiducials.fif`` file included with MNE—which contain the LPA, nasion, and RPA for the ``fsaverage`` subject—and transforms them to the given FreeSurfer subject's MRI space. The MRI of ``fsaverage`` is already in MNI Talairach space, so applying the inverse of the given subject's MNI Talairach affine transformation (``$SUBJECTS_DIR/$SUBJECT/mri/transforms/talairach.xfm``) is used to estimate the subject's fiducial locations. For more details about the coordinate systems and transformations involved, see https://surfer.nmr.mgh.harvard.edu/fswiki/CoordinateSystems and :ref:`plot_source_alignment`. """ # Eventually we might want to allow using the MNI Talairach with-skull # transformation rather than the standard brain-based MNI Talaranch # transformation, and/or project the points onto the head surface # (if available). fname_fids_fs = os.path.join(os.path.dirname(__file__), 'data', 'fsaverage', 'fsaverage-fiducials.fif') # Read fsaverage fiducials file and subject Talairach. fids, coord_frame = read_fiducials(fname_fids_fs) assert coord_frame == FIFF.FIFFV_COORD_MRI if subject == 'fsaverage': return fids # special short-circuit for fsaverage mni_mri_t = invert_transform(_read_talxfm(subject, subjects_dir)) # Convert to mm since this is Freesurfer's unit. lnr = np.array([f['r'] for f in fids]) * 1000. assert lnr.shape == (3, 3) # Apply transformation, to fsaverage (MNI) fiducials, convert back to m lnr = apply_trans(mni_mri_t, lnr) / 1000. for ii in range(3): fids[ii]['r'] = lnr[ii] return fids
37.822415
79
0.626009
794d8184bd8da1bd0ecf8ac3790f2f61d837a69f
796
py
Python
apps/rss_feeds/management/commands/task_feeds.py
louis-pre/NewsBlur
b4e9a56041ff187ef77b38dfd0778daf41b53f4f
[ "MIT" ]
3,073
2015-01-01T07:20:18.000Z
2022-03-31T20:33:41.000Z
apps/rss_feeds/management/commands/task_feeds.py
louis-pre/NewsBlur
b4e9a56041ff187ef77b38dfd0778daf41b53f4f
[ "MIT" ]
1,054
2015-01-02T13:32:35.000Z
2022-03-30T04:21:21.000Z
apps/rss_feeds/management/commands/task_feeds.py
louis-pre/NewsBlur
b4e9a56041ff187ef77b38dfd0778daf41b53f4f
[ "MIT" ]
676
2015-01-03T16:40:29.000Z
2022-03-30T14:00:40.000Z
from django.core.management.base import BaseCommand from django.conf import settings from apps.rss_feeds.tasks import TaskFeeds, TaskBrokenFeeds import datetime class Command(BaseCommand): def add_arguments(self, parser): parser.add_argument("-f", "--feed", default=None) parser.add_argument("-a", "--all", default=False, action='store_true') parser.add_argument("-b", "--broken", help="Task broken feeds that havent been fetched in a day.", default=False, action='store_true') parser.add_argument('-V', '--verbose', action='store_true', dest='verbose', default=False, help='Verbose output.') def handle(self, *args, **options): if options['broken']: TaskBrokenFeeds.apply() else: TaskFeeds.apply()
37.904762
142
0.665829
794d81e249626c1380ce00397bd0a43a46cd9e5b
3,116
py
Python
tests/testvcfparser/test_meta_header_parser.py
everestial/vcfparser
7e473d03e8ca9d040d2c5c310ce0febb07d874b2
[ "MIT" ]
2
2019-09-25T08:25:45.000Z
2019-12-25T16:24:00.000Z
tests/testvcfparser/test_meta_header_parser.py
everestial/vcfparser
7e473d03e8ca9d040d2c5c310ce0febb07d874b2
[ "MIT" ]
null
null
null
tests/testvcfparser/test_meta_header_parser.py
everestial/vcfparser
7e473d03e8ca9d040d2c5c310ce0febb07d874b2
[ "MIT" ]
null
null
null
import pytest import filecmp import os import collections import sys import json import csv from vcfparser import MetaDataParser from vcfparser.meta_header_parser import split_to_dict header_file = "tests/testfiles/vcf_parser_input/reference_input_test.vcf" meta_header_parser_obj = MetaDataParser(header_file) # 1 MetaDataParser._parse_gvcf_block def test_parse_gvcf_block(): pass # line = "" # data = meta_header_parser_obj._parse_gvcf_block(line) # write_data_and_check(1, data) # 2 MetaDataParser._parse_gatk_commands def test_parse_gatk_commands(): pass # line = "" # data = meta_header_parser_obj._parse_gatk_commands(line) # write_data_and_check(2, data) # 3 MetaDataParser.parse_lines def test_parse_lines(): pass # data = meta_header_parser_obj.parse_lines() # write_data_and_check(3, data) ##TODO:Bishwa - Problem in creating error file. # def test_parse_lines_exit1(): # fileformat_error = "tests/testfiles/vcf_parser_input/reference_input_test_fileformat_check.vcf" # meta_header_parser_obj2 = MetaDataParser(fileformat_error) # with pytest.raises(IndexError): # meta_header_parser_obj1.parse_lines() ##TODO:Bishwa - Problem in creating error file. # def test_parse_lines_exit2(): # reference_error = "tests/testfiles/vcf_parser_input/reference_input_test_reference_check.vcf" # meta_header_parser_obj2 = MetaDataParser(reference_error) # with pytest.raises(IndexError): # meta_header_parser_obj2.parse_lines() ##TODO:Bishwa - Problem in creating error file. # def test_parse_lines_exit3(): # format_error = "tests/testfiles/vcf_parser_input/reference_input_test_format_check.vcf" # meta_header_parser_obj3 = MetaDataParser(format_error) # with pytest.raises(IndexError): # meta_header_parser_obj3.parse_lines() ##TODO:Bishwa - Problem in creating error file. # def test_parse_lines_exit4(): # error = "tests/testfiles/vcf_parser_input/reference_input_test_error.vcf" # meta_header_parser_obj4 = MetaDataParser(error) # with pytest.raises(SyntaxError): # meta_header_parser_obj4.parse_lines() # 4 split_to_dict def test_split_to_dict(): # pass string = "<ID=HQ,Number=2,Type=Integer,Description='Haplotype Quality'>" data = split_to_dict(string) write_data_and_check(4, data) # Helping functions def write_data_and_check(i, data): output_file = f"tests/testfiles/vcf_parser_output/meta_header_parser_test_result{i}" with open(output_file + ".txt", "w") as w_file: for key, val in data.items(): w_file.write(f"{key}\n") w_file.write(f"{val}\n") w_file.write("\n") assert ( is_same_file( f"tests/testfiles/vcf_parser_output/meta_header_parser_test_result{i}.txt", f"tests/testfiles/vcf_parser_reference/meta_header_parser_test_ref{i}.txt", ) is True ) os.remove( f"tests/testfiles/vcf_parser_output/meta_header_parser_test_result{i}.txt" ) # Compare files def is_same_file(file1, file2): return filecmp.cmp(file1, file2)
32.123711
101
0.73973
794d8210f82d9ef750367fa280b3c0ac40361c74
7,336
py
Python
benchmarks/f3_wrong_hints/scaling_software_termination/15-2Nested_false-termination_4.py
EnricoMagnago/F3
c863215c318d7d5f258eb9be38c6962cf6863b52
[ "MIT" ]
3
2021-04-23T23:29:26.000Z
2022-03-23T10:00:30.000Z
benchmarks/f3_wrong_hints/scaling_software_termination/15-2Nested_false-termination_4.py
EnricoMagnago/F3
c863215c318d7d5f258eb9be38c6962cf6863b52
[ "MIT" ]
null
null
null
benchmarks/f3_wrong_hints/scaling_software_termination/15-2Nested_false-termination_4.py
EnricoMagnago/F3
c863215c318d7d5f258eb9be38c6962cf6863b52
[ "MIT" ]
1
2021-11-17T22:02:56.000Z
2021-11-17T22:02:56.000Z
from typing import Tuple, FrozenSet from pysmt.environment import Environment as PysmtEnv from pysmt.fnode import FNode import pysmt.typing as types from utils import symb_to_next from hint import Hint, Location def transition_system(env: PysmtEnv) -> Tuple[FrozenSet[FNode], FNode, FNode, FNode]: assert isinstance(env, PysmtEnv) mgr = env.formula_manager pc = mgr.Symbol("pc", types.INT) x = mgr.Symbol("x", types.INT) y = mgr.Symbol("y", types.INT) x_pc = symb_to_next(mgr, pc) x_x = symb_to_next(mgr, x) x_y = symb_to_next(mgr, y) symbols = frozenset([pc, x, y]) m_1 = mgr.Int(-1) n_locs = 3 max_int = n_locs ints = [] pcs = [] x_pcs = [] for idx in range(n_locs): num = mgr.Int(idx) ints.append(num) pcs.append(mgr.Equals(pc, num)) x_pcs.append(mgr.Equals(x_pc, num)) for idx in range(n_locs, max_int): num = mgr.Int(idx) ints.append(num) pcend = mgr.Equals(pc, m_1) x_pcend = mgr.Equals(x_pc, m_1) init = pcs[0] cfg = [] # pc = 0 & (x >= 0) -> pc' = 1 cond = mgr.GE(x, ints[0]) cfg.append(mgr.Implies(mgr.And(pcs[0], cond), x_pcs[1])) # pc = 0 & !(x >= 0) -> pc' = -1 cfg.append(mgr.Implies(mgr.And(pcs[0], mgr.Not(cond)), x_pcend)) # pc = 1 -> pc' = 2 cfg.append(mgr.Implies(pcs[1], x_pcs[2])) # pc = 2 -> pc' = 0 cfg.append(mgr.Implies(pcs[2], x_pcs[0])) # pc = -1 -> pc' = -1 cfg.append(mgr.Implies(pcend, x_pcend)) trans = [] same_x = mgr.Equals(x_x, x) same_y = mgr.Equals(x_y, y) same = mgr.And(same_x, same_y) # pc = 0 -> same trans.append(mgr.Implies(pcs[0], same)) # pc = 1 -> x' = x + y & same_y trans.append(mgr.Implies(pcs[1], mgr.And(mgr.Equals(x_x, mgr.Plus(x, y)), same_y))) # pc = 2 -> same_x & y' = y + 1 trans.append(mgr.Implies(pcs[2], mgr.And(same_x, mgr.Equals(x_y, mgr.Plus(y, ints[1]))))) # pc = end -> same trans.append(mgr.Implies(pcend, same)) trans = mgr.And(*cfg, *trans) fairness = mgr.Not(mgr.Equals(pc, m_1)) return symbols, init, trans, fairness def hints(env: PysmtEnv) -> FrozenSet[Hint]: assert isinstance(env, PysmtEnv) mgr = env.formula_manager pc = mgr.Symbol("pc", types.INT) x = mgr.Symbol("x", types.INT) y = mgr.Symbol("y", types.INT) symbs = frozenset([pc, x, y]) m_100 = mgr.Int(-100) m_1 = mgr.Int(-1) i_0 = mgr.Int(0) i_1 = mgr.Int(1) i_2 = mgr.Int(2) i_4 = mgr.Int(4) i_20 = mgr.Int(20) x_pc = symb_to_next(mgr, pc) x_x = symb_to_next(mgr, x) x_y = symb_to_next(mgr, y) res = [] stutter = mgr.Equals(x_y, y) loc = Location(env, mgr.TRUE(), mgr.LE(x, i_20), stutterT=stutter) loc.set_progress(0, mgr.Equals(x_y, mgr.Plus(i_1, y))) h_y = Hint("h_y0", env, frozenset([y]), symbs) h_y.set_locs([loc]) res.append(h_y) stutter = mgr.Equals(x_x, x) loc = Location(env, mgr.GE(x, i_20), mgr.GE(y, i_1), stutterT=stutter) loc.set_progress(0, mgr.Equals(x_x, mgr.Plus(x, y))) h_x = Hint("h_x0", env, frozenset([x]), symbs) h_x.set_locs([loc]) res.append(h_x) stutter = mgr.Equals(x_y, y) loc = Location(env, mgr.TRUE(), mgr.LE(x, i_20), stutterT=stutter) loc.set_progress(0, mgr.Equals(x_y, mgr.Plus(x, y))) h_y = Hint("h_y1", env, frozenset([y]), symbs) h_y.set_locs([loc]) res.append(h_y) stutter = mgr.Equals(x_x, x) loc = Location(env, mgr.GE(x, i_1), mgr.GE(y, i_1), stutterT=stutter) loc.set_progress(0, mgr.Equals(x_x, mgr.Plus(x, y))) h_x = Hint("h_x1", env, frozenset([x]), symbs) h_x.set_locs([loc]) res.append(h_x) loc0 = Location(env, mgr.GE(y, m_100), mgr.LE(x, i_20)) loc0.set_progress(1, mgr.Equals(x_y, mgr.Plus(x, y))) loc1 = Location(env, mgr.TRUE(), mgr.GE(x, m_100)) loc1.set_progress(0, mgr.Equals(x_y, m_100)) h_y = Hint("h_y2", env, frozenset([y]), symbs) h_y.set_locs([loc0, loc1]) res.append(h_y) loc0 = Location(env, mgr.GE(y, m_100), mgr.LE(x, i_20)) loc0.set_progress(1, mgr.Equals(x_y, mgr.Times(x, y))) loc1 = Location(env, mgr.TRUE(), mgr.GE(x, m_100)) loc1.set_progress(0, mgr.Equals(x_y, m_100)) h_y = Hint("h_y3", env, frozenset([y]), symbs) h_y.set_locs([loc0, loc1]) res.append(h_y) loc0 = Location(env, mgr.GE(x, i_1), mgr.GE(y, i_1)) loc0.set_progress(1, mgr.Equals(x_x, mgr.Times(x, y))) loc1 = Location(env, mgr.GE(x, i_1), mgr.GE(y, i_1)) loc1.set_progress(0, mgr.Equals(x_x, y)) h_x = Hint("h_x3", env, frozenset([x]), symbs) h_x.set_locs([loc0, loc1]) res.append(h_x) loc0 = Location(env, mgr.GE(x, i_1), mgr.GE(y, i_1)) loc0.set_progress(1, mgr.Equals(x_x, mgr.Times(x, y))) loc1 = Location(env, mgr.GE(x, i_1), mgr.GE(y, i_1)) loc1.set_progress(2, mgr.GT(x_x, y)) loc2 = Location(env, mgr.GE(x, i_2)) loc2.set_progress(0, mgr.GE(x_x, i_20)) h_x = Hint("h_x4", env, frozenset([x]), symbs) h_x.set_locs([loc0, loc1, loc2]) res.append(h_x) loc0 = Location(env, mgr.TRUE()) loc0.set_progress(0, mgr.TRUE()) h_pc = Hint("h_pc1", env, frozenset([pc]), symbs) h_pc.set_locs([loc0]) res.append(h_pc) loc0 = Location(env, mgr.GE(y, m_100)) loc0.set_progress(0, mgr.Equals(x_y, mgr.Times(y, y))) h_y = Hint("h_y5", env, frozenset([y]), symbs) h_y.set_locs([loc0]) res.append(h_y) loc0 = Location(env, mgr.LE(x, i_0)) loc0.set_progress(1, mgr.Equals(x_x, mgr.Times(x, x))) loc1 = Location(env, mgr.GE(x, i_0)) loc1.set_progress(0, mgr.LT(x_x, mgr.Times(m_1, x, x))) h_x = Hint("h_x5", env, frozenset([x]), symbs) h_x.set_locs([loc0, loc1]) res.append(h_x) loc0 = Location(env, mgr.GE(y, m_100)) loc0.set_progress(1, mgr.Equals(x_y, mgr.Times(y, y))) loc1 = Location(env, mgr.GE(y, i_0)) loc1.set_progress(0, mgr.GE(x_y, mgr.Plus(y, i_1))) h_y = Hint("h_y6", env, frozenset([y]), symbs) h_y.set_locs([loc0, loc1]) res.append(h_y) loc0 = Location(env, mgr.LE(x, i_20)) loc0.set_progress(1, mgr.Equals(x_x, mgr.Plus(mgr.Times(x, x), i_1))) loc1 = Location(env, mgr.GE(x, i_20)) loc1.set_progress(0, mgr.LT(x_x, mgr.Times(m_1, x, x))) h_x = Hint("h_x6", env, frozenset([x]), symbs) h_x.set_locs([loc0, loc1]) res.append(h_x) loc0 = Location(env, mgr.GE(y, i_0), mgr.GE(pc, i_1)) loc0.set_progress(1, mgr.Equals(x_y, mgr.Plus(y, pc))) loc1 = Location(env, mgr.GE(y, i_1)) loc1.set_progress(0, mgr.Equals(x_y, y)) h_y = Hint("h_y7", env, frozenset([y]), symbs) h_y.set_locs([loc0, loc1]) res.append(h_y) loc0 = Location(env, mgr.GE(x, i_1), mgr.GT(y, i_1)) loc0.set_progress(1, mgr.Equals(x_x, mgr.Plus(mgr.Times(x, y), i_1))) loc1 = Location(env, mgr.GE(x, i_2)) loc1.set_progress(2, mgr.LT(x_x, mgr.Times(m_1, x, x))) loc2 = Location(env, mgr.LE(x, i_4)) loc2.set_progress(0, mgr.GE(x_x, mgr.Div(x, x))) h_x = Hint("h_x7", env, frozenset([x]), symbs) h_x.set_locs([loc0, loc1, loc2]) res.append(h_x) return frozenset(res)
30.953586
77
0.585742
794d821fd149c82d8cbe4a56d6b0cf2150d74506
1,514
py
Python
openapi-python-client/test/test_schema_log_query_dto_sorting.py
yanavasileva/camunda-bpm-examples
051f8f28c62845e68ce4059ab64264c5a0bdc009
[ "Apache-2.0" ]
null
null
null
openapi-python-client/test/test_schema_log_query_dto_sorting.py
yanavasileva/camunda-bpm-examples
051f8f28c62845e68ce4059ab64264c5a0bdc009
[ "Apache-2.0" ]
null
null
null
openapi-python-client/test/test_schema_log_query_dto_sorting.py
yanavasileva/camunda-bpm-examples
051f8f28c62845e68ce4059ab64264c5a0bdc009
[ "Apache-2.0" ]
null
null
null
# coding: utf-8 """ Camunda BPM REST API OpenApi Spec for Camunda BPM REST API. # noqa: E501 The version of the OpenAPI document: 7.13.0 Generated by: https://openapi-generator.tech """ from __future__ import absolute_import import unittest import datetime import openapi_client from openapi_client.models.schema_log_query_dto_sorting import SchemaLogQueryDtoSorting # noqa: E501 from openapi_client.rest import ApiException class TestSchemaLogQueryDtoSorting(unittest.TestCase): """SchemaLogQueryDtoSorting unit test stubs""" def setUp(self): pass def tearDown(self): pass def make_instance(self, include_optional): """Test SchemaLogQueryDtoSorting include_option is a boolean, when False only required params are included, when True both required and optional params are included """ # model = openapi_client.models.schema_log_query_dto_sorting.SchemaLogQueryDtoSorting() # noqa: E501 if include_optional : return SchemaLogQueryDtoSorting( sort_by = 'timestamp', sort_order = 'asc' ) else : return SchemaLogQueryDtoSorting( ) def testSchemaLogQueryDtoSorting(self): """Test SchemaLogQueryDtoSorting""" inst_req_only = self.make_instance(include_optional=False) inst_req_and_optional = self.make_instance(include_optional=True) if __name__ == '__main__': unittest.main()
28.037037
109
0.688243
794d85317dd87f17f00143f49a2d99bbf3347b2c
12,659
py
Python
XBlock Integration Files/xdjangobb/xblock/lib/python2.7/site-packages/mako/lookup.py
DANCEcollaborative/forum-xblock
fe8a062b4e45966a5faa5282d85799be479ec28a
[ "MIT" ]
7
2016-02-09T06:43:14.000Z
2021-03-12T06:07:50.000Z
XBlock Integration Files/xdjangobb/xblock/lib/python2.7/site-packages/mako/lookup.py
DANCEcollaborative/forum-xblock
fe8a062b4e45966a5faa5282d85799be479ec28a
[ "MIT" ]
null
null
null
XBlock Integration Files/xdjangobb/xblock/lib/python2.7/site-packages/mako/lookup.py
DANCEcollaborative/forum-xblock
fe8a062b4e45966a5faa5282d85799be479ec28a
[ "MIT" ]
null
null
null
# mako/lookup.py # Copyright (C) 2006-2015 the Mako authors and contributors <see AUTHORS file> # # This module is part of Mako and is released under # the MIT License: http://www.opensource.org/licenses/mit-license.php import os import stat import posixpath import re from mako import exceptions, util from mako.template import Template try: import threading except: import dummy_threading as threading class TemplateCollection(object): """Represent a collection of :class:`.Template` objects, identifiable via URI. A :class:`.TemplateCollection` is linked to the usage of all template tags that address other templates, such as ``<%include>``, ``<%namespace>``, and ``<%inherit>``. The ``file`` attribute of each of those tags refers to a string URI that is passed to that :class:`.Template` object's :class:`.TemplateCollection` for resolution. :class:`.TemplateCollection` is an abstract class, with the usual default implementation being :class:`.TemplateLookup`. """ def has_template(self, uri): """Return ``True`` if this :class:`.TemplateLookup` is capable of returning a :class:`.Template` object for the given ``uri``. :param uri: String URI of the template to be resolved. """ try: self.get_template(uri) return True except exceptions.TemplateLookupException: return False def get_template(self, uri, relativeto=None): """Return a :class:`.Template` object corresponding to the given ``uri``. The default implementation raises :class:`.NotImplementedError`. Implementations should raise :class:`.TemplateLookupException` if the given ``uri`` cannot be resolved. :param uri: String URI of the template to be resolved. :param relativeto: if present, the given ``uri`` is assumed to be relative to this URI. """ raise NotImplementedError() def filename_to_uri(self, uri, filename): """Convert the given ``filename`` to a URI relative to this :class:`.TemplateCollection`.""" return uri def adjust_uri(self, uri, filename): """Adjust the given ``uri`` based on the calling ``filename``. When this method is called from the runtime, the ``filename`` parameter is taken directly to the ``filename`` attribute of the calling template. Therefore a custom :class:`.TemplateCollection` subclass can place any string identifier desired in the ``filename`` parameter of the :class:`.Template` objects it constructs and have them come back here. """ return uri class TemplateLookup(TemplateCollection): """Represent a collection of templates that locates template source files from the local filesystem. The primary argument is the ``directories`` argument, the list of directories to search: .. sourcecode:: python lookup = TemplateLookup(["/path/to/templates"]) some_template = lookup.get_template("/index.html") The :class:`.TemplateLookup` can also be given :class:`.Template` objects programatically using :meth:`.put_string` or :meth:`.put_template`: .. sourcecode:: python lookup = TemplateLookup() lookup.put_string("base.html", ''' <html><body>${self.next()}</body></html> ''') lookup.put_string("hello.html", ''' <%include file='base.html'/> Hello, world ! ''') :param directories: A list of directory names which will be searched for a particular template URI. The URI is appended to each directory and the filesystem checked. :param collection_size: Approximate size of the collection used to store templates. If left at its default of ``-1``, the size is unbounded, and a plain Python dictionary is used to relate URI strings to :class:`.Template` instances. Otherwise, a least-recently-used cache object is used which will maintain the size of the collection approximately to the number given. :param filesystem_checks: When at its default value of ``True``, each call to :meth:`.TemplateLookup.get_template()` will compare the filesystem last modified time to the time in which an existing :class:`.Template` object was created. This allows the :class:`.TemplateLookup` to regenerate a new :class:`.Template` whenever the original source has been updated. Set this to ``False`` for a very minor performance increase. :param modulename_callable: A callable which, when present, is passed the path of the source file as well as the requested URI, and then returns the full path of the generated Python module file. This is used to inject alternate schemes for Python module location. If left at its default of ``None``, the built in system of generation based on ``module_directory`` plus ``uri`` is used. All other keyword parameters available for :class:`.Template` are mirrored here. When new :class:`.Template` objects are created, the keywords established with this :class:`.TemplateLookup` are passed on to each new :class:`.Template`. """ def __init__(self, directories=None, module_directory=None, filesystem_checks=True, collection_size=-1, format_exceptions=False, error_handler=None, disable_unicode=False, bytestring_passthrough=False, output_encoding=None, encoding_errors='strict', cache_args=None, cache_impl='beaker', cache_enabled=True, cache_type=None, cache_dir=None, cache_url=None, modulename_callable=None, module_writer=None, default_filters=None, buffer_filters=(), strict_undefined=False, imports=None, future_imports=None, enable_loop=True, input_encoding=None, preprocessor=None, lexer_cls=None): self.directories = [posixpath.normpath(d) for d in util.to_list(directories, ()) ] self.module_directory = module_directory self.modulename_callable = modulename_callable self.filesystem_checks = filesystem_checks self.collection_size = collection_size if cache_args is None: cache_args = {} # transfer deprecated cache_* args if cache_dir: cache_args.setdefault('dir', cache_dir) if cache_url: cache_args.setdefault('url', cache_url) if cache_type: cache_args.setdefault('type', cache_type) self.template_args = { 'format_exceptions': format_exceptions, 'error_handler': error_handler, 'disable_unicode': disable_unicode, 'bytestring_passthrough': bytestring_passthrough, 'output_encoding': output_encoding, 'cache_impl': cache_impl, 'encoding_errors': encoding_errors, 'input_encoding': input_encoding, 'module_directory': module_directory, 'module_writer': module_writer, 'cache_args': cache_args, 'cache_enabled': cache_enabled, 'default_filters': default_filters, 'buffer_filters': buffer_filters, 'strict_undefined': strict_undefined, 'imports': imports, 'future_imports': future_imports, 'enable_loop': enable_loop, 'preprocessor': preprocessor, 'lexer_cls': lexer_cls } if collection_size == -1: self._collection = {} self._uri_cache = {} else: self._collection = util.LRUCache(collection_size) self._uri_cache = util.LRUCache(collection_size) self._mutex = threading.Lock() def get_template(self, uri): """Return a :class:`.Template` object corresponding to the given ``uri``. .. note:: The ``relativeto`` argument is not supported here at the moment. """ try: if self.filesystem_checks: return self._check(uri, self._collection[uri]) else: return self._collection[uri] except KeyError: u = re.sub(r'^\/+', '', uri) for dir in self.directories: srcfile = posixpath.normpath(posixpath.join(dir, u)) if os.path.isfile(srcfile): return self._load(srcfile, uri) else: raise exceptions.TopLevelLookupException( "Cant locate template for uri %r" % uri) def adjust_uri(self, uri, relativeto): """Adjust the given ``uri`` based on the given relative URI.""" key = (uri, relativeto) if key in self._uri_cache: return self._uri_cache[key] if uri[0] != '/': if relativeto is not None: v = self._uri_cache[key] = posixpath.join( posixpath.dirname(relativeto), uri) else: v = self._uri_cache[key] = '/' + uri else: v = self._uri_cache[key] = uri return v def filename_to_uri(self, filename): """Convert the given ``filename`` to a URI relative to this :class:`.TemplateCollection`.""" try: return self._uri_cache[filename] except KeyError: value = self._relativeize(filename) self._uri_cache[filename] = value return value def _relativeize(self, filename): """Return the portion of a filename that is 'relative' to the directories in this lookup. """ filename = posixpath.normpath(filename) for dir in self.directories: if filename[0:len(dir)] == dir: return filename[len(dir):] else: return None def _load(self, filename, uri): self._mutex.acquire() try: try: # try returning from collection one # more time in case concurrent thread already loaded return self._collection[uri] except KeyError: pass try: if self.modulename_callable is not None: module_filename = self.modulename_callable(filename, uri) else: module_filename = None self._collection[uri] = template = Template( uri=uri, filename=posixpath.normpath(filename), lookup=self, module_filename=module_filename, **self.template_args) return template except: # if compilation fails etc, ensure # template is removed from collection, # re-raise self._collection.pop(uri, None) raise finally: self._mutex.release() def _check(self, uri, template): if template.filename is None: return template try: template_stat = os.stat(template.filename) if template.module._modified_time < \ template_stat[stat.ST_MTIME]: self._collection.pop(uri, None) return self._load(template.filename, uri) else: return template except OSError: self._collection.pop(uri, None) raise exceptions.TemplateLookupException( "Cant locate template for uri %r" % uri) def put_string(self, uri, text): """Place a new :class:`.Template` object into this :class:`.TemplateLookup`, based on the given string of ``text``. """ self._collection[uri] = Template( text, lookup=self, uri=uri, **self.template_args) def put_template(self, uri, template): """Place a new :class:`.Template` object into this :class:`.TemplateLookup`, based on the given :class:`.Template` object. """ self._collection[uri] = template
34.682192
78
0.590647
794d8579073943f2db41da21161142aa4a33d0fc
1,135
py
Python
Love Babbar DSA Sheet Solutions/Dynamic Programming/Knapsack/Knapsack.py
dr490n1s3d-3d8/Data-Structures-and-Algorithms
f2bba2d836f448cc0cd894c4e9791819910571d5
[ "MIT" ]
1
2022-03-06T13:58:27.000Z
2022-03-06T13:58:27.000Z
Love Babbar DSA Sheet Solutions/Dynamic Programming/Knapsack/Knapsack.py
kumar-tushar/Data-Structures-and-Algorithms
c586847734abbdd8f3cfbd76ba588216369f45df
[ "MIT" ]
null
null
null
Love Babbar DSA Sheet Solutions/Dynamic Programming/Knapsack/Knapsack.py
kumar-tushar/Data-Structures-and-Algorithms
c586847734abbdd8f3cfbd76ba588216369f45df
[ "MIT" ]
1
2022-02-12T03:58:50.000Z
2022-02-12T03:58:50.000Z
""" You are given weights and values of N items, put these items in a knapsack of capacity W to get the maximum total value in the knapsack. Note that we have only one quantity of each item. In other words, given two integer arrays val[0..N-1] and wt[0..N-1] which represent values and weights associated with N items respectively. Also given an integer W which represents knapsack capacity, find out the maximum value subset of val[] such that sum of the weights of this subset is smaller than or equal to W. You cannot break an item, either pick the complete item, or don’t pick it (0-1 property). Example 1: Input: N = 3 W = 4 values[] = {1,2,3} weight[] = {4,5,1} Output: 3 """ def knapSack(w, wt, val, n): dp = [[0 for i in range(w+1)]for i in range(n+1)] for i in range(1,n+1): for j in range(1,w+1): if j >= wt[i-1]: dp[i][j] = max(dp[i-1][j],dp[i-1][j-wt[i-1]] + val[i-1]) # if the j is greater than weight then u can pick the val or not pick the value else: dp[i][j] = dp[i-1][j] return dp[n][w] """ TC - O(N * W) SC - O(N * W) """
35.46875
117
0.622907
794d8632803c3546123cb595249c3343c6d48fba
4,766
py
Python
ppdet/modeling/transformers/matchers.py
leakyH/PaddleDetection
aa15eb945711baf248177a02d4d3dd3bd3abc4e8
[ "Apache-2.0" ]
3
2022-03-23T08:48:06.000Z
2022-03-28T01:59:34.000Z
ppdet/modeling/transformers/matchers.py
leakyH/PaddleDetection
aa15eb945711baf248177a02d4d3dd3bd3abc4e8
[ "Apache-2.0" ]
null
null
null
ppdet/modeling/transformers/matchers.py
leakyH/PaddleDetection
aa15eb945711baf248177a02d4d3dd3bd3abc4e8
[ "Apache-2.0" ]
1
2021-08-09T14:53:28.000Z
2021-08-09T14:53:28.000Z
# Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # Modified from DETR (https://github.com/facebookresearch/detr) # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved from __future__ import absolute_import from __future__ import division from __future__ import print_function import paddle import paddle.nn as nn import paddle.nn.functional as F from scipy.optimize import linear_sum_assignment from ppdet.core.workspace import register, serializable from ..losses.iou_loss import GIoULoss from .utils import bbox_cxcywh_to_xyxy __all__ = ['HungarianMatcher'] @register @serializable class HungarianMatcher(nn.Layer): __shared__ = ['use_focal_loss'] def __init__(self, matcher_coeff={'class': 1, 'bbox': 5, 'giou': 2}, use_focal_loss=False, alpha=0.25, gamma=2.0): r""" Args: matcher_coeff (dict): The coefficient of hungarian matcher cost. """ super(HungarianMatcher, self).__init__() self.matcher_coeff = matcher_coeff self.use_focal_loss = use_focal_loss self.alpha = alpha self.gamma = gamma self.giou_loss = GIoULoss() def forward(self, boxes, logits, gt_bbox, gt_class): r""" Args: boxes (Tensor): [b, query, 4] logits (Tensor): [b, query, num_classes] gt_bbox (List(Tensor)): list[[n, 4]] gt_class (List(Tensor)): list[[n, 1]] Returns: A list of size batch_size, containing tuples of (index_i, index_j) where: - index_i is the indices of the selected predictions (in order) - index_j is the indices of the corresponding selected targets (in order) For each batch element, it holds: len(index_i) = len(index_j) = min(num_queries, num_target_boxes) """ bs, num_queries = boxes.shape[:2] num_gts = sum(len(a) for a in gt_class) if num_gts == 0: return [(paddle.to_tensor( [], dtype=paddle.int64), paddle.to_tensor( [], dtype=paddle.int64)) for _ in range(bs)] # We flatten to compute the cost matrices in a batch # [batch_size * num_queries, num_classes] out_prob = F.sigmoid(logits.flatten( 0, 1)) if self.use_focal_loss else F.softmax(logits.flatten(0, 1)) # [batch_size * num_queries, 4] out_bbox = boxes.flatten(0, 1) # Also concat the target labels and boxes tgt_ids = paddle.concat(gt_class).flatten() tgt_bbox = paddle.concat(gt_bbox) # Compute the classification cost if self.use_focal_loss: neg_cost_class = (1 - self.alpha) * (out_prob**self.gamma) * (-( 1 - out_prob + 1e-8).log()) pos_cost_class = self.alpha * ( (1 - out_prob)**self.gamma) * (-(out_prob + 1e-8).log()) cost_class = paddle.gather( pos_cost_class, tgt_ids, axis=1) - paddle.gather( neg_cost_class, tgt_ids, axis=1) else: cost_class = -paddle.gather(out_prob, tgt_ids, axis=1) # Compute the L1 cost between boxes cost_bbox = ( out_bbox.unsqueeze(1) - tgt_bbox.unsqueeze(0)).abs().sum(-1) # Compute the giou cost betwen boxes cost_giou = self.giou_loss( bbox_cxcywh_to_xyxy(out_bbox.unsqueeze(1)), bbox_cxcywh_to_xyxy(tgt_bbox.unsqueeze(0))).squeeze(-1) # Final cost matrix C = self.matcher_coeff['class'] * cost_class + self.matcher_coeff['bbox'] * cost_bbox + \ self.matcher_coeff['giou'] * cost_giou C = C.reshape([bs, num_queries, -1]) C = [a.squeeze(0) for a in C.chunk(bs)] sizes = [a.shape[0] for a in gt_bbox] indices = [ linear_sum_assignment(c.split(sizes, -1)[i].numpy()) for i, c in enumerate(C) ] return [(paddle.to_tensor( i, dtype=paddle.int64), paddle.to_tensor( j, dtype=paddle.int64)) for i, j in indices]
37.527559
97
0.610365
794d876058bc8af52e6adeb6a6cce3e6e537a844
17,234
py
Python
egs/cnceleb/v1/local/choose_utts_to_combine.py
kjw11/kaldi
e87928818ead5ef046b58d1290f46bc8551b2fe6
[ "Apache-2.0" ]
9
2019-11-02T03:28:06.000Z
2019-12-06T07:32:11.000Z
egs/cnceleb/v1/local/choose_utts_to_combine.py
kjw11/kaldi
e87928818ead5ef046b58d1290f46bc8551b2fe6
[ "Apache-2.0" ]
null
null
null
egs/cnceleb/v1/local/choose_utts_to_combine.py
kjw11/kaldi
e87928818ead5ef046b58d1290f46bc8551b2fe6
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/env python # Copyright 2016 Vijayaditya Peddinti # 2016 Johns Hopkins University (author: Daniel Povey) # 2019 Tsinghua University (author: Jiawen Kang) # # Copied from utils/data/internal/choose_utts_to_combine.py # Apache 2.0 from __future__ import print_function import argparse from random import randint import sys import os from collections import defaultdict parser = argparse.ArgumentParser(description=""" This script, called from data/utils/combine_short_segments.sh, chooses consecutive utterances to concatenate that will satisfy the minimum segment length. It uses the --spk2utt file to ensure that utterances from the same speaker are preferentially combined (as far as possible while respecting the minimum segment length). If it has to combine utterances across different speakers in order to satisfy the duration constraint, it will assign the combined utterances to the speaker which contributed the most to the duration of the combined utterances. The utt2uts output of this program is a map from new utterance-id to a list of old utterance-ids, so for example if the inputs were utt1, utt2 and utt3, and utterances 2 and 3 were combined, the output might look like: utt1 utt1 utt2-combine2 utt2 utt3 The utt2spk output of this program assigns utterances to the speakers of the input; in the (hopefully rare) case where utterances were combined across speakers, it will assign the utterance to whichever of the original speakers contributed the most to the grouped utterance. """) parser.add_argument("--min-duration", type = float, default = 1.55, help="Minimum utterance duration") parser.add_argument("spk2utt_in", type = str, metavar = "<spk2utt-in>", help="Filename of [input] speaker to utterance map needed " "because this script tries to merge utterances from the " "same speaker as much as possible, and also needs to produce" "an output utt2spk map.") parser.add_argument("utt2dur_in", type = str, metavar = "<utt2dur-in>", help="Filename of [input] utterance-to-duration map, with lines like 'utt1 1.23'.") parser.add_argument("utt2utts_out", type = str, metavar = "<utt2utts-out>", help="Filename of [output] new-utterance-to-old-utterances map, with lines " "like 'utt1 utt1' or 'utt2-comb2 utt2 utt3'") parser.add_argument("utt2spk_out", type = str, metavar = "<utt2spk-out>", help="Filename of [output] utt2spk map, which maps new utterances to original " "speakers. If utterances were combined across speakers, we map the new " "utterance to the speaker that contributed the most to them.") parser.add_argument("utt2dur_out", type = str, metavar = "<utt2spk-out>", help="Filename of [output] utt2dur map, which is just the summations of " "the durations of the source utterances.") args = parser.parse_args() # This LessThan is designed to be impervious to roundoff effects in cases where # numbers are really always separated by a distance >> 1.0e-05. It will return # false if x and y are almost identical, differing only by roundoff effects. def LessThan(x, y): return x < y - 1.0e-5 # This function implements the core of the utterance-combination code. # The input 'durations' is a list of durations, which must all be # >=0.0 This function tries to combine consecutive indexes # into groups such that for each group, the total duration is at # least 'min_duration'. It returns a list of (start,end) indexes. # For example, CombineList(0.1, [5.0,6.0,7.0]) would return # [ (0,1), (1,2), (2,3) ] because no combination is necessary; each # returned pair represents a singleton group. # Or CombineList(1.0, [0.5, 0.6, 0.7]) would return # [ (0,3) ]. # Or CombineList(1.0, [0.5, 0.6, 1.7]) would return # [ (0,2), (2,3) ]. # Note: if sum(durations) < min_duration, this function will # return everything in one group but of course the sum of durations # will be less than the total. def CombineList(min_duration, durations): assert min_duration >= 0.0 and min(durations) > 0.0 num_utts = len(durations) # for each utterance-index i, group_start[i] gives us the # start-index of the group of utterances of which it's currently # a member. group_start = list(range(num_utts)) # if utterance-index i currently corresponds to the start of a group # of utterances, then group_durations[i] is the total duration of # that utterance-group, otherwise undefined. group_durations = list(durations) # if utterance-index i currently corresponds to the start of a group # of utterances, then group_end[i] is the end-index (i.e. last index plus one # of that utterance-group, otherwise undefined. group_end = [ x + 1 for x in range(num_utts) ] queue = [ i for i in range(num_utts) if LessThan(group_durations[i], min_duration) ] while len(queue) > 0: i = queue.pop() if group_start[i] != i or not LessThan(group_durations[i], min_duration): # this group no longer exists or already has at least the minimum duration. continue this_dur = group_durations[i] # left_dur is the duration of the group to the left of this group, # or 0.0 if there is no such group. left_dur = group_durations[group_start[i-1]] if i > 0 else 0.0 # right_dur is the duration of the group to the right of this group, # or 0.0 if there is no such group. right_dur = group_durations[group_end[i]] if group_end[i] < num_utts else 0.0 if left_dur == 0.0 and right_dur == 0.0: # there is only one group. Nothing more to merge; break assert group_start[i] == 0 and group_end[i] == num_utts break # work out whether to combine left or right, # by means of the combine_left variable [ True or False ] if left_dur == 0.0: combine_left = False elif right_dur == 0.0: combine_left = True elif LessThan(left_dur + this_dur, min_duration): # combining left would still be below the minimum duration-> # combine right... if it's above the min duration then good; # otherwise it still doesn't really matter so we might as well # pick one. combine_left = False elif LessThan(right_dur + this_dur, min_duration): # combining right would still be below the minimum duration, # and combining left would be >= the min duration (else we wouldn't # have reached this line) -> combine left. combine_left = True elif LessThan(left_dur, right_dur): # if we reached here then combining either way would take us >= the # minimum duration; but if left_dur < right_dur then we combine left # because that would give us more evenly sized segments. combine_left = True else: # if we reached here then combining either way would take us >= the # minimum duration; but left_dur >= right_dur, so we combine right # because that would give us more evenly sized segments. combine_left = False if combine_left: assert left_dur != 0.0 new_group_start = group_start[i-1] group_end[new_group_start] = group_end[i] for j in range(group_start[i], group_end[i]): group_start[j] = new_group_start group_durations[new_group_start] += durations[j] # note: there is no need to add group_durations[new_group_start] to # the queue even if it is still below the minimum length, because it # would have previously had to have been below the minimum length, # therefore it would already be in the queue. else: assert right_dur != 0.0 # group start doesn't change, group end changes. old_group_end = group_end[i] new_group_end = group_end[old_group_end] group_end[i] = new_group_end for j in range(old_group_end, new_group_end): group_durations[i] += durations[j] group_start[j] = i if LessThan(group_durations[i], min_duration): # the group starting at i is still below the minimum length, so # we need to put it back on the queue. queue.append(i) ans = [] cur_group_start = 0 while cur_group_start < num_utts: ans.append( (cur_group_start, group_end[cur_group_start]) ) cur_group_start = group_end[cur_group_start] return ans def SelfTest(): assert CombineList(0.1, [5.0, 6.0, 7.0]) == [ (0,1), (1,2), (2,3) ] assert CombineList(0.5, [0.1, 6.0, 7.0]) == [ (0,2), (2,3) ] assert CombineList(0.5, [6.0, 7.0, 0.1]) == [ (0,1), (1,3) ] # in the two examples below, it combines with the shorter one if both would # be above min-dur. assert CombineList(0.5, [6.0, 0.1, 7.0]) == [ (0,2), (2,3) ] assert CombineList(0.5, [7.0, 0.1, 6.0]) == [ (0,1), (1,3) ] # in the example below, it combines with whichever one would # take it above the min-dur, if there is only one such. # note, it tests the 0.1 first as the queue is popped from the end. assert CombineList(1.0, [1.0, 0.5, 0.1, 6.0]) == [ (0,2), (2,4) ] for x in range(100): min_duration = 0.05 num_utts = randint(1, 15) durations = [] for i in range(num_utts): durations.append(0.01 * randint(1, 10)) ranges = CombineList(min_duration, durations) if len(ranges) > 1: # check that each range's duration is >= min_duration for j in range(len(ranges)): (start, end) = ranges[j] this_dur = sum([ durations[k] for k in range(start, end) ]) assert not LessThan(this_dur, min_duration) # check that the list returned is not affected by very tiny differences # in the inputs. durations2 = list(durations) for i in range(len(durations2)): durations2[i] += 1.0e-07 * randint(-5, 5) ranges2 = CombineList(min_duration, durations2) assert ranges2 == ranges # This function figures out the grouping of utterances. # The input is: # 'min_duration' which is the minimum utterance length in seconds. # 'spk2utt' which is a list of pairs (speaker-id, [list-of-utterances]) # 'utt2dur' which is a dict from utterance-id to duration (as a float) # It returns a lists of lists of utterances; each list corresponds to # a group, e.g. # [ ['utt1'], ['utt2', 'utt3'] ] def GetUtteranceGroups(min_duration, spk2utt, utt2dur): # utt_groups will be a list of lists of utterance-ids formed from the # first pass of combination. utt_groups = [] # group_durations will be the durations of the corresponding elements of # 'utt_groups'. group_durations = [] # This block calls CombineList for the utterances of each speaker # separately, in the 'first pass' of combination. for i in range(len(spk2utt)): (spk, utts) = spk2utt[i] durations = [] # durations for this group of utts. for utt in utts: try: durations.append(utt2dur[utt]) except: sys.exit("choose_utts_to_combine.py: no duration available " "in utt2dur file {0} for utterance {1}".format( args.utt2dur_in, utt)) ranges = CombineList(min_duration, durations) for start, end in ranges: # each element of 'ranges' is a 2-tuple (start, end) utt_groups.append( [ utts[i] for i in range(start, end) ]) group_durations.append(sum([ durations[i] for i in range(start, end) ])) old_dur_sum = sum(utt2dur.values()) new_dur_sum = sum(group_durations) if abs(old_dur_sum - new_dur_sum) > 0.0001 * old_dur_sum: print("choose_utts_to_combine.py: large difference in total " "durations: {0} vs {1} ".format(old_dur_sum, new_dur_sum), file = sys.stderr) # Now we combine the groups obtained above, in case we had situations where # the combination of all the utterances of one speaker were still below # the minimum duration. new_utt_groups = [] ranges = CombineList(min_duration, group_durations) for start, end in ranges: # the following code is destructive of 'utt_groups' but it doesn't # matter. this_group = utt_groups[start] for i in range(start + 1, end): this_group += utt_groups[i] new_utt_groups.append(this_group) print("choose_utts_to_combine.py: combined {0} utterances to {1} utterances " "while respecting speaker boundaries, and then to {2} utterances " "with merging across speaker boundaries.".format( len(utt2dur), len(utt_groups), len(new_utt_groups)), file = sys.stderr) return utt_groups SelfTest() if args.min_duration < 0.0: print("choose_utts_to_combine.py: bad minium duration {0}".format( args.min_duration)) # spk2utt is a list of 2-tuples (speaker-id, [list-of-utterances]) spk2utt = [] # utt2spk is a dict from speaker-id to utternace-id. utt2spk = dict() try: f = open(args.spk2utt_in) except: sys.exit("choose_utts_to_combine.py: error opening --spk2utt={0}".format(args.spk2utt_in)) while True: line = f.readline() if line == '': break a = line.split() if len(a) < 2: sys.exit("choose_utts_to_combine.py: bad line in spk2utt file: " + line) spk = a[0] utts = a[1:] spk2utt.append((spk, utts)) for utt in utts: if utt in utt2spk: sys.exit("choose_utts_to_combine.py: utterance {0} is listed more than once" "in the spk2utt file {1}".format(utt, args.spk2utt_in)) utt2spk[utt] = spk f.close() # utt2dur is a dict from utterance-id (as a string) to duration in seconds (as a float) utt2dur = dict() try: f = open(args.utt2dur_in) except: sys.exit("choose_utts_to_combine.py: error opening utt2dur file {0}".format(args.utt2dur_in)) while True: line = f.readline() if line == '': break try: [ utt, dur ] = line.split() dur = float(dur) utt2dur[utt] = dur except: sys.exit("choose_utts_to_combine.py: bad line in utt2dur file {0}: {1}".format( args.utt2dur_in, line)) utt_groups = GetUtteranceGroups(args.min_duration, spk2utt, utt2dur) # set utt_group names to an array like [ 'utt1', 'utt2-comb2', 'utt4', ... ] utt_group_names = [ group[0] if len(group)==1 else "{0}-comb{1}".format(group[0], len(group)) for group in utt_groups ] # write the utt2utts file. try: with open(args.utt2utts_out, 'w') as f: for i in range(len(utt_groups)): print(utt_group_names[i], ' '.join(utt_groups[i]), file = f) except Exception as e: sys.exit("choose_utts_to_combine.py: exception writing to " "<utt2utts-out>={0}: {1}".format(args.utt2utts_out, str(e))) # write the utt2spk file. try: with open(args.utt2spk_out, 'w') as f: for i in range(len(utt_groups)): utt_group = utt_groups[i] spk_list = [ utt2spk[utt] for utt in utt_group ] if spk_list == [ spk_list[0] ] * len(utt_group): spk = spk_list[0] else: spk2dur = defaultdict(float) # spk2dur is a map from the speaker-id to the duration within this # utt, that it comprises. for utt in utt_group: spk2dur[utt2spk[utt]] += utt2dur[utt] # the following code, which picks the speaker that contributed # the most to the duration of this utterance, is a little # complex because we want to break ties in a deterministic way # picking the earlier spaker in case of a tied duration. longest_spk_dur = -1.0 spk = None for this_spk in sorted(spk2dur.keys()): if LessThan(longest_spk_dur, spk2dur[this_spk]): longest_spk_dur = spk2dur[this_spk] spk = this_spk assert spk != None new_utt = utt_group_names[i] print(new_utt, spk, file = f) except Exception as e: sys.exit("choose_utts_to_combine.py: exception writing to " "<utt2spk-out>={0}: {1}".format(args.utt2spk_out, str(e))) # write the utt2dur file. try: with open(args.utt2dur_out, 'w') as f: for i in range(len(utt_groups)): utt_name = utt_group_names[i] duration = sum([ utt2dur[utt] for utt in utt_groups[i]]) print(utt_name, duration, file = f) except Exception as e: sys.exit("choose_utts_to_combine.py: exception writing to " "<utt2dur-out>={0}: {1}".format(args.utt2dur_out, str(e)))
44.417526
103
0.635836
794d87e1f5b4d340fd89e07a1323dc8a4c510432
4,981
py
Python
web/core.py
jagoPG/-restaurant-ml-inspector
4efc7855401cc8cfa9d5e470c14685158a607448
[ "Apache-2.0" ]
1
2018-07-10T12:53:35.000Z
2018-07-10T12:53:35.000Z
web/core.py
jagoPG/-restaurant-ml-inspector
4efc7855401cc8cfa9d5e470c14685158a607448
[ "Apache-2.0" ]
null
null
null
web/core.py
jagoPG/-restaurant-ml-inspector
4efc7855401cc8cfa9d5e470c14685158a607448
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/env # -*- coding: utf-8 -*- """ Copyright 2017-2018 Jagoba Pérez-Gómez Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from flask import Flask, send_from_directory, session from flask_oauthlib.client import OAuth from flask_mail import Mail from logging.handlers import SMTPHandler from neomodel import config from src.application.config import NEO4J_CONFIG, GOOGLE_OAUTH_ID, GOOGLE_OAUTH_SECRET from src.infrastructure.dependency_injector import DependencyInjector from web.routing.admin import admin_page from web.routing.client import client_page from web.routing.api import api_page from web.routing.user_authentication import user_authentication import logging import os def download_dependencies(): import nltk nltk.download('stopwords') nltk.download('punkt') def setup_database(): if os.environ['DEBUG'] == 'True': config.DATABASE_URL = 'bolt://%s:%s@localhost:7687' % (NEO4J_CONFIG['user'], NEO4J_CONFIG['passwd']) else: BOLT_URL = os.environ['GRAPHENEDB_BOLT_URL'] BOLT_USER = os.environ['GRAPHENEDB_BOLT_USER'] BOLT_PASSWORD = os.environ['GRAPHENEDB_BOLT_PASSWORD'] config.DATABASE_URL = 'bolt://%s:%s@%s' % (BOLT_USER, BOLT_PASSWORD, BOLT_URL[BOLT_URL.find('//') + 2:]) def setup_email(app): app.config['MAIL_SERVER'] = 'smtp.gmail.com' app.config['MAIL_PORT'] = '465' app.config['MAIL_USE_SSL'] = True app.config['MAIL_USERNAME'] = os.environ['MAIL_USERNAME'] app.config['MAIL_PASSWORD'] = os.environ['MAIL_PASSWORD'] mail = Mail(app) app.config['email'] = mail def setup_logging(app): # Set up Flask logging app.config['PROPAGATE_EXCEPTIONS'] = True formatter = logging.Formatter( '[%(asctime)s] {%(pathname)s:%(lineno)d} %(levelname)s - %(message)s' ) stream_handler = logging.StreamHandler() stream_handler.setFormatter(formatter) app.logger.addHandler(stream_handler) if os.environ['DEBUG'] == 'True': app.logger.setLevel(logging.DEBUG) else: app.logger.setLevel(logging.WARNING) # mail_handler = SMTPHandler( # mailhost='smtp', # fromaddr=app.config['MAIL_USERNAME'], # credentials=app.config['MAIL_PASSWORD'], # toaddrs='jagobapg@protonmail.com', # subject='[FATAL ERROR] Restaurant Review App' # ) # mail_handler.setFormatter(formatter) # mail_handler.setLevel(logging.CRITICAL) # app.logger.addHandler(mail_handler) # Set up Python logging logging.basicConfig( format='%(levelname)s:%(message)s', level=logging.ERROR ) def setup_google_oauth(app): app.config['GOOGLE_ID'] = GOOGLE_OAUTH_ID app.config['GOOGLE_SECRET'] = GOOGLE_OAUTH_SECRET oauth = OAuth(app) google = oauth.remote_app( 'google', consumer_key=app.config.get('GOOGLE_ID'), consumer_secret=app.config.get('GOOGLE_SECRET'), request_token_params={ 'scope': 'email' }, base_url='https://www.googleapis.com/oauth2/v1/', request_token_url=None, access_token_method='POST', access_token_url='https://accounts.google.com/o/oauth2/token', authorize_url='https://accounts.google.com/o/oauth2/auth', ) app.config['GOOGLE'] = google return google def register_blueprints(app): app.register_blueprint(admin_page, url_prefix='/admin') app.register_blueprint(client_page, url_prefix='/') app.register_blueprint(api_page, url_prefix='/api') app.register_blueprint(user_authentication, url_prefix='/auth') def load_basic_data(app): dependency_injector = app.config['dependency_injector'] loader = dependency_injector.get('app.fixtures.basic_structure_loader') loader.execute() # Initialise database and application setup_database() download_dependencies() app = Flask(__name__) app.secret_key = os.environ['SECRET'] setup_logging(app) setup_email(app) google = setup_google_oauth(app) register_blueprints(app) app.config['dependency_injector'] = DependencyInjector() load_basic_data(app) @google.tokengetter def get_google_oauth_token(): return session.get('google_token') # General requests @app.route('/static/<path:path>') def get_static(path): return send_from_directory('js', path) @app.route('/images/<path:path>') def get_images(path): return send_from_directory('static/images', path) @app.route('/file/<path:path>') def get_file(path): return send_from_directory('static/files', path)
31.525316
112
0.71572
794d89dc747df7e3a5cbc78c64698a52060e01a0
4,701
py
Python
ceilometer/tests/ipmi/pollsters/test_sensor.py
gitter-badger/ceilometer
21bde474f4836b4a933caf41baf4e20e5ff25135
[ "Apache-2.0" ]
null
null
null
ceilometer/tests/ipmi/pollsters/test_sensor.py
gitter-badger/ceilometer
21bde474f4836b4a933caf41baf4e20e5ff25135
[ "Apache-2.0" ]
null
null
null
ceilometer/tests/ipmi/pollsters/test_sensor.py
gitter-badger/ceilometer
21bde474f4836b4a933caf41baf4e20e5ff25135
[ "Apache-2.0" ]
3
2015-10-08T20:03:36.000Z
2020-02-05T10:45:50.000Z
# Copyright 2014 Intel Corp. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from oslo_config import cfg from ceilometer.ipmi.pollsters import sensor from ceilometer.tests.ipmi.notifications import ipmi_test_data from ceilometer.tests.ipmi.pollsters import base CONF = cfg.CONF CONF.import_opt('host', 'ceilometer.service') TEMPERATURE_SENSOR_DATA = { 'Temperature': ipmi_test_data.TEMPERATURE_DATA } CURRENT_SENSOR_DATA = { 'Current': ipmi_test_data.CURRENT_DATA } FAN_SENSOR_DATA = { 'Fan': ipmi_test_data.FAN_DATA } VOLTAGE_SENSOR_DATA = { 'Voltage': ipmi_test_data.VOLTAGE_DATA } MISSING_SENSOR_DATA = ipmi_test_data.MISSING_SENSOR['payload']['payload'] MALFORMED_SENSOR_DATA = ipmi_test_data.BAD_SENSOR['payload']['payload'] MISSING_ID_SENSOR_DATA = ipmi_test_data.NO_SENSOR_ID['payload']['payload'] class TestTemperatureSensorPollster(base.TestPollsterBase): def fake_sensor_data(self, sensor_type): return TEMPERATURE_SENSOR_DATA def fake_data(self): # No use for Sensor test return None def make_pollster(self): return sensor.TemperatureSensorPollster() @mock.patch('ceilometer.pipeline.setup_pipeline', mock.MagicMock()) def test_get_samples(self): self._test_get_samples() self._verify_metering(10, float(32), CONF.host) class TestMissingSensorData(base.TestPollsterBase): def fake_sensor_data(self, sensor_type): return MISSING_SENSOR_DATA def fake_data(self): # No use for Sensor test return None def make_pollster(self): return sensor.TemperatureSensorPollster() @mock.patch('ceilometer.pipeline.setup_pipeline', mock.MagicMock()) def test_get_samples(self): self._test_get_samples() self._verify_metering(0) class TestMalformedSensorData(base.TestPollsterBase): def fake_sensor_data(self, sensor_type): return MALFORMED_SENSOR_DATA def fake_data(self): # No use for Sensor test return None def make_pollster(self): return sensor.TemperatureSensorPollster() @mock.patch('ceilometer.pipeline.setup_pipeline', mock.MagicMock()) def test_get_samples(self): self._test_get_samples() self._verify_metering(0) class TestMissingSensorId(base.TestPollsterBase): def fake_sensor_data(self, sensor_type): return MISSING_ID_SENSOR_DATA def fake_data(self): # No use for Sensor test return None def make_pollster(self): return sensor.TemperatureSensorPollster() @mock.patch('ceilometer.pipeline.setup_pipeline', mock.MagicMock()) def test_get_samples(self): self._test_get_samples() self._verify_metering(0) class TestFanSensorPollster(base.TestPollsterBase): def fake_sensor_data(self, sensor_type): return FAN_SENSOR_DATA def fake_data(self): # No use for Sensor test return None def make_pollster(self): return sensor.FanSensorPollster() @mock.patch('ceilometer.pipeline.setup_pipeline', mock.MagicMock()) def test_get_samples(self): self._test_get_samples() self._verify_metering(12, float(7140), CONF.host) class TestCurrentSensorPollster(base.TestPollsterBase): def fake_sensor_data(self, sensor_type): return CURRENT_SENSOR_DATA def fake_data(self): # No use for Sensor test return None def make_pollster(self): return sensor.CurrentSensorPollster() @mock.patch('ceilometer.pipeline.setup_pipeline', mock.MagicMock()) def test_get_samples(self): self._test_get_samples() self._verify_metering(1, float(130), CONF.host) class TestVoltageSensorPollster(base.TestPollsterBase): def fake_sensor_data(self, sensor_type): return VOLTAGE_SENSOR_DATA def fake_data(self): # No use for Sensor test return None def make_pollster(self): return sensor.VoltageSensorPollster() @mock.patch('ceilometer.pipeline.setup_pipeline', mock.MagicMock()) def test_get_samples(self): self._test_get_samples() self._verify_metering(4, float(3.309), CONF.host)
26.862857
75
0.721549
794d89efa0e58125e32dc57ec4d4290b697afdd9
723
py
Python
tpdatasrc/tpgamefiles/scr/tpModifiers/tiefling.py
edoipi/TemplePlus
f0e552289822fea908f16daa379fa568b1bd286d
[ "MIT" ]
69
2015-05-05T14:09:25.000Z
2022-02-15T06:13:04.000Z
tpdatasrc/tpgamefiles/scr/tpModifiers/tiefling.py
edoipi/TemplePlus
f0e552289822fea908f16daa379fa568b1bd286d
[ "MIT" ]
457
2015-05-01T22:07:45.000Z
2022-03-31T02:19:10.000Z
tpdatasrc/tpgamefiles/scr/tpModifiers/tiefling.py
edoipi/TemplePlus
f0e552289822fea908f16daa379fa568b1bd286d
[ "MIT" ]
25
2016-02-04T21:19:53.000Z
2021-11-15T23:14:51.000Z
from templeplus.pymod import PythonModifier from toee import * import tpdp import race_utils ################################################### def GetConditionName(): return "Tiefling" raceEnum = race_human + (3 << 5) raceSpecModule = __import__('race096_tiefling') ################################################### raceSpecObj = PythonModifier(GetConditionName(), 0) race_utils.AddAbilityModifierHooks(raceSpecObj, raceSpecModule) race_utils.AddSkillBonuses(raceSpecObj, {skill_bluff: 2, skill_hide: 2}) race_utils.AddBaseMoveSpeed(raceSpecObj, 30) race_utils.AddFavoredClassHook(raceSpecObj, stat_level_rogue) race_utils.AddDamageResistances(raceSpecObj, {D20DT_FIRE: 5, D20DT_COLD: 5, D20DT_ELECTRICITY: 5})
30.125
98
0.701245
794d8ab5b21e05679d025723f900e1f9ebff2230
2,807
py
Python
tests/conftest.py
shkumagai/drf_renderer_svgheatmap
43e829a8c28c855d87d51ff81082717d236545e6
[ "Apache-2.0" ]
null
null
null
tests/conftest.py
shkumagai/drf_renderer_svgheatmap
43e829a8c28c855d87d51ff81082717d236545e6
[ "Apache-2.0" ]
139
2019-09-02T14:33:01.000Z
2022-03-29T01:49:07.000Z
tests/conftest.py
shkumagai/drf_renderer_svgheatmap
43e829a8c28c855d87d51ff81082717d236545e6
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- # Copyright 2019 Shoji Kumagai # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os.path import sys import django def pytest_addoption(parser): parser.addoption( "--no-pkgroot", action="store_true", default=False, help=( "Remove package root directory from sys.path, ensuring that " "drf_renderer_svgheatmap is imported from the installed site-packages. " "Used for testing the distribution." ), ) def pytest_configure(config): from django.conf import settings settings.configure( DEBUG_PROPAGATE_EXCEPTION=True, DATABASES={ "default": { "ENGINE": "django.db.backends.sqlite3", "NAME": ":memory:", } }, SITE_ID=1, SECRET_KEY="not very secret in tests", USE_I18N=True, USE_L10N=True, # USE_DEPRECATED_PYTZ=False, STATIC_URL="/static/", ROOT_URLCONF="tests.urls", TEMPLATES=[ { "BACKEND": "django.template.backends.django.DjangoTemplates", "APP_DIRS": True, "OPTIONS": { "debug": True, }, }, ], MODDLEWARE=( "django.middleware.common.CommonMiddleware", "django.contrib.sessions.middleware.SessionMiddleware", "django.contrib.auth.middleware.AuthenticationMiddleware", "django.contrib.messages.middleware.MessageMiddleware", ), INSTALLED_APPS=( "django.contrib.admin", "django.contrib.auth", "django.contrib.contenttypes", "django.contrib.sessions", "django.contrib.staticfiles", "rest_framework", "rest_framework.authtoken", "drf_renderer_svgheatmap", "tests", ), PASSWORD_HASHERS=("django.contrib.auth.hashers.MD5PasswordHasher",), ) if config.getoption("--no-pkgroot"): sys.path.pop(0) import drf_renderer_svgheatmap package_dir = os.path.join(os.getcwd(), "drf_renderer_svgheatmap") assert not drf_renderer_svgheatmap.__file__.startswith(package_dir) django.setup()
30.846154
84
0.603491