query stringlengths 5 1.23k | positive stringlengths 53 15.2k | id_ int64 0 252k | task_name stringlengths 87 242 | negative listlengths 20 553 |
|---|---|---|---|---|
Returns username from the configfile . | def get_auth ( self ) : return ( self . _cfgparse . get ( self . _section , 'username' ) , self . _cfgparse . get ( self . _section , 'password' ) ) | 251,000 | https://github.com/paragbaxi/qualysapi/blob/2c8bf1d5d300117403062885c8e10b5665eb4615/qualysapi/config.py#L211-L213 | [
"def",
"reserve_udp_port",
"(",
"self",
",",
"port",
",",
"project",
")",
":",
"if",
"port",
"in",
"self",
".",
"_used_udp_ports",
":",
"raise",
"HTTPConflict",
"(",
"text",
"=",
"\"UDP port {} already in use on host {}\"",
".",
"format",
"(",
"port",
",",
"se... |
Return a QGAPIConnect object for v1 API pulling settings from config file . | def connect ( config_file = qcs . default_filename , section = 'info' , remember_me = False , remember_me_always = False ) : # Retrieve login credentials. conf = qcconf . QualysConnectConfig ( filename = config_file , section = section , remember_me = remember_me , remember_me_always = remember_me_always ) connect = qcconn . QGConnector ( conf . get_auth ( ) , conf . get_hostname ( ) , conf . proxies , conf . max_retries ) logger . info ( "Finished building connector." ) return connect | 251,001 | https://github.com/paragbaxi/qualysapi/blob/2c8bf1d5d300117403062885c8e10b5665eb4615/qualysapi/util.py#L18-L30 | [
"def",
"medial_wall_to_nan",
"(",
"in_file",
",",
"subjects_dir",
",",
"target_subject",
",",
"newpath",
"=",
"None",
")",
":",
"import",
"nibabel",
"as",
"nb",
"import",
"numpy",
"as",
"np",
"import",
"os",
"fn",
"=",
"os",
".",
"path",
".",
"basename",
... |
Return QualysGuard API version for api_version specified . | def format_api_version ( self , api_version ) : # Convert to int. if type ( api_version ) == str : api_version = api_version . lower ( ) if api_version [ 0 ] == 'v' and api_version [ 1 ] . isdigit ( ) : # Remove first 'v' in case the user typed 'v1' or 'v2', etc. api_version = api_version [ 1 : ] # Check for input matching Qualys modules. if api_version in ( 'asset management' , 'assets' , 'tag' , 'tagging' , 'tags' ) : # Convert to Asset Management API. api_version = 'am' elif api_version in ( 'am2' ) : # Convert to Asset Management API v2 api_version = 'am2' elif api_version in ( 'webapp' , 'web application scanning' , 'webapp scanning' ) : # Convert to WAS API. api_version = 'was' elif api_version in ( 'pol' , 'pc' ) : # Convert PC module to API number 2. api_version = 2 else : api_version = int ( api_version ) return api_version | 251,002 | https://github.com/paragbaxi/qualysapi/blob/2c8bf1d5d300117403062885c8e10b5665eb4615/qualysapi/connector.py#L72-L97 | [
"def",
"_SeparateTypes",
"(",
"self",
",",
"metadata_value_pairs",
")",
":",
"registry_pairs",
"=",
"[",
"]",
"file_pairs",
"=",
"[",
"]",
"match_pairs",
"=",
"[",
"]",
"for",
"metadata",
",",
"result",
"in",
"metadata_value_pairs",
":",
"if",
"(",
"result",... |
Return QualysGuard API version for api_call specified . | def which_api_version ( self , api_call ) : # Leverage patterns of calls to API methods. if api_call . endswith ( '.php' ) : # API v1. return 1 elif api_call . startswith ( 'api/2.0/' ) : # API v2. return 2 elif '/am/' in api_call : # Asset Management API. return 'am' elif '/was/' in api_call : # WAS API. return 'was' return False | 251,003 | https://github.com/paragbaxi/qualysapi/blob/2c8bf1d5d300117403062885c8e10b5665eb4615/qualysapi/connector.py#L99-L116 | [
"def",
"_SeparateTypes",
"(",
"self",
",",
"metadata_value_pairs",
")",
":",
"registry_pairs",
"=",
"[",
"]",
"file_pairs",
"=",
"[",
"]",
"match_pairs",
"=",
"[",
"]",
"for",
"metadata",
",",
"result",
"in",
"metadata_value_pairs",
":",
"if",
"(",
"result",... |
Return base API url string for the QualysGuard api_version and server . | def url_api_version ( self , api_version ) : # Set base url depending on API version. if api_version == 1 : # QualysGuard API v1 url. url = "https://%s/msp/" % ( self . server , ) elif api_version == 2 : # QualysGuard API v2 url. url = "https://%s/" % ( self . server , ) elif api_version == 'was' : # QualysGuard REST v3 API url (Portal API). url = "https://%s/qps/rest/3.0/" % ( self . server , ) elif api_version == 'am' : # QualysGuard REST v1 API url (Portal API). url = "https://%s/qps/rest/1.0/" % ( self . server , ) elif api_version == 'am2' : # QualysGuard REST v1 API url (Portal API). url = "https://%s/qps/rest/2.0/" % ( self . server , ) else : raise Exception ( "Unknown QualysGuard API Version Number (%s)" % ( api_version , ) ) logger . debug ( "Base url =\n%s" % ( url ) ) return url | 251,004 | https://github.com/paragbaxi/qualysapi/blob/2c8bf1d5d300117403062885c8e10b5665eb4615/qualysapi/connector.py#L118-L141 | [
"def",
"object",
"(",
"self",
")",
":",
"if",
"self",
".",
"type",
"==",
"EntryType",
".",
"category",
":",
"return",
"self",
".",
"category",
"elif",
"self",
".",
"type",
"==",
"EntryType",
".",
"event",
":",
"return",
"self",
".",
"event",
"elif",
... |
Return QualysGuard API http method with POST preferred .. | def format_http_method ( self , api_version , api_call , data ) : # Define get methods for automatic http request methodology. # # All API v2 requests are POST methods. if api_version == 2 : return 'post' elif api_version == 1 : if api_call in self . api_methods [ '1 post' ] : return 'post' else : return 'get' elif api_version == 'was' : # WAS API call. # Because WAS API enables user to GET API resources in URI, let's chop off the resource. # '/download/was/report/18823' --> '/download/was/report/' api_call_endpoint = api_call [ : api_call . rfind ( '/' ) + 1 ] if api_call_endpoint in self . api_methods [ 'was get' ] : return 'get' # Post calls with no payload will result in HTTPError: 415 Client Error: Unsupported Media Type. if data is None : # No post data. Some calls change to GET with no post data. if api_call_endpoint in self . api_methods [ 'was no data get' ] : return 'get' else : return 'post' else : # Call with post data. return 'post' else : # Asset Management API call. if api_call in self . api_methods [ 'am get' ] : return 'get' else : return 'post' | 251,005 | https://github.com/paragbaxi/qualysapi/blob/2c8bf1d5d300117403062885c8e10b5665eb4615/qualysapi/connector.py#L143-L179 | [
"def",
"calc_regenerated",
"(",
"self",
",",
"lastvotetime",
")",
":",
"delta",
"=",
"datetime",
".",
"utcnow",
"(",
")",
"-",
"datetime",
".",
"strptime",
"(",
"lastvotetime",
",",
"'%Y-%m-%dT%H:%M:%S'",
")",
"td",
"=",
"delta",
".",
"days",
"ts",
"=",
... |
Return properly formatted QualysGuard API call . | def preformat_call ( self , api_call ) : # Remove possible starting slashes or trailing question marks in call. api_call_formatted = api_call . lstrip ( '/' ) api_call_formatted = api_call_formatted . rstrip ( '?' ) if api_call != api_call_formatted : # Show difference logger . debug ( 'api_call post strip =\n%s' % api_call_formatted ) return api_call_formatted | 251,006 | https://github.com/paragbaxi/qualysapi/blob/2c8bf1d5d300117403062885c8e10b5665eb4615/qualysapi/connector.py#L181-L191 | [
"def",
"load_metafile",
"(",
"filepath",
")",
":",
"try",
":",
"with",
"open",
"(",
"filepath",
",",
"'r'",
",",
"encoding",
"=",
"'utf-8'",
")",
"as",
"file",
":",
"return",
"email",
".",
"message_from_file",
"(",
"file",
")",
"except",
"FileNotFoundError... |
Return properly formatted QualysGuard API call according to api_version etiquette . | def format_call ( self , api_version , api_call ) : # Remove possible starting slashes or trailing question marks in call. api_call = api_call . lstrip ( '/' ) api_call = api_call . rstrip ( '?' ) logger . debug ( 'api_call post strip =\n%s' % api_call ) # Make sure call always ends in slash for API v2 calls. if ( api_version == 2 and api_call [ - 1 ] != '/' ) : # Add slash. logger . debug ( 'Adding "/" to api_call.' ) api_call += '/' if api_call in self . api_methods_with_trailing_slash [ api_version ] : # Add slash. logger . debug ( 'Adding "/" to api_call.' ) api_call += '/' return api_call | 251,007 | https://github.com/paragbaxi/qualysapi/blob/2c8bf1d5d300117403062885c8e10b5665eb4615/qualysapi/connector.py#L193-L210 | [
"def",
"stream_filesystem_node",
"(",
"path",
",",
"recursive",
"=",
"False",
",",
"patterns",
"=",
"'**'",
",",
"chunk_size",
"=",
"default_chunk_size",
")",
":",
"is_dir",
"=",
"isinstance",
"(",
"path",
",",
"six",
".",
"string_types",
")",
"and",
"os",
... |
Return appropriate QualysGuard API call . | def format_payload ( self , api_version , data ) : # Check if payload is for API v1 or API v2. if ( api_version in ( 1 , 2 ) ) : # Check if string type. if type ( data ) == str : # Convert to dictionary. logger . debug ( 'Converting string to dict:\n%s' % data ) # Remove possible starting question mark & ending ampersands. data = data . lstrip ( '?' ) data = data . rstrip ( '&' ) # Convert to dictionary. data = parse_qs ( data ) logger . debug ( 'Converted:\n%s' % str ( data ) ) elif api_version in ( 'am' , 'was' , 'am2' ) : if type ( data ) == etree . _Element : logger . debug ( 'Converting lxml.builder.E to string' ) data = etree . tostring ( data ) logger . debug ( 'Converted:\n%s' % data ) return data | 251,008 | https://github.com/paragbaxi/qualysapi/blob/2c8bf1d5d300117403062885c8e10b5665eb4615/qualysapi/connector.py#L212-L233 | [
"def",
"_result_is_lyrics",
"(",
"self",
",",
"song_title",
")",
":",
"default_terms",
"=",
"[",
"'track\\\\s?list'",
",",
"'album art(work)?'",
",",
"'liner notes'",
",",
"'booklet'",
",",
"'credits'",
",",
"'interview'",
",",
"'skit'",
",",
"'instrumental'",
","... |
Wait for all jobs to finish then exit successfully . | def travis_after ( ini , envlist ) : # after-all disabled for pull requests if os . environ . get ( 'TRAVIS_PULL_REQUEST' , 'false' ) != 'false' : return if not after_config_matches ( ini , envlist ) : return # This is not the one that needs to wait github_token = os . environ . get ( 'GITHUB_TOKEN' ) if not github_token : print ( 'No GitHub token given.' , file = sys . stderr ) sys . exit ( NO_GITHUB_TOKEN ) api_url = os . environ . get ( 'TRAVIS_API_URL' , 'https://api.travis-ci.org' ) build_id = os . environ . get ( 'TRAVIS_BUILD_ID' ) job_number = os . environ . get ( 'TRAVIS_JOB_NUMBER' ) try : polling_interval = int ( os . environ . get ( 'TRAVIS_POLLING_INTERVAL' , 5 ) ) except ValueError : print ( 'Invalid polling interval given: {0}' . format ( repr ( os . environ . get ( 'TRAVIS_POLLING_INTERVAL' ) ) ) , file = sys . stderr ) sys . exit ( INVALID_POLLING_INTERVAL ) if not all ( [ api_url , build_id , job_number ] ) : print ( 'Required Travis environment not given.' , file = sys . stderr ) sys . exit ( INCOMPLETE_TRAVIS_ENVIRONMENT ) # This may raise an Exception, and it should be printed job_statuses = get_job_statuses ( github_token , api_url , build_id , polling_interval , job_number ) if not all ( job_statuses ) : print ( 'Some jobs were not successful.' ) sys . exit ( JOBS_FAILED ) print ( 'All required jobs were successful.' ) | 251,009 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/after.py#L25-L62 | [
"def",
"characterize_local_files",
"(",
"filedir",
",",
"max_bytes",
"=",
"MAX_FILE_DEFAULT",
")",
":",
"file_data",
"=",
"{",
"}",
"logging",
".",
"info",
"(",
"'Characterizing files in {}'",
".",
"format",
"(",
"filedir",
")",
")",
"for",
"filename",
"in",
"... |
Determine if this job should wait for the others . | def after_config_matches ( ini , envlist ) : section = ini . sections . get ( 'travis:after' , { } ) if not section : return False # Never wait if it's not configured if 'envlist' in section or 'toxenv' in section : if 'toxenv' in section : print ( 'The "toxenv" key of the [travis:after] section is ' 'deprecated in favor of the "envlist" key.' , file = sys . stderr ) toxenv = section . get ( 'toxenv' ) required = set ( split_env ( section . get ( 'envlist' , toxenv ) or '' ) ) actual = set ( envlist ) if required - actual : return False # Translate travis requirements to env requirements env_requirements = [ ( TRAVIS_FACTORS [ factor ] , value ) for factor , value in parse_dict ( section . get ( 'travis' , '' ) ) . items ( ) if factor in TRAVIS_FACTORS ] + [ ( name , value ) for name , value in parse_dict ( section . get ( 'env' , '' ) ) . items ( ) ] return all ( [ os . environ . get ( name ) == value for name , value in env_requirements ] ) | 251,010 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/after.py#L65-L96 | [
"def",
"get_model_file",
"(",
"name",
",",
"root",
"=",
"os",
".",
"path",
".",
"join",
"(",
"base",
".",
"data_dir",
"(",
")",
",",
"'models'",
")",
")",
":",
"file_name",
"=",
"'{name}-{short_hash}'",
".",
"format",
"(",
"name",
"=",
"name",
",",
"... |
Wait for all the travis jobs to complete . | def get_job_statuses ( github_token , api_url , build_id , polling_interval , job_number ) : auth = get_json ( '{api_url}/auth/github' . format ( api_url = api_url ) , data = { 'github_token' : github_token } ) [ 'access_token' ] while True : build = get_json ( '{api_url}/builds/{build_id}' . format ( api_url = api_url , build_id = build_id ) , auth = auth ) jobs = [ job for job in build [ 'jobs' ] if job [ 'number' ] != job_number and not job [ 'allow_failure' ] ] # Ignore allowed failures if all ( job [ 'finished_at' ] for job in jobs ) : break # All the jobs have completed elif any ( job [ 'state' ] != 'passed' for job in jobs if job [ 'finished_at' ] ) : break # Some required job that finished did not pass print ( 'Waiting for jobs to complete: {job_numbers}' . format ( job_numbers = [ job [ 'number' ] for job in jobs if not job [ 'finished_at' ] ] ) ) time . sleep ( polling_interval ) return [ job [ 'state' ] == 'passed' for job in jobs ] | 251,011 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/after.py#L99-L127 | [
"def",
"del_reference",
"(",
"self",
",",
"source",
",",
"target",
",",
"*",
"*",
"kwargs",
")",
":",
"rc",
"=",
"api",
".",
"get_tool",
"(",
"\"reference_catalog\"",
")",
"uid",
"=",
"api",
".",
"get_uid",
"(",
"target",
")",
"rc",
".",
"deleteReferen... |
Make a GET request and return the response as parsed JSON . | def get_json ( url , auth = None , data = None ) : headers = { 'Accept' : 'application/vnd.travis-ci.2+json' , 'User-Agent' : 'Travis/Tox-Travis-1.0a' , # User-Agent must start with "Travis/" in order to work } if auth : headers [ 'Authorization' ] = 'token {auth}' . format ( auth = auth ) params = { } if data : headers [ 'Content-Type' ] = 'application/json' params [ 'data' ] = json . dumps ( data ) . encode ( 'utf-8' ) request = urllib2 . Request ( url , headers = headers , * * params ) response = urllib2 . urlopen ( request ) . read ( ) return json . loads ( response . decode ( 'utf-8' ) ) | 251,012 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/after.py#L130-L147 | [
"def",
"update",
"(",
"self",
")",
"->",
"None",
":",
"mask",
"=",
"self",
".",
"mask",
"weights",
"=",
"self",
".",
"refweights",
"[",
"mask",
"]",
"self",
"[",
"~",
"mask",
"]",
"=",
"numpy",
".",
"nan",
"self",
"[",
"mask",
"]",
"=",
"weights"... |
Default envlist automatically based on the Travis environment . | def detect_envlist ( ini ) : # Find the envs that tox knows about declared_envs = get_declared_envs ( ini ) # Find all the envs for all the desired factors given desired_factors = get_desired_factors ( ini ) # Reduce desired factors desired_envs = [ '-' . join ( env ) for env in product ( * desired_factors ) ] # Find matching envs return match_envs ( declared_envs , desired_envs , passthru = len ( desired_factors ) == 1 ) | 251,013 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/envlist.py#L14-L27 | [
"def",
"close",
"(",
"self",
")",
":",
"# Instead of actually closing the connection,",
"# unshare it and/or return it to the pool.",
"if",
"self",
".",
"_con",
":",
"self",
".",
"_pool",
".",
"unshare",
"(",
"self",
".",
"_shared_con",
")",
"self",
".",
"_shared_co... |
Make the envconfigs for undeclared envs . | def autogen_envconfigs ( config , envs ) : prefix = 'tox' if config . toxinipath . basename == 'setup.cfg' else None reader = tox . config . SectionReader ( "tox" , config . _cfg , prefix = prefix ) distshare_default = "{homedir}/.tox/distshare" reader . addsubstitutions ( toxinidir = config . toxinidir , homedir = config . homedir ) reader . addsubstitutions ( toxworkdir = config . toxworkdir ) config . distdir = reader . getpath ( "distdir" , "{toxworkdir}/dist" ) reader . addsubstitutions ( distdir = config . distdir ) config . distshare = reader . getpath ( "distshare" , distshare_default ) reader . addsubstitutions ( distshare = config . distshare ) try : make_envconfig = tox . config . ParseIni . make_envconfig # tox 3.4.0+ except AttributeError : make_envconfig = tox . config . parseini . make_envconfig # Dig past the unbound method in Python 2 make_envconfig = getattr ( make_envconfig , '__func__' , make_envconfig ) # Create the undeclared envs for env in envs : section = tox . config . testenvprefix + env config . envconfigs [ env ] = make_envconfig ( config , env , section , reader . _subs , config ) | 251,014 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/envlist.py#L30-L59 | [
"def",
"_strip_ctype",
"(",
"name",
",",
"ctype",
",",
"protocol",
"=",
"2",
")",
":",
"# parse channel type from name (e.g. 'L1:GDS-CALIB_STRAIN,reduced')",
"try",
":",
"name",
",",
"ctypestr",
"=",
"name",
".",
"rsplit",
"(",
"','",
",",
"1",
")",
"except",
... |
Get the full list of envs from the tox ini . | def get_declared_envs ( ini ) : tox_section_name = 'tox:tox' if ini . path . endswith ( 'setup.cfg' ) else 'tox' tox_section = ini . sections . get ( tox_section_name , { } ) envlist = split_env ( tox_section . get ( 'envlist' , [ ] ) ) # Add additional envs that are declared as sections in the ini section_envs = [ section [ 8 : ] for section in sorted ( ini . sections , key = ini . lineof ) if section . startswith ( 'testenv:' ) ] return envlist + [ env for env in section_envs if env not in envlist ] | 251,015 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/envlist.py#L62-L81 | [
"def",
"correlator",
"(",
"A",
",",
"B",
")",
":",
"correlators",
"=",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"len",
"(",
"A",
")",
")",
":",
"correlator_row",
"=",
"[",
"]",
"for",
"j",
"in",
"range",
"(",
"len",
"(",
"B",
")",
")",
":",
... |
Get version info from the sys module . | def get_version_info ( ) : overrides = os . environ . get ( '__TOX_TRAVIS_SYS_VERSION' ) if overrides : version , major , minor = overrides . split ( ',' ) [ : 3 ] major , minor = int ( major ) , int ( minor ) else : version , ( major , minor ) = sys . version , sys . version_info [ : 2 ] return version , major , minor | 251,016 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/envlist.py#L84-L95 | [
"def",
"compare",
"(",
"self",
",",
"control_result",
",",
"experimental_result",
")",
":",
"_compare",
"=",
"getattr",
"(",
"self",
",",
"'_compare'",
",",
"lambda",
"x",
",",
"y",
":",
"x",
"==",
"y",
")",
"return",
"(",
"# Mismatch if only one of the resu... |
Guess the default python env to use . | def guess_python_env ( ) : version , major , minor = get_version_info ( ) if 'PyPy' in version : return 'pypy3' if major == 3 else 'pypy' return 'py{major}{minor}' . format ( major = major , minor = minor ) | 251,017 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/envlist.py#L98-L103 | [
"def",
"revoke_session",
"(",
"self",
",",
"sid",
"=",
"''",
",",
"token",
"=",
"''",
")",
":",
"if",
"not",
"sid",
":",
"if",
"token",
":",
"sid",
"=",
"self",
".",
"handler",
".",
"sid",
"(",
"token",
")",
"else",
":",
"raise",
"ValueError",
"(... |
Parse a default tox env based on the version . | def get_default_envlist ( version ) : if version in [ 'pypy' , 'pypy3' ] : return version # Assume single digit major and minor versions match = re . match ( r'^(\d)\.(\d)(?:\.\d+)?$' , version or '' ) if match : major , minor = match . groups ( ) return 'py{major}{minor}' . format ( major = major , minor = minor ) return guess_python_env ( ) | 251,018 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/envlist.py#L106-L122 | [
"def",
"share_nans",
"(",
"self",
")",
":",
"def",
"f",
"(",
"_",
",",
"s",
",",
"channels",
")",
":",
"outs",
"=",
"wt_kit",
".",
"share_nans",
"(",
"*",
"[",
"c",
"[",
"s",
"]",
"for",
"c",
"in",
"channels",
"]",
")",
"for",
"c",
",",
"o",
... |
Get the list of desired envs per declared factor . | def get_desired_factors ( ini ) : # Find configuration based on known travis factors travis_section = ini . sections . get ( 'travis' , { } ) found_factors = [ ( factor , parse_dict ( travis_section [ factor ] ) ) for factor in TRAVIS_FACTORS if factor in travis_section ] # Backward compatibility with the old tox:travis section if 'tox:travis' in ini . sections : print ( 'The [tox:travis] section is deprecated in favor of' ' the "python" key of the [travis] section.' , file = sys . stderr ) found_factors . append ( ( 'python' , ini . sections [ 'tox:travis' ] ) ) # Inject any needed autoenv version = os . environ . get ( 'TRAVIS_PYTHON_VERSION' ) if version : default_envlist = get_default_envlist ( version ) if not any ( factor == 'python' for factor , _ in found_factors ) : found_factors . insert ( 0 , ( 'python' , { version : default_envlist } ) ) python_factors = [ ( factor , mapping ) for factor , mapping in found_factors if version and factor == 'python' ] for _ , mapping in python_factors : mapping . setdefault ( version , default_envlist ) # Convert known travis factors to env factors, # and combine with declared env factors. env_factors = [ ( TRAVIS_FACTORS [ factor ] , mapping ) for factor , mapping in found_factors ] + [ ( name , parse_dict ( value ) ) for name , value in ini . sections . get ( 'travis:env' , { } ) . items ( ) ] # Choose the correct envlists based on the factor values return [ split_env ( mapping [ os . environ [ name ] ] ) for name , mapping in env_factors if name in os . environ and os . environ [ name ] in mapping ] | 251,019 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/envlist.py#L125-L197 | [
"async",
"def",
"on_raw_422",
"(",
"self",
",",
"message",
")",
":",
"await",
"self",
".",
"_registration_completed",
"(",
"message",
")",
"self",
".",
"motd",
"=",
"None",
"await",
"self",
".",
"on_connect",
"(",
")"
] |
Determine the envs that match the desired_envs . | def match_envs ( declared_envs , desired_envs , passthru ) : matched = [ declared for declared in declared_envs if any ( env_matches ( declared , desired ) for desired in desired_envs ) ] return desired_envs if not matched and passthru else matched | 251,020 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/envlist.py#L200-L215 | [
"def",
"_read_body_by_chunk",
"(",
"self",
",",
"response",
",",
"file",
",",
"raw",
"=",
"False",
")",
":",
"reader",
"=",
"ChunkedTransferReader",
"(",
"self",
".",
"_connection",
")",
"file_is_async",
"=",
"hasattr",
"(",
"file",
",",
"'drain'",
")",
"w... |
Determine if a declared env matches a desired env . | def env_matches ( declared , desired ) : desired_factors = desired . split ( '-' ) declared_factors = declared . split ( '-' ) return all ( factor in declared_factors for factor in desired_factors ) | 251,021 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/envlist.py#L218-L228 | [
"def",
"updateLink",
"(",
"page",
",",
"lnk",
")",
":",
"CheckParent",
"(",
"page",
")",
"annot",
"=",
"getLinkText",
"(",
"page",
",",
"lnk",
")",
"if",
"annot",
"==",
"\"\"",
":",
"raise",
"ValueError",
"(",
"\"link kind not supported\"",
")",
"page",
... |
Decide whether to override ignore_outcomes . | def override_ignore_outcome ( ini ) : travis_reader = tox . config . SectionReader ( "travis" , ini ) return travis_reader . getbool ( 'unignore_outcomes' , False ) | 251,022 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/envlist.py#L231-L234 | [
"def",
"get_connection",
"(",
"self",
",",
"command",
",",
"args",
"=",
"(",
")",
")",
":",
"# TODO: find a better way to determine if connection is free",
"# and not havily used.",
"command",
"=",
"command",
".",
"upper",
"(",
")",
".",
"strip",
"(",
")",
"... |
Add arguments and needed monkeypatches . | def tox_addoption ( parser ) : parser . add_argument ( '--travis-after' , dest = 'travis_after' , action = 'store_true' , help = 'Exit successfully after all Travis jobs complete successfully.' ) if 'TRAVIS' in os . environ : pypy_version_monkeypatch ( ) subcommand_test_monkeypatch ( tox_subcommand_test_post ) | 251,023 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/hooks.py#L19-L27 | [
"def",
"stop_capture",
"(",
"self",
")",
":",
"if",
"self",
".",
"_capture_node",
":",
"yield",
"from",
"self",
".",
"_capture_node",
"[",
"\"node\"",
"]",
".",
"post",
"(",
"\"/adapters/{adapter_number}/ports/{port_number}/stop_capture\"",
".",
"format",
"(",
"ad... |
Check for the presence of the added options . | def tox_configure ( config ) : if 'TRAVIS' not in os . environ : return ini = config . _cfg # envlist if 'TOXENV' not in os . environ and not config . option . env : envlist = detect_envlist ( ini ) undeclared = set ( envlist ) - set ( config . envconfigs ) if undeclared : print ( 'Matching undeclared envs is deprecated. Be sure all the ' 'envs that Tox should run are declared in the tox config.' , file = sys . stderr ) autogen_envconfigs ( config , undeclared ) config . envlist = envlist # Override ignore_outcomes if override_ignore_outcome ( ini ) : for envconfig in config . envconfigs . values ( ) : envconfig . ignore_outcome = False # after if config . option . travis_after : print ( 'The after all feature has been deprecated. Check out Travis\' ' 'build stages, which are a better solution. ' 'See https://tox-travis.readthedocs.io/en/stable/after.html ' 'for more details.' , file = sys . stderr ) | 251,024 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/hooks.py#L31-L59 | [
"def",
"_read_body_until_close",
"(",
"self",
",",
"response",
",",
"file",
")",
":",
"_logger",
".",
"debug",
"(",
"'Reading body until close.'",
")",
"file_is_async",
"=",
"hasattr",
"(",
"file",
",",
"'drain'",
")",
"while",
"True",
":",
"data",
"=",
"yie... |
Parse a dict value from the tox config . | def parse_dict ( value ) : lines = [ line . strip ( ) for line in value . strip ( ) . splitlines ( ) ] pairs = [ line . split ( ':' , 1 ) for line in lines if line ] return dict ( ( k . strip ( ) , v . strip ( ) ) for k , v in pairs ) | 251,025 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/utils.py#L11-L32 | [
"def",
"calculate_com",
"(",
"self",
",",
"first",
"=",
"0",
",",
"last",
"=",
"None",
")",
":",
"if",
"last",
"is",
"None",
":",
"last",
"=",
"self",
".",
"N_real",
"clibrebound",
".",
"reb_get_com_range",
".",
"restype",
"=",
"Particle",
"return",
"c... |
Patch Tox to work with non - default PyPy 3 versions . | def pypy_version_monkeypatch ( ) : # Travis virtualenv do not provide `pypy3`, which tox tries to execute. # This doesnt affect Travis python version `pypy3`, as the pyenv pypy3 # is in the PATH. # https://github.com/travis-ci/travis-ci/issues/6304 # Force use of the virtualenv `python`. version = os . environ . get ( 'TRAVIS_PYTHON_VERSION' ) if version and default_factors and version . startswith ( 'pypy3.3-' ) : default_factors [ 'pypy3' ] = 'python' | 251,026 | https://github.com/tox-dev/tox-travis/blob/d97a966c19abb020298a7e4b91fe83dd1d0a4517/src/tox_travis/hacks.py#L9-L18 | [
"def",
"_mem",
"(",
"self",
")",
":",
"value",
"=",
"int",
"(",
"psutil",
".",
"virtual_memory",
"(",
")",
".",
"percent",
")",
"set_metric",
"(",
"\"memory\"",
",",
"value",
",",
"category",
"=",
"self",
".",
"category",
")",
"gauge",
"(",
"\"memory\"... |
The direction of the pin either True for an input or False for an output . | def direction ( self ) : if _get_bit ( self . _mcp . iodir , self . _pin ) : return digitalio . Direction . INPUT return digitalio . Direction . OUTPUT | 251,027 | https://github.com/adafruit/Adafruit_CircuitPython_MCP230xx/blob/da9480befecef31c2428062919b9f3da6f428d15/adafruit_mcp230xx.py#L148-L154 | [
"def",
"drawLeftStatus",
"(",
"self",
",",
"scr",
",",
"vs",
")",
":",
"cattr",
"=",
"CursesAttr",
"(",
"colors",
".",
"color_status",
")",
"attr",
"=",
"cattr",
".",
"attr",
"error_attr",
"=",
"cattr",
".",
"update_attr",
"(",
"colors",
".",
"color_erro... |
Enable or disable internal pull - up resistors for this pin . A value of digitalio . Pull . UP will enable a pull - up resistor and None will disable it . Pull - down resistors are NOT supported! | def pull ( self ) : if _get_bit ( self . _mcp . gppu , self . _pin ) : return digitalio . Pull . UP return None | 251,028 | https://github.com/adafruit/Adafruit_CircuitPython_MCP230xx/blob/da9480befecef31c2428062919b9f3da6f428d15/adafruit_mcp230xx.py#L166-L173 | [
"def",
"remove_targets",
"(",
"self",
",",
"type",
",",
"kept",
"=",
"None",
")",
":",
"if",
"kept",
"is",
"None",
":",
"kept",
"=",
"[",
"i",
"for",
"i",
",",
"x",
"in",
"enumerate",
"(",
"self",
".",
"_targets",
")",
"if",
"not",
"isinstance",
... |
Returns the number of throttled read events during a given time frame | def get_throttled_read_event_count ( table_name , lookback_window_start = 15 , lookback_period = 5 ) : try : metrics = __get_aws_metric ( table_name , lookback_window_start , lookback_period , 'ReadThrottleEvents' ) except BotoServerError : raise if metrics : throttled_read_events = int ( metrics [ 0 ] [ 'Sum' ] ) else : throttled_read_events = 0 logger . info ( '{0} - Read throttle count: {1:d}' . format ( table_name , throttled_read_events ) ) return throttled_read_events | 251,029 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/statistics/table.py#L58-L86 | [
"def",
"status",
"(",
"server",
",",
"output",
",",
"verbose",
")",
":",
"if",
"verbose",
":",
"log",
"(",
"\"Connecting to Entity Matching Server: {}\"",
".",
"format",
"(",
"server",
")",
")",
"service_status",
"=",
"server_get_status",
"(",
"server",
")",
"... |
Returns the number of throttled read events in percent of consumption | def get_throttled_by_consumed_read_percent ( table_name , lookback_window_start = 15 , lookback_period = 5 ) : try : metrics1 = __get_aws_metric ( table_name , lookback_window_start , lookback_period , 'ConsumedReadCapacityUnits' ) metrics2 = __get_aws_metric ( table_name , lookback_window_start , lookback_period , 'ReadThrottleEvents' ) except BotoServerError : raise if metrics1 and metrics2 : lookback_seconds = lookback_period * 60 throttled_by_consumed_read_percent = ( ( ( float ( metrics2 [ 0 ] [ 'Sum' ] ) / float ( lookback_seconds ) ) / ( float ( metrics1 [ 0 ] [ 'Sum' ] ) / float ( lookback_seconds ) ) ) * 100 ) else : throttled_by_consumed_read_percent = 0 logger . info ( '{0} - Throttled read percent by consumption: {1:.2f}%' . format ( table_name , throttled_by_consumed_read_percent ) ) return throttled_by_consumed_read_percent | 251,030 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/statistics/table.py#L132-L171 | [
"def",
"_set_all_zone_stereo",
"(",
"self",
",",
"zst_on",
")",
":",
"command_url",
"=",
"self",
".",
"_urls",
".",
"command_set_all_zone_stereo",
"if",
"zst_on",
":",
"command_url",
"+=",
"\"ZST ON\"",
"else",
":",
"command_url",
"+=",
"\"ZST OFF\"",
"try",
":"... |
Returns the number of throttled write events in percent of consumption | def get_throttled_by_consumed_write_percent ( table_name , lookback_window_start = 15 , lookback_period = 5 ) : try : metrics1 = __get_aws_metric ( table_name , lookback_window_start , lookback_period , 'ConsumedWriteCapacityUnits' ) metrics2 = __get_aws_metric ( table_name , lookback_window_start , lookback_period , 'WriteThrottleEvents' ) except BotoServerError : raise if metrics1 and metrics2 : lookback_seconds = lookback_period * 60 throttled_by_consumed_write_percent = ( ( ( float ( metrics2 [ 0 ] [ 'Sum' ] ) / float ( lookback_seconds ) ) / ( float ( metrics1 [ 0 ] [ 'Sum' ] ) / float ( lookback_seconds ) ) ) * 100 ) else : throttled_by_consumed_write_percent = 0 logger . info ( '{0} - Throttled write percent by consumption: {1:.2f}%' . format ( table_name , throttled_by_consumed_write_percent ) ) return throttled_by_consumed_write_percent | 251,031 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/statistics/table.py#L291-L332 | [
"def",
"discover_remote_results",
"(",
"self",
",",
"response",
",",
"name",
")",
":",
"host",
",",
"port",
"=",
"response",
".",
"source",
"if",
"response",
".",
"code",
"==",
"defines",
".",
"Codes",
".",
"CONTENT",
".",
"number",
":",
"resource",
"=",... |
Returns a metric list from the AWS CloudWatch service may return None if no metric exists | def __get_aws_metric ( table_name , lookback_window_start , lookback_period , metric_name ) : try : now = datetime . utcnow ( ) start_time = now - timedelta ( minutes = lookback_window_start ) end_time = now - timedelta ( minutes = lookback_window_start - lookback_period ) return cloudwatch_connection . get_metric_statistics ( period = lookback_period * 60 , start_time = start_time , end_time = end_time , metric_name = metric_name , namespace = 'AWS/DynamoDB' , statistics = [ 'Sum' ] , dimensions = { 'TableName' : table_name } , unit = 'Count' ) except BotoServerError as error : logger . error ( 'Unknown boto error. Status: "{0}". ' 'Reason: "{1}". Message: {2}' . format ( error . status , error . reason , error . message ) ) raise | 251,032 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/statistics/table.py#L340-L378 | [
"def",
"restrict_input_to_index",
"(",
"df_or_dict",
",",
"column_id",
",",
"index",
")",
":",
"if",
"isinstance",
"(",
"df_or_dict",
",",
"pd",
".",
"DataFrame",
")",
":",
"df_or_dict_restricted",
"=",
"df_or_dict",
"[",
"df_or_dict",
"[",
"column_id",
"]",
"... |
Ensure that provisioning is correct for Global Secondary Indexes | def ensure_provisioning ( table_name , table_key , gsi_name , gsi_key , num_consec_read_checks , num_consec_write_checks ) : if get_global_option ( 'circuit_breaker_url' ) or get_gsi_option ( table_key , gsi_key , 'circuit_breaker_url' ) : if circuit_breaker . is_open ( table_name , table_key , gsi_name , gsi_key ) : logger . warning ( 'Circuit breaker is OPEN!' ) return ( 0 , 0 ) logger . info ( '{0} - Will ensure provisioning for global secondary index {1}' . format ( table_name , gsi_name ) ) # Handle throughput alarm checks __ensure_provisioning_alarm ( table_name , table_key , gsi_name , gsi_key ) try : read_update_needed , updated_read_units , num_consec_read_checks = __ensure_provisioning_reads ( table_name , table_key , gsi_name , gsi_key , num_consec_read_checks ) write_update_needed , updated_write_units , num_consec_write_checks = __ensure_provisioning_writes ( table_name , table_key , gsi_name , gsi_key , num_consec_write_checks ) if read_update_needed : num_consec_read_checks = 0 if write_update_needed : num_consec_write_checks = 0 # Handle throughput updates if read_update_needed or write_update_needed : logger . info ( '{0} - GSI: {1} - Changing provisioning to {2:d} ' 'read units and {3:d} write units' . format ( table_name , gsi_name , int ( updated_read_units ) , int ( updated_write_units ) ) ) __update_throughput ( table_name , table_key , gsi_name , gsi_key , updated_read_units , updated_write_units ) else : logger . info ( '{0} - GSI: {1} - No need to change provisioning' . format ( table_name , gsi_name ) ) except JSONResponseError : raise except BotoServerError : raise return num_consec_read_checks , num_consec_write_checks | 251,033 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/core/gsi.py#L13-L93 | [
"def",
"content",
"(",
"self",
")",
":",
"if",
"not",
"getattr",
"(",
"self",
",",
"'_content'",
",",
"False",
")",
":",
"query_params",
"=",
"{",
"'prop'",
":",
"'extracts|revisions'",
",",
"'explaintext'",
":",
"''",
",",
"'rvprop'",
":",
"'ids'",
"}",... |
Update throughput on the GSI | def __update_throughput ( table_name , table_key , gsi_name , gsi_key , read_units , write_units ) : try : current_ru = dynamodb . get_provisioned_gsi_read_units ( table_name , gsi_name ) current_wu = dynamodb . get_provisioned_gsi_write_units ( table_name , gsi_name ) except JSONResponseError : raise # Check table status try : gsi_status = dynamodb . get_gsi_status ( table_name , gsi_name ) except JSONResponseError : raise logger . debug ( '{0} - GSI: {1} - GSI status is {2}' . format ( table_name , gsi_name , gsi_status ) ) if gsi_status != 'ACTIVE' : logger . warning ( '{0} - GSI: {1} - Not performing throughput changes when GSI ' 'status is {2}' . format ( table_name , gsi_name , gsi_status ) ) return # If this setting is True, we will only scale down when # BOTH reads AND writes are low if get_gsi_option ( table_key , gsi_key , 'always_decrease_rw_together' ) : read_units , write_units = __calculate_always_decrease_rw_values ( table_name , gsi_name , read_units , current_ru , write_units , current_wu ) if read_units == current_ru and write_units == current_wu : logger . info ( '{0} - GSI: {1} - No changes to perform' . format ( table_name , gsi_name ) ) return dynamodb . update_gsi_provisioning ( table_name , table_key , gsi_name , gsi_key , int ( read_units ) , int ( write_units ) ) | 251,034 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/core/gsi.py#L1027-L1088 | [
"def",
"diff",
"(",
"vault_client",
",",
"opt",
")",
":",
"if",
"opt",
".",
"thaw_from",
":",
"opt",
".",
"secrets",
"=",
"tempfile",
".",
"mkdtemp",
"(",
"'aomi-thaw'",
")",
"auto_thaw",
"(",
"vault_client",
",",
"opt",
")",
"ctx",
"=",
"Context",
"."... |
Checks whether the circuit breaker is open | def is_open ( table_name = None , table_key = None , gsi_name = None , gsi_key = None ) : logger . debug ( 'Checking circuit breaker status' ) # Parse the URL to make sure it is OK pattern = re . compile ( r'^(?P<scheme>http(s)?://)' r'((?P<username>.+):(?P<password>.+)@){0,1}' r'(?P<url>.*)$' ) url = timeout = None if gsi_name : url = get_gsi_option ( table_key , gsi_key , 'circuit_breaker_url' ) timeout = get_gsi_option ( table_key , gsi_key , 'circuit_breaker_timeout' ) elif table_name : url = get_table_option ( table_key , 'circuit_breaker_url' ) timeout = get_table_option ( table_key , 'circuit_breaker_timeout' ) if not url : url = get_global_option ( 'circuit_breaker_url' ) timeout = get_global_option ( 'circuit_breaker_timeout' ) match = pattern . match ( url ) if not match : logger . error ( 'Malformatted URL: {0}' . format ( url ) ) sys . exit ( 1 ) use_basic_auth = False if match . group ( 'username' ) and match . group ( 'password' ) : use_basic_auth = True # Make the actual URL to call auth = ( ) if use_basic_auth : url = '{scheme}{url}' . format ( scheme = match . group ( 'scheme' ) , url = match . group ( 'url' ) ) auth = ( match . group ( 'username' ) , match . group ( 'password' ) ) headers = { } if table_name : headers [ "x-table-name" ] = table_name if gsi_name : headers [ "x-gsi-name" ] = gsi_name # Make the actual request try : response = requests . get ( url , auth = auth , timeout = timeout / 1000.00 , headers = headers ) if int ( response . status_code ) >= 200 and int ( response . status_code ) < 300 : logger . info ( 'Circuit breaker is closed' ) return False else : logger . warning ( 'Circuit breaker returned with status code {0:d}' . format ( response . status_code ) ) except requests . exceptions . SSLError as error : logger . warning ( 'Circuit breaker: {0}' . format ( error ) ) except requests . exceptions . Timeout as error : logger . warning ( 'Circuit breaker: {0}' . format ( error ) ) except requests . exceptions . ConnectionError as error : logger . warning ( 'Circuit breaker: {0}' . format ( error ) ) except requests . exceptions . HTTPError as error : logger . warning ( 'Circuit breaker: {0}' . format ( error ) ) except requests . exceptions . TooManyRedirects as error : logger . warning ( 'Circuit breaker: {0}' . format ( error ) ) except Exception as error : logger . error ( 'Unhandled exception: {0}' . format ( error ) ) logger . error ( 'Please file a bug at ' 'https://github.com/sebdah/dynamic-dynamodb/issues' ) return True | 251,035 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/core/circuit_breaker.py#L13-L97 | [
"def",
"generate_http_manifest",
"(",
"self",
")",
":",
"base_path",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"self",
".",
"translate_path",
"(",
"self",
".",
"path",
")",
")",
"self",
".",
"dataset",
"=",
"dtoolcore",
".",
"DataSet",
".",
"from_uri"... |
Ensure connection to CloudWatch | def __get_connection_cloudwatch ( ) : region = get_global_option ( 'region' ) try : if ( get_global_option ( 'aws_access_key_id' ) and get_global_option ( 'aws_secret_access_key' ) ) : logger . debug ( 'Authenticating to CloudWatch using ' 'credentials in configuration file' ) connection = cloudwatch . connect_to_region ( region , aws_access_key_id = get_global_option ( 'aws_access_key_id' ) , aws_secret_access_key = get_global_option ( 'aws_secret_access_key' ) ) else : logger . debug ( 'Authenticating using boto\'s authentication handler' ) connection = cloudwatch . connect_to_region ( region ) except Exception as err : logger . error ( 'Failed connecting to CloudWatch: {0}' . format ( err ) ) logger . error ( 'Please report an issue at: ' 'https://github.com/sebdah/dynamic-dynamodb/issues' ) raise logger . debug ( 'Connected to CloudWatch in {0}' . format ( region ) ) return connection | 251,036 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/aws/cloudwatch.py#L9-L36 | [
"def",
"accept_freeware_license",
"(",
")",
":",
"ntab",
"=",
"3",
"if",
"version",
"(",
")",
".",
"startswith",
"(",
"'6.6.'",
")",
"else",
"2",
"for",
"_",
"in",
"range",
"(",
"ntab",
")",
":",
"EasyProcess",
"(",
"'xdotool key KP_Tab'",
")",
".",
"c... |
Get a set of tables and gsis and their configuration keys | def get_tables_and_gsis ( ) : table_names = set ( ) configured_tables = get_configured_tables ( ) not_used_tables = set ( configured_tables ) # Add regexp table names for table_instance in list_tables ( ) : for key_name in configured_tables : try : if re . match ( key_name , table_instance . table_name ) : logger . debug ( "Table {0} match with config key {1}" . format ( table_instance . table_name , key_name ) ) # Notify users about regexps that match multiple tables if table_instance . table_name in [ x [ 0 ] for x in table_names ] : logger . warning ( 'Table {0} matches more than one regexp in config, ' 'skipping this match: "{1}"' . format ( table_instance . table_name , key_name ) ) else : table_names . add ( ( table_instance . table_name , key_name ) ) not_used_tables . discard ( key_name ) else : logger . debug ( "Table {0} did not match with config key {1}" . format ( table_instance . table_name , key_name ) ) except re . error : logger . error ( 'Invalid regular expression: "{0}"' . format ( key_name ) ) sys . exit ( 1 ) if not_used_tables : logger . warning ( 'No tables matching the following configured ' 'tables found: {0}' . format ( ', ' . join ( not_used_tables ) ) ) return sorted ( table_names ) | 251,037 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/aws/dynamodb.py#L21-L65 | [
"def",
"send_request",
"(",
"self",
",",
"request",
")",
":",
"assert",
"self",
".",
"_handshake_performed",
",",
"\"Perform a handshake first.\"",
"assert",
"request",
".",
"id",
"not",
"in",
"self",
".",
"_outbound_pending_call",
",",
"(",
"\"Message ID '%d' alrea... |
Return list of DynamoDB tables available from AWS | def list_tables ( ) : tables = [ ] try : table_list = DYNAMODB_CONNECTION . list_tables ( ) while True : for table_name in table_list [ u'TableNames' ] : tables . append ( get_table ( table_name ) ) if u'LastEvaluatedTableName' in table_list : table_list = DYNAMODB_CONNECTION . list_tables ( table_list [ u'LastEvaluatedTableName' ] ) else : break except DynamoDBResponseError as error : dynamodb_error = error . body [ '__type' ] . rsplit ( '#' , 1 ) [ 1 ] if dynamodb_error == 'ResourceNotFoundException' : logger . error ( 'No tables found' ) elif dynamodb_error == 'AccessDeniedException' : logger . debug ( 'Your AWS API keys lack access to listing tables. ' 'That is an issue if you are trying to use regular ' 'expressions in your table configuration.' ) elif dynamodb_error == 'UnrecognizedClientException' : logger . error ( 'Invalid security token. Are your AWS API keys correct?' ) else : logger . error ( ( 'Unhandled exception: {0}: {1}. ' 'Please file a bug report at ' 'https://github.com/sebdah/dynamic-dynamodb/issues' ) . format ( dynamodb_error , error . body [ 'message' ] ) ) except JSONResponseError as error : logger . error ( 'Communication error: {0}' . format ( error ) ) sys . exit ( 1 ) return tables | 251,038 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/aws/dynamodb.py#L214-L260 | [
"def",
"agp",
"(",
"args",
")",
":",
"from",
"jcvi",
".",
"formats",
".",
"base",
"import",
"DictFile",
"p",
"=",
"OptionParser",
"(",
"agp",
".",
"__doc__",
")",
"opts",
",",
"args",
"=",
"p",
".",
"parse_args",
"(",
"args",
")",
"if",
"len",
"(",... |
Returns a list of GSIs for the given table | def table_gsis ( table_name ) : try : desc = DYNAMODB_CONNECTION . describe_table ( table_name ) [ u'Table' ] except JSONResponseError : raise if u'GlobalSecondaryIndexes' in desc : return desc [ u'GlobalSecondaryIndexes' ] return [ ] | 251,039 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/aws/dynamodb.py#L602-L617 | [
"def",
"_restart_session",
"(",
"self",
",",
"session",
")",
":",
"# remove old session key, if socket is None, that means the",
"# session was closed by user and there is no need to restart.",
"if",
"session",
".",
"socket",
"is",
"not",
"None",
":",
"self",
".",
"log",
".... |
Ensure connection to DynamoDB | def __get_connection_dynamodb ( retries = 3 ) : connected = False region = get_global_option ( 'region' ) while not connected : if ( get_global_option ( 'aws_access_key_id' ) and get_global_option ( 'aws_secret_access_key' ) ) : logger . debug ( 'Authenticating to DynamoDB using ' 'credentials in configuration file' ) connection = dynamodb2 . connect_to_region ( region , aws_access_key_id = get_global_option ( 'aws_access_key_id' ) , aws_secret_access_key = get_global_option ( 'aws_secret_access_key' ) ) else : logger . debug ( 'Authenticating using boto\'s authentication handler' ) connection = dynamodb2 . connect_to_region ( region ) if not connection : if retries == 0 : logger . error ( 'Failed to connect to DynamoDB. Giving up.' ) raise else : logger . error ( 'Failed to connect to DynamoDB. Retrying in 5 seconds' ) retries -= 1 time . sleep ( 5 ) else : connected = True logger . debug ( 'Connected to DynamoDB in {0}' . format ( region ) ) return connection | 251,040 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/aws/dynamodb.py#L620-L658 | [
"def",
"from_string",
"(",
"contents",
")",
":",
"if",
"contents",
"[",
"-",
"1",
"]",
"!=",
"\"\\n\"",
":",
"contents",
"+=",
"\"\\n\"",
"white_space",
"=",
"r\"[ \\t\\r\\f\\v]\"",
"natoms_line",
"=",
"white_space",
"+",
"r\"*\\d+\"",
"+",
"white_space",
"+",... |
Checks that the current time is within the maintenance window | def __is_gsi_maintenance_window ( table_name , gsi_name , maintenance_windows ) : # Example string '00:00-01:00,10:00-11:00' maintenance_window_list = [ ] for window in maintenance_windows . split ( ',' ) : try : start , end = window . split ( '-' , 1 ) except ValueError : logger . error ( '{0} - GSI: {1} - ' 'Malformatted maintenance window' . format ( table_name , gsi_name ) ) return False maintenance_window_list . append ( ( start , end ) ) now = datetime . datetime . utcnow ( ) . strftime ( '%H%M' ) for maintenance_window in maintenance_window_list : start = '' . join ( maintenance_window [ 0 ] . split ( ':' ) ) end = '' . join ( maintenance_window [ 1 ] . split ( ':' ) ) if now >= start and now <= end : return True return False | 251,041 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/aws/dynamodb.py#L661-L692 | [
"def",
"unbind",
"(",
"self",
",",
"devices_to_unbind",
")",
":",
"if",
"self",
".",
"entity_api_key",
"==",
"\"\"",
":",
"return",
"{",
"'status'",
":",
"'failure'",
",",
"'response'",
":",
"'No API key found in request'",
"}",
"url",
"=",
"self",
".",
"bas... |
Publish a notification for a specific GSI | def publish_gsi_notification ( table_key , gsi_key , message , message_types , subject = None ) : topic = get_gsi_option ( table_key , gsi_key , 'sns_topic_arn' ) if not topic : return for message_type in message_types : if ( message_type in get_gsi_option ( table_key , gsi_key , 'sns_message_types' ) ) : __publish ( topic , message , subject ) return | 251,042 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/aws/sns.py#L11-L40 | [
"def",
"_ar_matrix",
"(",
"self",
")",
":",
"Y",
"=",
"np",
".",
"array",
"(",
"self",
".",
"data",
"[",
"self",
".",
"max_lag",
":",
"self",
".",
"data",
".",
"shape",
"[",
"0",
"]",
"]",
")",
"X",
"=",
"self",
".",
"data",
"[",
"(",
"self",... |
Publish a notification for a specific table | def publish_table_notification ( table_key , message , message_types , subject = None ) : topic = get_table_option ( table_key , 'sns_topic_arn' ) if not topic : return for message_type in message_types : if message_type in get_table_option ( table_key , 'sns_message_types' ) : __publish ( topic , message , subject ) return | 251,043 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/aws/sns.py#L43-L68 | [
"def",
"start_wrap_console",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"_wrap_console",
"or",
"self",
".",
"_console_type",
"!=",
"\"telnet\"",
":",
"return",
"remaining_trial",
"=",
"60",
"while",
"True",
":",
"try",
":",
"(",
"reader",
",",
"write... |
Publish a message to a SNS topic | def __publish ( topic , message , subject = None ) : try : SNS_CONNECTION . publish ( topic = topic , message = message , subject = subject ) logger . info ( 'Sent SNS notification to {0}' . format ( topic ) ) except BotoServerError as error : logger . error ( 'Problem sending SNS notification: {0}' . format ( error . message ) ) return | 251,044 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/aws/sns.py#L71-L89 | [
"def",
"_openResources",
"(",
"self",
")",
":",
"try",
":",
"rate",
",",
"data",
"=",
"scipy",
".",
"io",
".",
"wavfile",
".",
"read",
"(",
"self",
".",
"_fileName",
",",
"mmap",
"=",
"True",
")",
"except",
"Exception",
"as",
"ex",
":",
"logger",
"... |
Ensure connection to SNS | def __get_connection_SNS ( ) : region = get_global_option ( 'region' ) try : if ( get_global_option ( 'aws_access_key_id' ) and get_global_option ( 'aws_secret_access_key' ) ) : logger . debug ( 'Authenticating to SNS using ' 'credentials in configuration file' ) connection = sns . connect_to_region ( region , aws_access_key_id = get_global_option ( 'aws_access_key_id' ) , aws_secret_access_key = get_global_option ( 'aws_secret_access_key' ) ) else : logger . debug ( 'Authenticating using boto\'s authentication handler' ) connection = sns . connect_to_region ( region ) except Exception as err : logger . error ( 'Failed connecting to SNS: {0}' . format ( err ) ) logger . error ( 'Please report an issue at: ' 'https://github.com/sebdah/dynamic-dynamodb/issues' ) raise logger . debug ( 'Connected to SNS in {0}' . format ( region ) ) return connection | 251,045 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/aws/sns.py#L92-L121 | [
"def",
"group_experiments_greedy",
"(",
"tomo_expt",
":",
"TomographyExperiment",
")",
":",
"diag_sets",
"=",
"_max_tpb_overlap",
"(",
"tomo_expt",
")",
"grouped_expt_settings_list",
"=",
"list",
"(",
"diag_sets",
".",
"values",
"(",
")",
")",
"grouped_tomo_expt",
"... |
Calculate values for always - decrease - rw - together | def __calculate_always_decrease_rw_values ( table_name , read_units , provisioned_reads , write_units , provisioned_writes ) : if read_units <= provisioned_reads and write_units <= provisioned_writes : return ( read_units , write_units ) if read_units < provisioned_reads : logger . info ( '{0} - Reads could be decreased, but we are waiting for ' 'writes to get lower than the threshold before ' 'scaling down' . format ( table_name ) ) read_units = provisioned_reads elif write_units < provisioned_writes : logger . info ( '{0} - Writes could be decreased, but we are waiting for ' 'reads to get lower than the threshold before ' 'scaling down' . format ( table_name ) ) write_units = provisioned_writes return ( read_units , write_units ) | 251,046 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/core/table.py#L81-L121 | [
"def",
"postprocess_segments",
"(",
"self",
")",
":",
"# make segs a list of mask arrays, it's easier to store",
"# as there is a hdf5 equivalent",
"for",
"iseg",
",",
"seg",
"in",
"enumerate",
"(",
"self",
".",
"segs",
")",
":",
"mask",
"=",
"np",
".",
"zeros",
"("... |
Update throughput on the DynamoDB table | def __update_throughput ( table_name , key_name , read_units , write_units ) : try : current_ru = dynamodb . get_provisioned_table_read_units ( table_name ) current_wu = dynamodb . get_provisioned_table_write_units ( table_name ) except JSONResponseError : raise # Check table status try : table_status = dynamodb . get_table_status ( table_name ) except JSONResponseError : raise logger . debug ( '{0} - Table status is {1}' . format ( table_name , table_status ) ) if table_status != 'ACTIVE' : logger . warning ( '{0} - Not performing throughput changes when table ' 'is {1}' . format ( table_name , table_status ) ) return # If this setting is True, we will only scale down when # BOTH reads AND writes are low if get_table_option ( key_name , 'always_decrease_rw_together' ) : read_units , write_units = __calculate_always_decrease_rw_values ( table_name , read_units , current_ru , write_units , current_wu ) if read_units == current_ru and write_units == current_wu : logger . info ( '{0} - No changes to perform' . format ( table_name ) ) return dynamodb . update_table_provisioning ( table_name , key_name , int ( read_units ) , int ( write_units ) ) | 251,047 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/core/table.py#L897-L945 | [
"def",
"from_string",
"(",
"contents",
")",
":",
"if",
"contents",
"[",
"-",
"1",
"]",
"!=",
"\"\\n\"",
":",
"contents",
"+=",
"\"\\n\"",
"white_space",
"=",
"r\"[ \\t\\r\\f\\v]\"",
"natoms_line",
"=",
"white_space",
"+",
"r\"*\\d+\"",
"+",
"white_space",
"+",... |
Get the configuration from command line and config files | def get_configuration ( ) : # This is the dict we will return configuration = { 'global' : { } , 'logging' : { } , 'tables' : ordereddict ( ) } # Read the command line options cmd_line_options = command_line_parser . parse ( ) # If a configuration file is specified, read that as well conf_file_options = None if 'config' in cmd_line_options : conf_file_options = config_file_parser . parse ( cmd_line_options [ 'config' ] ) # Extract global config configuration [ 'global' ] = __get_global_options ( cmd_line_options , conf_file_options ) # Extract logging config configuration [ 'logging' ] = __get_logging_options ( cmd_line_options , conf_file_options ) # Extract table configuration # If the --table cmd line option is set, it indicates that only table # options from the command line should be used if 'table_name' in cmd_line_options : configuration [ 'tables' ] = __get_cmd_table_options ( cmd_line_options ) else : configuration [ 'tables' ] = __get_config_table_options ( conf_file_options ) # Ensure some basic rules __check_gsi_rules ( configuration ) __check_logging_rules ( configuration ) __check_table_rules ( configuration ) return configuration | 251,048 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/config/__init__.py#L162-L203 | [
"def",
"local_regon_checksum",
"(",
"digits",
")",
":",
"weights_for_check_digit",
"=",
"[",
"2",
",",
"4",
",",
"8",
",",
"5",
",",
"0",
",",
"9",
",",
"7",
",",
"3",
",",
"6",
",",
"1",
",",
"2",
",",
"4",
",",
"8",
"]",
"check_digit",
"=",
... |
Get all table options from the command line | def __get_cmd_table_options ( cmd_line_options ) : table_name = cmd_line_options [ 'table_name' ] options = { table_name : { } } for option in DEFAULT_OPTIONS [ 'table' ] . keys ( ) : options [ table_name ] [ option ] = DEFAULT_OPTIONS [ 'table' ] [ option ] if option in cmd_line_options : options [ table_name ] [ option ] = cmd_line_options [ option ] return options | 251,049 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/config/__init__.py#L206-L222 | [
"def",
"face_index",
"(",
"vertices",
")",
":",
"new_verts",
"=",
"[",
"]",
"face_indices",
"=",
"[",
"]",
"for",
"wall",
"in",
"vertices",
":",
"face_wall",
"=",
"[",
"]",
"for",
"vert",
"in",
"wall",
":",
"if",
"new_verts",
":",
"if",
"not",
"np",
... |
Get all table options from the config file | def __get_config_table_options ( conf_file_options ) : options = ordereddict ( ) if not conf_file_options : return options for table_name in conf_file_options [ 'tables' ] : options [ table_name ] = { } # Regular table options for option in DEFAULT_OPTIONS [ 'table' ] . keys ( ) : options [ table_name ] [ option ] = DEFAULT_OPTIONS [ 'table' ] [ option ] if option not in conf_file_options [ 'tables' ] [ table_name ] : continue if option == 'sns_message_types' : try : raw_list = conf_file_options [ 'tables' ] [ table_name ] [ option ] options [ table_name ] [ option ] = [ i . strip ( ) for i in raw_list . split ( ',' ) ] except : print ( 'Error parsing the "sns-message-types" ' 'option: {0}' . format ( conf_file_options [ 'tables' ] [ table_name ] [ option ] ) ) else : options [ table_name ] [ option ] = conf_file_options [ 'tables' ] [ table_name ] [ option ] # GSI specific options if 'gsis' in conf_file_options [ 'tables' ] [ table_name ] : for gsi_name in conf_file_options [ 'tables' ] [ table_name ] [ 'gsis' ] : for option in DEFAULT_OPTIONS [ 'gsi' ] . keys ( ) : opt = DEFAULT_OPTIONS [ 'gsi' ] [ option ] if 'gsis' not in options [ table_name ] : options [ table_name ] [ 'gsis' ] = { } if gsi_name not in options [ table_name ] [ 'gsis' ] : options [ table_name ] [ 'gsis' ] [ gsi_name ] = { } if ( option not in conf_file_options [ 'tables' ] [ table_name ] [ 'gsis' ] [ gsi_name ] ) : options [ table_name ] [ 'gsis' ] [ gsi_name ] [ option ] = opt continue if option == 'sns_message_types' : try : raw_list = conf_file_options [ 'tables' ] [ table_name ] [ 'gsis' ] [ gsi_name ] [ option ] opt = [ i . strip ( ) for i in raw_list . split ( ',' ) ] except : print ( 'Error parsing the "sns-message-types" ' 'option: {0}' . format ( conf_file_options [ 'tables' ] [ table_name ] [ 'gsis' ] [ gsi_name ] [ option ] ) ) else : opt = conf_file_options [ 'tables' ] [ table_name ] [ 'gsis' ] [ gsi_name ] [ option ] options [ table_name ] [ 'gsis' ] [ gsi_name ] [ option ] = opt return options | 251,050 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/config/__init__.py#L225-L296 | [
"def",
"_compile_arithmetic_expression",
"(",
"self",
",",
"expr",
":",
"Expression",
",",
"scope",
":",
"Dict",
"[",
"str",
",",
"TensorFluent",
"]",
",",
"batch_size",
":",
"Optional",
"[",
"int",
"]",
"=",
"None",
",",
"noise",
":",
"Optional",
"[",
"... |
Get all global options | def __get_global_options ( cmd_line_options , conf_file_options = None ) : options = { } for option in DEFAULT_OPTIONS [ 'global' ] . keys ( ) : options [ option ] = DEFAULT_OPTIONS [ 'global' ] [ option ] if conf_file_options and option in conf_file_options : options [ option ] = conf_file_options [ option ] if cmd_line_options and option in cmd_line_options : options [ option ] = cmd_line_options [ option ] return options | 251,051 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/config/__init__.py#L299-L319 | [
"def",
"_openResources",
"(",
"self",
")",
":",
"try",
":",
"rate",
",",
"data",
"=",
"scipy",
".",
"io",
".",
"wavfile",
".",
"read",
"(",
"self",
".",
"_fileName",
",",
"mmap",
"=",
"True",
")",
"except",
"Exception",
"as",
"ex",
":",
"logger",
"... |
Get all logging options | def __get_logging_options ( cmd_line_options , conf_file_options = None ) : options = { } for option in DEFAULT_OPTIONS [ 'logging' ] . keys ( ) : options [ option ] = DEFAULT_OPTIONS [ 'logging' ] [ option ] if conf_file_options and option in conf_file_options : options [ option ] = conf_file_options [ option ] if cmd_line_options and option in cmd_line_options : options [ option ] = cmd_line_options [ option ] return options | 251,052 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/config/__init__.py#L322-L342 | [
"def",
"_get_purecn_dx_files",
"(",
"paired",
",",
"out",
")",
":",
"out_base",
"=",
"\"%s-dx\"",
"%",
"utils",
".",
"splitext_plus",
"(",
"out",
"[",
"\"rds\"",
"]",
")",
"[",
"0",
"]",
"all_files",
"=",
"[",
"]",
"for",
"key",
",",
"ext",
"in",
"["... |
Check that the logging values are proper | def __check_logging_rules ( configuration ) : valid_log_levels = [ 'debug' , 'info' , 'warning' , 'error' ] if configuration [ 'logging' ] [ 'log_level' ] . lower ( ) not in valid_log_levels : print ( 'Log level must be one of {0}' . format ( ', ' . join ( valid_log_levels ) ) ) sys . exit ( 1 ) | 251,053 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/config/__init__.py#L505-L516 | [
"def",
"createAltHistoryPlot",
"(",
"self",
")",
":",
"self",
".",
"altHistRect",
"=",
"patches",
".",
"Rectangle",
"(",
"(",
"self",
".",
"leftPos",
"+",
"(",
"self",
".",
"vertSize",
"/",
"10.0",
")",
",",
"-",
"0.25",
")",
",",
"0.5",
",",
"0.5",
... |
Determines if the currently consumed capacity is over the proposed capacity for this table | def is_consumed_over_proposed ( current_provisioning , proposed_provisioning , consumed_units_percent ) : consumption_based_current_provisioning = int ( math . ceil ( current_provisioning * ( consumed_units_percent / 100 ) ) ) return consumption_based_current_provisioning > proposed_provisioning | 251,054 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/calculators.py#L342-L358 | [
"def",
"concatenate_json",
"(",
"source_folder",
",",
"destination_file",
")",
":",
"matches",
"=",
"[",
"]",
"for",
"root",
",",
"dirnames",
",",
"filenames",
"in",
"os",
".",
"walk",
"(",
"source_folder",
")",
":",
"for",
"filename",
"in",
"fnmatch",
"."... |
Get the minimum number of reads to current_provisioning | def __get_min_reads ( current_provisioning , min_provisioned_reads , log_tag ) : # Fallback value to ensure that we always have at least 1 read reads = 1 if min_provisioned_reads : reads = int ( min_provisioned_reads ) if reads > int ( current_provisioning * 2 ) : reads = int ( current_provisioning * 2 ) logger . debug ( '{0} - ' 'Cannot reach min-provisioned-reads as max scale up ' 'is 100% of current provisioning' . format ( log_tag ) ) logger . debug ( '{0} - Setting min provisioned reads to {1}' . format ( log_tag , min_provisioned_reads ) ) return reads | 251,055 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/calculators.py#L361-L389 | [
"def",
"weld_invert",
"(",
"array",
")",
":",
"obj_id",
",",
"weld_obj",
"=",
"create_weld_object",
"(",
"array",
")",
"weld_template",
"=",
"\"\"\"result(\n for({array},\n appender[bool],\n |b: appender[bool], i: i64, e: bool|\n if(e, merge(b, false), mer... |
Get the minimum number of writes to current_provisioning | def __get_min_writes ( current_provisioning , min_provisioned_writes , log_tag ) : # Fallback value to ensure that we always have at least 1 read writes = 1 if min_provisioned_writes : writes = int ( min_provisioned_writes ) if writes > int ( current_provisioning * 2 ) : writes = int ( current_provisioning * 2 ) logger . debug ( '{0} - ' 'Cannot reach min-provisioned-writes as max scale up ' 'is 100% of current provisioning' . format ( log_tag ) ) logger . debug ( '{0} - Setting min provisioned writes to {1}' . format ( log_tag , min_provisioned_writes ) ) return writes | 251,056 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/calculators.py#L392-L420 | [
"def",
"retract",
"(",
"self",
",",
"e",
",",
"a",
",",
"v",
")",
":",
"ta",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"ret",
"=",
"u\"[:db/retract %i :%s %s]\"",
"%",
"(",
"e",
",",
"a",
",",
"dump_edn_val",
"(",
"v",
")",
")",
"rs... |
Restart the daemon | def restart ( self , * args , * * kwargs ) : self . stop ( ) try : self . start ( * args , * * kwargs ) except IOError : raise | 251,057 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/daemon.py#L132-L138 | [
"def",
"_adapt_WSDateTime",
"(",
"dt",
")",
":",
"try",
":",
"ts",
"=",
"int",
"(",
"(",
"dt",
".",
"replace",
"(",
"tzinfo",
"=",
"pytz",
".",
"utc",
")",
"-",
"datetime",
"(",
"1970",
",",
"1",
",",
"1",
",",
"tzinfo",
"=",
"pytz",
".",
"utc"... |
Parse the section options | def __parse_options ( config_file , section , options ) : configuration = { } for option in options : try : if option . get ( 'type' ) == 'str' : configuration [ option . get ( 'key' ) ] = config_file . get ( section , option . get ( 'option' ) ) elif option . get ( 'type' ) == 'int' : try : configuration [ option . get ( 'key' ) ] = config_file . getint ( section , option . get ( 'option' ) ) except ValueError : print ( 'Error: Expected an integer value for {0}' . format ( option . get ( 'option' ) ) ) sys . exit ( 1 ) elif option . get ( 'type' ) == 'float' : try : configuration [ option . get ( 'key' ) ] = config_file . getfloat ( section , option . get ( 'option' ) ) except ValueError : print ( 'Error: Expected an float value for {0}' . format ( option . get ( 'option' ) ) ) sys . exit ( 1 ) elif option . get ( 'type' ) == 'bool' : try : configuration [ option . get ( 'key' ) ] = config_file . getboolean ( section , option . get ( 'option' ) ) except ValueError : print ( 'Error: Expected an boolean value for {0}' . format ( option . get ( 'option' ) ) ) sys . exit ( 1 ) elif option . get ( 'type' ) == 'dict' : configuration [ option . get ( 'key' ) ] = ast . literal_eval ( config_file . get ( section , option . get ( 'option' ) ) ) else : configuration [ option . get ( 'key' ) ] = config_file . get ( section , option . get ( 'option' ) ) except ConfigParser . NoOptionError : if option . get ( 'required' ) : print ( 'Missing [{0}] option "{1}" in configuration' . format ( section , option . get ( 'option' ) ) ) sys . exit ( 1 ) return configuration | 251,058 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/config/config_file_parser.py#L392-L453 | [
"def",
"recv_raw",
"(",
"self",
",",
"timeout",
",",
"opcodes",
",",
"*",
"*",
"kwargs",
")",
":",
"orig_timeout",
"=",
"self",
".",
"get_timeout",
"(",
"timeout",
")",
"timeout",
"=",
"orig_timeout",
"while",
"timeout",
">",
"0.0",
":",
"start",
"=",
... |
Main function called from dynamic - dynamodb | def main ( ) : try : if get_global_option ( 'show_config' ) : print json . dumps ( config . get_configuration ( ) , indent = 2 ) elif get_global_option ( 'daemon' ) : daemon = DynamicDynamoDBDaemon ( '{0}/dynamic-dynamodb.{1}.pid' . format ( get_global_option ( 'pid_file_dir' ) , get_global_option ( 'instance' ) ) ) if get_global_option ( 'daemon' ) == 'start' : logger . debug ( 'Starting daemon' ) try : daemon . start ( ) logger . info ( 'Daemon started' ) except IOError as error : logger . error ( 'Could not create pid file: {0}' . format ( error ) ) logger . error ( 'Daemon not started' ) elif get_global_option ( 'daemon' ) == 'stop' : logger . debug ( 'Stopping daemon' ) daemon . stop ( ) logger . info ( 'Daemon stopped' ) sys . exit ( 0 ) elif get_global_option ( 'daemon' ) == 'restart' : logger . debug ( 'Restarting daemon' ) daemon . restart ( ) logger . info ( 'Daemon restarted' ) elif get_global_option ( 'daemon' ) in [ 'foreground' , 'fg' ] : logger . debug ( 'Starting daemon in foreground' ) daemon . run ( ) logger . info ( 'Daemon started in foreground' ) else : print ( 'Valid options for --daemon are start, ' 'stop, restart, and foreground' ) sys . exit ( 1 ) else : if get_global_option ( 'run_once' ) : execute ( ) else : while True : execute ( ) except Exception as error : logger . exception ( error ) | 251,059 | https://github.com/sebdah/dynamic-dynamodb/blob/bfd0ca806b1c3301e724696de90ef0f973410493/dynamic_dynamodb/__init__.py#L56-L104 | [
"def",
"leave_swarm",
"(",
"force",
"=",
"bool",
")",
":",
"salt_return",
"=",
"{",
"}",
"__context__",
"[",
"'client'",
"]",
".",
"swarm",
".",
"leave",
"(",
"force",
"=",
"force",
")",
"output",
"=",
"__context__",
"[",
"'server_name'",
"]",
"+",
"' ... |
Generic byte decoding function for TVE parameters . | def decode_tve_parameter ( data ) : ( nontve , ) = struct . unpack ( nontve_header , data [ : nontve_header_len ] ) if nontve == 1023 : # customparameter ( size , ) = struct . unpack ( '!H' , data [ nontve_header_len : nontve_header_len + 2 ] ) ( subtype , ) = struct . unpack ( '!H' , data [ size - 4 : size - 2 ] ) param_name , param_fmt = ext_param_formats [ subtype ] ( unpacked , ) = struct . unpack ( param_fmt , data [ size - 2 : size ] ) return { param_name : unpacked } , size # decode the TVE field's header (1 bit "reserved" + 7-bit type) ( msgtype , ) = struct . unpack ( tve_header , data [ : tve_header_len ] ) if not msgtype & 0b10000000 : # not a TV-encoded param return None , 0 msgtype = msgtype & 0x7f try : param_name , param_fmt = tve_param_formats [ msgtype ] logger . debug ( 'found %s (type=%s)' , param_name , msgtype ) except KeyError : return None , 0 # decode the body nbytes = struct . calcsize ( param_fmt ) end = tve_header_len + nbytes try : unpacked = struct . unpack ( param_fmt , data [ tve_header_len : end ] ) return { param_name : unpacked } , end except struct . error : return None , 0 | 251,060 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp_decoder.py#L39-L74 | [
"def",
"bind",
"(",
"self",
",",
"container",
")",
":",
"# if there's already a matching bound instance, return that",
"shared",
"=",
"container",
".",
"shared_extensions",
".",
"get",
"(",
"self",
".",
"sharing_key",
")",
"if",
"shared",
":",
"return",
"shared",
... |
Get the long description from a file . | def read ( filename ) : fname = os . path . join ( here , filename ) with codecs . open ( fname , encoding = 'utf-8' ) as f : return f . read ( ) | 251,061 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/setup.py#L10-L16 | [
"def",
"get_form_kwargs",
"(",
"self",
")",
":",
"kwargs",
"=",
"super",
"(",
")",
".",
"get_form_kwargs",
"(",
")",
"if",
"self",
".",
"request",
".",
"method",
"==",
"'POST'",
":",
"data",
"=",
"copy",
"(",
"self",
".",
"request",
".",
"POST",
")",... |
Turns a sequence of bytes into a message dictionary . | def deserialize ( self ) : if self . msgbytes is None : raise LLRPError ( 'No message bytes to deserialize.' ) data = self . msgbytes msgtype , length , msgid = struct . unpack ( self . full_hdr_fmt , data [ : self . full_hdr_len ] ) ver = ( msgtype >> 10 ) & BITMASK ( 3 ) msgtype = msgtype & BITMASK ( 10 ) try : name = Message_Type2Name [ msgtype ] logger . debug ( 'deserializing %s command' , name ) decoder = Message_struct [ name ] [ 'decode' ] except KeyError : raise LLRPError ( 'Cannot find decoder for message type ' '{}' . format ( msgtype ) ) body = data [ self . full_hdr_len : length ] try : self . msgdict = { name : dict ( decoder ( body ) ) } self . msgdict [ name ] [ 'Ver' ] = ver self . msgdict [ name ] [ 'Type' ] = msgtype self . msgdict [ name ] [ 'ID' ] = msgid logger . debug ( 'done deserializing %s command' , name ) except ValueError : logger . exception ( 'Unable to decode body %s, %s' , body , decoder ( body ) ) except LLRPError : logger . exception ( 'Problem with %s message format' , name ) return '' return '' | 251,062 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L66-L97 | [
"def",
"is_period_current",
"(",
"self",
")",
":",
"return",
"self",
".",
"current_period_end",
">",
"timezone",
".",
"now",
"(",
")",
"or",
"(",
"self",
".",
"trial_end",
"and",
"self",
".",
"trial_end",
">",
"timezone",
".",
"now",
"(",
")",
")"
] |
Parse a reader configuration dictionary . | def parseReaderConfig ( self , confdict ) : logger . debug ( 'parseReaderConfig input: %s' , confdict ) conf = { } for k , v in confdict . items ( ) : if not k . startswith ( 'Parameter' ) : continue ty = v [ 'Type' ] data = v [ 'Data' ] vendor = None subtype = None try : vendor , subtype = v [ 'Vendor' ] , v [ 'Subtype' ] except KeyError : pass if ty == 1023 : if vendor == 25882 and subtype == 37 : tempc = struct . unpack ( '!H' , data ) [ 0 ] conf . update ( temperature = tempc ) else : conf [ ty ] = data return conf | 251,063 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L291-L326 | [
"def",
"moreland_adjusthue",
"(",
"msh",
",",
"m_unsat",
")",
":",
"if",
"msh",
"[",
"M",
"]",
">=",
"m_unsat",
":",
"return",
"msh",
"[",
"H",
"]",
"# \"Best we can do\"",
"hspin",
"=",
"(",
"msh",
"[",
"S",
"]",
"*",
"np",
".",
"sqrt",
"(",
"m_un... |
Parse a capabilities dictionary and adjust instance settings . | def parseCapabilities ( self , capdict ) : # check requested antenna set gdc = capdict [ 'GeneralDeviceCapabilities' ] max_ant = gdc [ 'MaxNumberOfAntennaSupported' ] if max ( self . antennas ) > max_ant : reqd = ',' . join ( map ( str , self . antennas ) ) avail = ',' . join ( map ( str , range ( 1 , max_ant + 1 ) ) ) errmsg = ( 'Invalid antenna set specified: requested={},' ' available={}; ignoring invalid antennas' . format ( reqd , avail ) ) raise ReaderConfigurationError ( errmsg ) logger . debug ( 'set antennas: %s' , self . antennas ) # parse available transmit power entries, set self.tx_power bandcap = capdict [ 'RegulatoryCapabilities' ] [ 'UHFBandCapabilities' ] self . tx_power_table = self . parsePowerTable ( bandcap ) logger . debug ( 'tx_power_table: %s' , self . tx_power_table ) self . setTxPower ( self . tx_power ) # parse list of reader's supported mode identifiers regcap = capdict [ 'RegulatoryCapabilities' ] modes = regcap [ 'UHFBandCapabilities' ] [ 'UHFRFModeTable' ] mode_list = [ modes [ k ] for k in sorted ( modes . keys ( ) , key = natural_keys ) ] # select a mode by matching available modes to requested parameters if self . mode_identifier is not None : logger . debug ( 'Setting mode from mode_identifier=%s' , self . mode_identifier ) try : mode = [ mo for mo in mode_list if mo [ 'ModeIdentifier' ] == self . mode_identifier ] [ 0 ] self . reader_mode = mode except IndexError : valid_modes = sorted ( mo [ 'ModeIdentifier' ] for mo in mode_list ) errstr = ( 'Invalid mode_identifier; valid mode_identifiers' ' are {}' . format ( valid_modes ) ) raise ReaderConfigurationError ( errstr ) # if we're trying to set Tari explicitly, but the selected mode doesn't # support the requested Tari, that's a configuration error. if self . reader_mode and self . tari : if self . reader_mode [ 'MinTari' ] < self . tari < self . reader_mode [ 'MaxTari' ] : logger . debug ( 'Overriding mode Tari %s with requested Tari %s' , self . reader_mode [ 'MaxTari' ] , self . tari ) else : errstr = ( 'Requested Tari {} is incompatible with selected ' 'mode {}' . format ( self . tari , self . reader_mode ) ) logger . info ( 'using reader mode: %s' , self . reader_mode ) | 251,064 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L328-L393 | [
"def",
"lockfile",
"(",
"path",
")",
":",
"with",
"genfile",
"(",
"path",
")",
"as",
"fd",
":",
"fcntl",
".",
"lockf",
"(",
"fd",
",",
"fcntl",
".",
"LOCK_EX",
")",
"yield",
"None"
] |
Add a ROSpec to the reader and enable it . | def startInventory ( self , proto = None , force_regen_rospec = False ) : if self . state == LLRPClient . STATE_INVENTORYING : logger . warn ( 'ignoring startInventory() while already inventorying' ) return None rospec = self . getROSpec ( force_new = force_regen_rospec ) [ 'ROSpec' ] logger . info ( 'starting inventory' ) # upside-down chain of callbacks: add, enable, start ROSpec # started_rospec = defer.Deferred() # started_rospec.addCallback(self._setState_wrapper, # LLRPClient.STATE_INVENTORYING) # started_rospec.addErrback(self.panic, 'START_ROSPEC failed') # logger.debug('made started_rospec') enabled_rospec = defer . Deferred ( ) enabled_rospec . addCallback ( self . _setState_wrapper , LLRPClient . STATE_INVENTORYING ) # enabled_rospec.addCallback(self.send_START_ROSPEC, rospec, # onCompletion=started_rospec) enabled_rospec . addErrback ( self . panic , 'ENABLE_ROSPEC failed' ) logger . debug ( 'made enabled_rospec' ) added_rospec = defer . Deferred ( ) added_rospec . addCallback ( self . send_ENABLE_ROSPEC , rospec , onCompletion = enabled_rospec ) added_rospec . addErrback ( self . panic , 'ADD_ROSPEC failed' ) logger . debug ( 'made added_rospec' ) self . send_ADD_ROSPEC ( rospec , onCompletion = added_rospec ) | 251,065 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L1038-L1069 | [
"def",
"destroy",
"(",
"name",
",",
"call",
"=",
"None",
")",
":",
"if",
"call",
"==",
"'function'",
":",
"raise",
"SaltCloudSystemExit",
"(",
"'The destroy action must be called with -d, --destroy, '",
"'-a or --action.'",
")",
"__utils__",
"[",
"'cloud.fire_event'",
... |
Delete all active ROSpecs . Return a Deferred that will be called when the DELETE_ROSPEC_RESPONSE comes back . | def stopPolitely ( self , disconnect = False ) : logger . info ( 'stopping politely' ) if disconnect : logger . info ( 'will disconnect when stopped' ) self . disconnecting = True self . sendMessage ( { 'DELETE_ACCESSSPEC' : { 'Ver' : 1 , 'Type' : 41 , 'ID' : 0 , 'AccessSpecID' : 0 # all AccessSpecs } } ) self . setState ( LLRPClient . STATE_SENT_DELETE_ACCESSSPEC ) d = defer . Deferred ( ) d . addCallback ( self . stopAllROSpecs ) d . addErrback ( self . panic , 'DELETE_ACCESSSPEC failed' ) self . _deferreds [ 'DELETE_ACCESSSPEC_RESPONSE' ] . append ( d ) return d | 251,066 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L1103-L1124 | [
"def",
"match",
"(",
"self",
",",
"string",
")",
":",
"m",
"=",
"self",
".",
"regex",
".",
"match",
"(",
"string",
")",
"if",
"m",
":",
"c",
"=",
"self",
".",
"type_converters",
"return",
"dict",
"(",
"(",
"k",
",",
"c",
"[",
"k",
"]",
"(",
"... |
Parse the transmit power table | def parsePowerTable ( uhfbandcap ) : bandtbl = { k : v for k , v in uhfbandcap . items ( ) if k . startswith ( 'TransmitPowerLevelTableEntry' ) } tx_power_table = [ 0 ] * ( len ( bandtbl ) + 1 ) for k , v in bandtbl . items ( ) : idx = v [ 'Index' ] tx_power_table [ idx ] = int ( v [ 'TransmitPowerValue' ] ) / 100.0 return tx_power_table | 251,067 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L1143-L1163 | [
"def",
"_http_call",
"(",
"the_url",
",",
"method",
",",
"authorization",
",",
"*",
"*",
"kw",
")",
":",
"params",
"=",
"None",
"boundary",
"=",
"None",
"if",
"method",
"==",
"_HTTP_UPLOAD",
":",
"# fix sina upload url:",
"the_url",
"=",
"the_url",
".",
"r... |
Validates tx_power against self . tx_power_table | def get_tx_power ( self , tx_power ) : if not self . tx_power_table : logger . warn ( 'get_tx_power(): tx_power_table is empty!' ) return { } logger . debug ( 'requested tx_power: %s' , tx_power ) min_power = self . tx_power_table . index ( min ( self . tx_power_table ) ) max_power = self . tx_power_table . index ( max ( self . tx_power_table ) ) ret = { } for antid , tx_power in tx_power . items ( ) : if tx_power == 0 : # tx_power = 0 means max power max_power_dbm = max ( self . tx_power_table ) tx_power = self . tx_power_table . index ( max_power_dbm ) ret [ antid ] = ( tx_power , max_power_dbm ) try : power_dbm = self . tx_power_table [ tx_power ] ret [ antid ] = ( tx_power , power_dbm ) except IndexError : raise LLRPError ( 'Invalid tx_power for antenna {}: ' 'requested={}, min_available={}, ' 'max_available={}' . format ( antid , self . tx_power , min_power , max_power ) ) return ret | 251,068 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L1165-L1199 | [
"def",
"dateDict",
"(",
"date",
",",
"prefix",
"=",
"\"\"",
")",
":",
"res",
"=",
"{",
"}",
"res",
"[",
"prefix",
"+",
"\"a\"",
"]",
"=",
"date",
".",
"strftime",
"(",
"\"%a\"",
")",
"res",
"[",
"prefix",
"+",
"\"A\"",
"]",
"=",
"date",
".",
"s... |
Set the transmission power for one or more antennas . | def setTxPower ( self , tx_power ) : tx_pow_validated = self . get_tx_power ( tx_power ) logger . debug ( 'tx_pow_validated: %s' , tx_pow_validated ) needs_update = False for ant , ( tx_pow_idx , tx_pow_dbm ) in tx_pow_validated . items ( ) : if self . tx_power [ ant ] != tx_pow_idx : self . tx_power [ ant ] = tx_pow_idx needs_update = True logger . debug ( 'tx_power for antenna %s: %s (%s dBm)' , ant , tx_pow_idx , tx_pow_dbm ) if needs_update and self . state == LLRPClient . STATE_INVENTORYING : logger . debug ( 'changing tx power; will stop politely, then resume' ) d = self . stopPolitely ( ) d . addCallback ( self . startInventory , force_regen_rospec = True ) | 251,069 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L1201-L1220 | [
"def",
"get_hostname",
"(",
"cls",
",",
"container_name",
",",
"client_name",
"=",
"None",
")",
":",
"base_name",
"=",
"container_name",
"for",
"old",
",",
"new",
"in",
"cls",
".",
"hostname_replace",
":",
"base_name",
"=",
"base_name",
".",
"replace",
"(",
... |
Pause an inventory operation for a set amount of time . | def pause ( self , duration_seconds = 0 , force = False , force_regen_rospec = False ) : logger . debug ( 'pause(%s)' , duration_seconds ) if self . state != LLRPClient . STATE_INVENTORYING : if not force : logger . info ( 'ignoring pause(); not inventorying (state==%s)' , self . getStateName ( self . state ) ) return None else : logger . info ( 'forcing pause()' ) if duration_seconds : logger . info ( 'pausing for %s seconds' , duration_seconds ) rospec = self . getROSpec ( force_new = force_regen_rospec ) [ 'ROSpec' ] self . sendMessage ( { 'DISABLE_ROSPEC' : { 'Ver' : 1 , 'Type' : 25 , 'ID' : 0 , 'ROSpecID' : rospec [ 'ROSpecID' ] } } ) self . setState ( LLRPClient . STATE_PAUSING ) d = defer . Deferred ( ) d . addCallback ( self . _setState_wrapper , LLRPClient . STATE_PAUSED ) d . addErrback ( self . complain , 'pause() failed' ) self . _deferreds [ 'DISABLE_ROSPEC_RESPONSE' ] . append ( d ) if duration_seconds > 0 : startAgain = task . deferLater ( reactor , duration_seconds , lambda : None ) startAgain . addCallback ( lambda _ : self . resume ( ) ) return d | 251,070 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L1222-L1257 | [
"def",
"files_comments_delete",
"(",
"self",
",",
"*",
",",
"file",
":",
"str",
",",
"id",
":",
"str",
",",
"*",
"*",
"kwargs",
")",
"->",
"SlackResponse",
":",
"kwargs",
".",
"update",
"(",
"{",
"\"file\"",
":",
"file",
",",
"\"id\"",
":",
"id",
"... |
Serialize and send a dict LLRP Message | def sendMessage ( self , msg_dict ) : sent_ids = [ ] for name in msg_dict : self . last_msg_id += 1 msg_dict [ name ] [ 'ID' ] = self . last_msg_id sent_ids . append ( ( name , self . last_msg_id ) ) llrp_msg = LLRPMessage ( msgdict = msg_dict ) assert llrp_msg . msgbytes , "LLRPMessage is empty" self . transport . write ( llrp_msg . msgbytes ) return sent_ids | 251,071 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L1283-L1299 | [
"def",
"available_formats",
"(",
")",
":",
"loaders",
"=",
"mesh_formats",
"(",
")",
"loaders",
".",
"extend",
"(",
"path_formats",
"(",
")",
")",
"loaders",
".",
"extend",
"(",
"compressed_loaders",
".",
"keys",
"(",
")",
")",
"return",
"loaders"
] |
Get a new LLRP client protocol object . | def buildProtocol ( self , addr ) : self . resetDelay ( ) # reset reconnection backoff state clargs = self . client_args . copy ( ) # optionally configure antennas from self.antenna_dict, which looks # like {'10.0.0.1:5084': {'1': 'ant1', '2': 'ant2'}} hostport = '{}:{}' . format ( addr . host , addr . port ) logger . debug ( 'Building protocol for %s' , hostport ) if hostport in self . antenna_dict : clargs [ 'antennas' ] = [ int ( x ) for x in self . antenna_dict [ hostport ] . keys ( ) ] elif addr . host in self . antenna_dict : clargs [ 'antennas' ] = [ int ( x ) for x in self . antenna_dict [ addr . host ] . keys ( ) ] logger . debug ( 'Antennas in buildProtocol: %s' , clargs . get ( 'antennas' ) ) logger . debug ( '%s start_inventory: %s' , hostport , clargs . get ( 'start_inventory' ) ) if self . start_first and not self . protocols : # this is the first protocol, so let's start it inventorying clargs [ 'start_inventory' ] = True proto = LLRPClient ( factory = self , * * clargs ) # register state-change callbacks with new client for state , cbs in self . _state_callbacks . items ( ) : for cb in cbs : proto . addStateCallback ( state , cb ) # register message callbacks with new client for msg_type , cbs in self . _message_callbacks . items ( ) : for cb in cbs : proto . addMessageCallback ( msg_type , cb ) return proto | 251,072 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L1339-L1376 | [
"def",
"setRepoData",
"(",
"self",
",",
"searchString",
",",
"category",
"=",
"\"\"",
",",
"extension",
"=",
"\"\"",
",",
"math",
"=",
"False",
",",
"game",
"=",
"False",
",",
"searchFiles",
"=",
"False",
")",
":",
"self",
".",
"searchString",
"=",
"se... |
Set the transmit power on one or all readers | def setTxPower ( self , tx_power , peername = None ) : if peername : protocols = [ p for p in self . protocols if p . peername [ 0 ] == peername ] else : protocols = self . protocols for proto in protocols : proto . setTxPower ( tx_power ) | 251,073 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L1411-L1423 | [
"def",
"get_cdn_metadata",
"(",
"self",
",",
"container",
")",
":",
"uri",
"=",
"\"%s/%s\"",
"%",
"(",
"self",
".",
"uri_base",
",",
"utils",
".",
"get_name",
"(",
"container",
")",
")",
"resp",
",",
"resp_body",
"=",
"self",
".",
"api",
".",
"cdn_requ... |
Stop inventory on all connected readers . | def politeShutdown ( self ) : protoDeferreds = [ ] for proto in self . protocols : protoDeferreds . append ( proto . stopPolitely ( disconnect = True ) ) return defer . DeferredList ( protoDeferreds ) | 251,074 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp.py#L1425-L1430 | [
"def",
"add",
"(",
"self",
",",
"pk",
",",
"quantity",
"=",
"1",
",",
"*",
"*",
"kwargs",
")",
":",
"pk",
"=",
"str",
"(",
"pk",
")",
"if",
"pk",
"in",
"self",
".",
"items",
":",
"existing_item",
"=",
"self",
".",
"items",
"[",
"pk",
"]",
"ex... |
Given a SGTIN - 96 hex string parse each segment . Returns a dictionary of the segments . | def parse_sgtin_96 ( sgtin_96 ) : if not sgtin_96 : raise Exception ( 'Pass in a value.' ) if not sgtin_96 . startswith ( "30" ) : # not a sgtin, not handled raise Exception ( 'Not SGTIN-96.' ) binary = "{0:020b}" . format ( int ( sgtin_96 , 16 ) ) . zfill ( 96 ) header = int ( binary [ : 8 ] , 2 ) tag_filter = int ( binary [ 8 : 11 ] , 2 ) partition = binary [ 11 : 14 ] partition_value = int ( partition , 2 ) m , l , n , k = SGTIN_96_PARTITION_MAP [ partition_value ] company_start = 8 + 3 + 3 company_end = company_start + m company_data = int ( binary [ company_start : company_end ] , 2 ) if company_data > pow ( 10 , l ) : # can't be too large raise Exception ( 'Company value is too large' ) company_prefix = str ( company_data ) . zfill ( l ) item_start = company_end item_end = item_start + n item_data = binary [ item_start : item_end ] item_number = int ( item_data , 2 ) item_reference = str ( item_number ) . zfill ( k ) serial = int ( binary [ - 38 : ] , 2 ) return { "header" : header , "filter" : tag_filter , "partition" : partition , "company_prefix" : company_prefix , "item_reference" : item_reference , "serial" : serial } | 251,075 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/epc/sgtin_96.py#L27-L71 | [
"def",
"pixelwise_or",
"(",
"self",
",",
"binary_im",
")",
":",
"data",
"=",
"np",
".",
"copy",
"(",
"self",
".",
"_data",
")",
"ind",
"=",
"np",
".",
"where",
"(",
"binary_im",
".",
"data",
">",
"0",
")",
"data",
"[",
"ind",
"[",
"0",
"]",
","... |
Decode any parameter to a byte sequence . | def decode_param ( data ) : logger . debug ( 'decode_param data: %r' , data ) header_len = struct . calcsize ( '!HH' ) partype , parlen = struct . unpack ( '!HH' , data [ : header_len ] ) pardata = data [ header_len : parlen ] logger . debug ( 'decode_param pardata: %r' , pardata ) ret = { 'Type' : partype , } if partype == 1023 : vsfmt = '!II' vendor , subtype = struct . unpack ( vsfmt , pardata [ : struct . calcsize ( vsfmt ) ] ) ret [ 'Vendor' ] = vendor ret [ 'Subtype' ] = subtype ret [ 'Data' ] = pardata [ struct . calcsize ( vsfmt ) : ] else : ret [ 'Data' ] = pardata , return ret , data [ parlen : ] | 251,076 | https://github.com/ransford/sllurp/blob/d744b7e17d7ba64a24d9a31bde6cba65d91ad9b1/sllurp/llrp_proto.py#L341-L369 | [
"def",
"cublasGetVersion",
"(",
"handle",
")",
":",
"version",
"=",
"ctypes",
".",
"c_int",
"(",
")",
"status",
"=",
"_libcublas",
".",
"cublasGetVersion_v2",
"(",
"handle",
",",
"ctypes",
".",
"byref",
"(",
"version",
")",
")",
"cublasCheckStatus",
"(",
"... |
Download the latest data . | def download_files ( file_list ) : for _ , source_data_file in file_list : sql_gz_name = source_data_file [ 'name' ] . split ( '/' ) [ - 1 ] msg = 'Downloading: %s' % ( sql_gz_name ) log . debug ( msg ) new_data = objectstore . get_object ( handelsregister_conn , source_data_file , 'handelsregister' ) # save output to file! with open ( 'data/{}' . format ( sql_gz_name ) , 'wb' ) as outputzip : outputzip . write ( new_data ) | 251,077 | https://github.com/Amsterdam/objectstore/blob/15852e8a3f159e4e6eb018866df0a84feb2c7f68/examples/handelsregister.py#L48-L60 | [
"async",
"def",
"create",
"(",
"self",
",",
"config",
":",
"dict",
"=",
"None",
",",
"access",
":",
"str",
"=",
"None",
",",
"replace",
":",
"bool",
"=",
"False",
")",
"->",
"Wallet",
":",
"LOGGER",
".",
"debug",
"(",
"'WalletManager.create >>> config %s... |
get an objectsctore connection | def get_connection ( store_settings : dict = { } ) -> Connection : store = store_settings if not store_settings : store = make_config_from_env ( ) os_options = { 'tenant_id' : store [ 'TENANT_ID' ] , 'region_name' : store [ 'REGION_NAME' ] , # 'endpoint_type': 'internalURL' } # when we are running in cloudvps we should use internal urls use_internal = os . getenv ( 'OBJECTSTORE_LOCAL' , '' ) if use_internal : os_options [ 'endpoint_type' ] = 'internalURL' connection = Connection ( authurl = store [ 'AUTHURL' ] , user = store [ 'USER' ] , key = store [ 'PASSWORD' ] , tenant_name = store [ 'TENANT_NAME' ] , auth_version = store [ 'VERSION' ] , os_options = os_options ) return connection | 251,078 | https://github.com/Amsterdam/objectstore/blob/15852e8a3f159e4e6eb018866df0a84feb2c7f68/objectstore/objectstore.py#L67-L96 | [
"def",
"run_role",
"(",
"self",
",",
"name",
",",
"options",
"=",
"None",
",",
"content",
"=",
"None",
")",
":",
"if",
"options",
"is",
"None",
":",
"options",
"=",
"{",
"}",
"if",
"content",
"is",
"None",
":",
"content",
"=",
"[",
"]",
"role_fn",
... |
Download object from objectstore . object_meta_data is an object retured when using get_full_container_list | def get_object ( connection , object_meta_data : dict , dirname : str ) : return connection . get_object ( dirname , object_meta_data [ 'name' ] ) [ 1 ] | 251,079 | https://github.com/Amsterdam/objectstore/blob/15852e8a3f159e4e6eb018866df0a84feb2c7f68/objectstore/objectstore.py#L117-L123 | [
"def",
"update_feature_type_rates",
"(",
"sender",
",",
"instance",
",",
"created",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"created",
":",
"for",
"role",
"in",
"ContributorRole",
".",
"objects",
".",
"all",
"(",
")",
":",
"FeatureType... |
Put file to objectstore | def put_object ( connection , container : str , object_name : str , contents , content_type : str ) -> None : connection . put_object ( container , object_name , contents = contents , content_type = content_type ) | 251,080 | https://github.com/Amsterdam/objectstore/blob/15852e8a3f159e4e6eb018866df0a84feb2c7f68/objectstore/objectstore.py#L126-L140 | [
"def",
"derivativeZ",
"(",
"self",
",",
"mLvl",
",",
"pLvl",
",",
"MedShk",
")",
":",
"xLvl",
"=",
"self",
".",
"xFunc",
"(",
"mLvl",
",",
"pLvl",
",",
"MedShk",
")",
"dxdShk",
"=",
"self",
".",
"xFunc",
".",
"derivativeZ",
"(",
"mLvl",
",",
"pLvl"... |
Delete single object from objectstore | def delete_object ( connection , container : str , object_meta_data : dict ) -> None : connection . delete_object ( container , object_meta_data [ 'name' ] ) | 251,081 | https://github.com/Amsterdam/objectstore/blob/15852e8a3f159e4e6eb018866df0a84feb2c7f68/objectstore/objectstore.py#L143-L147 | [
"def",
"_restart_session",
"(",
"self",
",",
"session",
")",
":",
"# remove old session key, if socket is None, that means the",
"# session was closed by user and there is no need to restart.",
"if",
"session",
".",
"socket",
"is",
"not",
"None",
":",
"self",
".",
"log",
".... |
Given connecton and container find database dumps | def return_file_objects ( connection , container , prefix = 'database' ) : options = [ ] meta_data = objectstore . get_full_container_list ( connection , container , prefix = 'database' ) env = ENV . upper ( ) for o_info in meta_data : expected_file = f'database.{ENV}' if o_info [ 'name' ] . startswith ( expected_file ) : dt = dateparser . parse ( o_info [ 'last_modified' ] ) now = datetime . datetime . now ( ) delta = now - dt LOG . debug ( 'AGE: %d %s' , delta . days , expected_file ) options . append ( ( dt , o_info ) ) options . sort ( ) return options | 251,082 | https://github.com/Amsterdam/objectstore/blob/15852e8a3f159e4e6eb018866df0a84feb2c7f68/objectstore/databasedumps.py#L66-L91 | [
"def",
"_verify_options",
"(",
"options",
")",
":",
"# sanity check all vals used for bitwise operations later",
"bitwise_args",
"=",
"[",
"(",
"'level'",
",",
"options",
"[",
"'level'",
"]",
")",
",",
"(",
"'facility'",
",",
"options",
"[",
"'facility'",
"]",
")"... |
Remove dumps older than x days | def remove_old_dumps ( connection , container : str , days = None ) : if not days : return if days < 20 : LOG . error ( 'A minimum of 20 backups is stored' ) return options = return_file_objects ( connection , container ) for dt , o_info in options : now = datetime . datetime . now ( ) delta = now - dt if delta . days > days : LOG . info ( 'Deleting %s' , o_info [ 'name' ] ) objectstore . delete_object ( connection , container , o_info ) | 251,083 | https://github.com/Amsterdam/objectstore/blob/15852e8a3f159e4e6eb018866df0a84feb2c7f68/objectstore/databasedumps.py#L94-L112 | [
"def",
"render_cvmfs_sc",
"(",
"cvmfs_volume",
")",
":",
"name",
"=",
"CVMFS_REPOSITORIES",
"[",
"cvmfs_volume",
"]",
"rendered_template",
"=",
"dict",
"(",
"REANA_CVMFS_SC_TEMPLATE",
")",
"rendered_template",
"[",
"'metadata'",
"]",
"[",
"'name'",
"]",
"=",
"\"cs... |
Download database dump | def download_database ( connection , container : str , target : str = "" ) : meta_data = objectstore . get_full_container_list ( connection , container , prefix = 'database' ) options = return_file_objects ( connection , container ) for o_info in meta_data : expected_file = f'database.{ENV}' LOG . info ( o_info [ 'name' ] ) if o_info [ 'name' ] . startswith ( expected_file ) : dt = dateparser . parse ( o_info [ 'last_modified' ] ) # now = datetime.datetime.now() options . append ( ( dt , o_info ) ) options . sort ( ) if not options : LOG . error ( 'Dumps missing? ENVIRONMENT wrong? (acceptance / production' ) LOG . error ( 'Environtment {ENV}' ) sys . exit ( 1 ) newest = options [ - 1 ] [ 1 ] LOG . debug ( 'Downloading: %s' , ( newest [ 'name' ] ) ) target_file = os . path . join ( target , expected_file ) LOG . info ( 'TARGET: %s' , target_file ) if os . path . exists ( target_file ) : LOG . info ( 'Already downloaded' ) return LOG . error ( 'TARGET does not exists downloading...' ) new_data = objectstore . get_object ( connection , newest , container ) # save output to file! with open ( target_file , 'wb' ) as outputzip : outputzip . write ( new_data ) | 251,084 | https://github.com/Amsterdam/objectstore/blob/15852e8a3f159e4e6eb018866df0a84feb2c7f68/objectstore/databasedumps.py#L115-L160 | [
"def",
"__numero_tres_cifras",
"(",
"self",
",",
"number",
",",
"indice",
"=",
"None",
",",
"sing",
"=",
"False",
")",
":",
"number",
"=",
"int",
"(",
"number",
")",
"if",
"number",
"<",
"30",
":",
"if",
"sing",
":",
"return",
"especiales_apocopado",
"... |
Remember the authenticated identity . | def remember ( self , user_name ) : log . debug ( 'Repoze OAuth remember' ) environ = toolkit . request . environ rememberer = self . _get_rememberer ( environ ) identity = { 'repoze.who.userid' : user_name } headers = rememberer . remember ( environ , identity ) for header , value in headers : toolkit . response . headers . add ( header , value ) | 251,085 | https://github.com/conwetlab/ckanext-oauth2/blob/e68dd2664229b7563d77b2c8fc869fe57b747c88/ckanext/oauth2/oauth2.py#L206-L218 | [
"def",
"_set_cdn_access",
"(",
"self",
",",
"container",
",",
"public",
",",
"ttl",
"=",
"None",
")",
":",
"headers",
"=",
"{",
"\"X-Cdn-Enabled\"",
":",
"\"%s\"",
"%",
"public",
"}",
"if",
"public",
"and",
"ttl",
":",
"headers",
"[",
"\"X-Ttl\"",
"]",
... |
Redirect to the callback URL after a successful authentication . | def redirect_from_callback ( self ) : state = toolkit . request . params . get ( 'state' ) came_from = get_came_from ( state ) toolkit . response . status = 302 toolkit . response . location = came_from | 251,086 | https://github.com/conwetlab/ckanext-oauth2/blob/e68dd2664229b7563d77b2c8fc869fe57b747c88/ckanext/oauth2/oauth2.py#L220-L225 | [
"def",
"locations_within",
"(",
"a",
",",
"b",
",",
"tolerance",
")",
":",
"ret",
"=",
"''",
"# Clone b so that we can destroy it.",
"b",
"=",
"dict",
"(",
"b",
")",
"for",
"(",
"key",
",",
"value",
")",
"in",
"a",
".",
"items",
"(",
")",
":",
"if",
... |
Return True if user can share folder . | def can_share_folder ( self , user , folder ) : return folder . parent_id is None and folder . author_id == user . id | 251,087 | https://github.com/pinax/pinax-documents/blob/b8c6a748976ec4b22cff9b195eb426b46c7a2a1e/pinax/documents/hooks.py#L20-L24 | [
"def",
"_error_messages",
"(",
"self",
",",
"driver_id",
")",
":",
"assert",
"isinstance",
"(",
"driver_id",
",",
"ray",
".",
"DriverID",
")",
"message",
"=",
"self",
".",
"redis_client",
".",
"execute_command",
"(",
"\"RAY.TABLE_LOOKUP\"",
",",
"ray",
".",
... |
Return labels indicating amount of storage used . | def storage_color ( self , user_storage ) : p = user_storage . percentage if p >= 0 and p < 60 : return "success" if p >= 60 and p < 90 : return "warning" if p >= 90 and p <= 100 : return "danger" raise ValueError ( "percentage out of range" ) | 251,088 | https://github.com/pinax/pinax-documents/blob/b8c6a748976ec4b22cff9b195eb426b46c7a2a1e/pinax/documents/hooks.py#L26-L37 | [
"def",
"_init_publisher_ws",
"(",
"self",
")",
":",
"logging",
".",
"debug",
"(",
"\"Initializing new web socket connection.\"",
")",
"url",
"=",
"(",
"'wss://%s/v1/stream/messages/'",
"%",
"self",
".",
"eventhub_client",
".",
"host",
")",
"headers",
"=",
"self",
... |
Send messages . success message after successful folder creation . | def folder_created_message ( self , request , folder ) : messages . success ( request , _ ( "Folder {} was created" . format ( folder ) ) ) | 251,089 | https://github.com/pinax/pinax-documents/blob/b8c6a748976ec4b22cff9b195eb426b46c7a2a1e/pinax/documents/hooks.py#L39-L43 | [
"def",
"main",
"(",
")",
":",
"save_settings",
"=",
"None",
"stdin_fd",
"=",
"-",
"1",
"try",
":",
"import",
"termios",
"stdin_fd",
"=",
"sys",
".",
"stdin",
".",
"fileno",
"(",
")",
"save_settings",
"=",
"termios",
".",
"tcgetattr",
"(",
"stdin_fd",
"... |
Send messages . success message after successful document creation . | def document_created_message ( self , request , document ) : messages . success ( request , _ ( "Document {} was created" . format ( document ) ) ) | 251,090 | https://github.com/pinax/pinax-documents/blob/b8c6a748976ec4b22cff9b195eb426b46c7a2a1e/pinax/documents/hooks.py#L45-L49 | [
"def",
"set_output_data_rate",
"(",
"self",
",",
"output_data_rate",
")",
":",
"# self.standby()",
"register",
"=",
"self",
".",
"MMA8452Q_Register",
"[",
"'CTRL_REG1'",
"]",
"self",
".",
"board",
".",
"i2c_read_request",
"(",
"self",
".",
"address",
",",
"regis... |
Send messages . success message after successful share . | def folder_shared_message ( self , request , user , folder ) : messages . success ( request , _ ( "Folder {} is now shared with {}" . format ( folder , user ) ) ) | 251,091 | https://github.com/pinax/pinax-documents/blob/b8c6a748976ec4b22cff9b195eb426b46c7a2a1e/pinax/documents/hooks.py#L51-L55 | [
"def",
"_ir_calibrate",
"(",
"self",
",",
"data",
")",
":",
"# No radiance -> no temperature",
"data",
"=",
"da",
".",
"where",
"(",
"data",
"==",
"0",
",",
"np",
".",
"float32",
"(",
"np",
".",
"nan",
")",
",",
"data",
")",
"cwl",
"=",
"self",
".",
... |
Perform folder operations prior to deletions . For example deleting all contents . | def folder_pre_delete ( self , request , folder ) : for m in folder . members ( ) : if m . __class__ == folder . __class__ : self . folder_pre_delete ( request , m ) m . delete ( ) | 251,092 | https://github.com/pinax/pinax-documents/blob/b8c6a748976ec4b22cff9b195eb426b46c7a2a1e/pinax/documents/hooks.py#L75-L82 | [
"def",
"api_headers_tween_factory",
"(",
"handler",
",",
"registry",
")",
":",
"def",
"api_headers_tween",
"(",
"request",
")",
":",
"response",
"=",
"handler",
"(",
"request",
")",
"set_version",
"(",
"request",
",",
"response",
")",
"set_req_guid",
"(",
"req... |
Callable passed to the FileField s upload_to kwarg on Document . file | def file_upload_to ( self , instance , filename ) : ext = filename . split ( "." ) [ - 1 ] filename = "{}.{}" . format ( uuid . uuid4 ( ) , ext ) return os . path . join ( "document" , filename ) | 251,093 | https://github.com/pinax/pinax-documents/blob/b8c6a748976ec4b22cff9b195eb426b46c7a2a1e/pinax/documents/hooks.py#L84-L90 | [
"def",
"_wait_for_machine_booted",
"(",
"name",
",",
"suffictinet_texts",
"=",
"None",
")",
":",
"# TODO: rewrite it using probes module in utils",
"suffictinet_texts",
"=",
"suffictinet_texts",
"or",
"[",
"\"systemd-logind\"",
"]",
"# optionally use: \"Unit: machine\"",
"for",... |
All folders the given user can do something with . | def for_user ( self , user ) : qs = SharedMemberQuerySet ( model = self . model , using = self . _db , user = user ) qs = qs . filter ( Q ( author = user ) | Q ( foldershareduser__user = user ) ) return qs . distinct ( ) & self . distinct ( ) | 251,094 | https://github.com/pinax/pinax-documents/blob/b8c6a748976ec4b22cff9b195eb426b46c7a2a1e/pinax/documents/managers.py#L31-L37 | [
"def",
"remove",
"(",
"self",
",",
"obj",
")",
":",
"relationship_table",
"=",
"self",
".",
"params",
"[",
"'relationship_table'",
"]",
"with",
"self",
".",
"obj",
".",
"backend",
".",
"transaction",
"(",
"implicit",
"=",
"True",
")",
":",
"condition",
"... |
Parse command line arguments to construct a dictionary of cluster parameters that can be used to determine which clusters to list . | def _parse_list ( cls , args ) : argparser = ArgumentParser ( prog = "cluster list" ) group = argparser . add_mutually_exclusive_group ( ) group . add_argument ( "--id" , dest = "cluster_id" , help = "show cluster with this id" ) group . add_argument ( "--label" , dest = "label" , help = "show cluster with this label" ) group . add_argument ( "--state" , dest = "state" , action = "store" , choices = [ 'up' , 'down' , 'pending' , 'terminating' ] , help = "list only clusters in the given state" ) pagination_group = group . add_argument_group ( ) pagination_group . add_argument ( "--page" , dest = "page" , action = "store" , type = int , help = "page number" ) pagination_group . add_argument ( "--per-page" , dest = "per_page" , action = "store" , type = int , help = "number of clusters to be retrieved per page" ) arguments = argparser . parse_args ( args ) return vars ( arguments ) | 251,095 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/cluster.py#L31-L62 | [
"def",
"setOverlayTextureColorSpace",
"(",
"self",
",",
"ulOverlayHandle",
",",
"eTextureColorSpace",
")",
":",
"fn",
"=",
"self",
".",
"function_table",
".",
"setOverlayTextureColorSpace",
"result",
"=",
"fn",
"(",
"ulOverlayHandle",
",",
"eTextureColorSpace",
")",
... |
Parse command line arguments for cluster manage commands . | def _parse_cluster_manage_command ( cls , args , action ) : argparser = ArgumentParser ( prog = "cluster_manage_command" ) group = argparser . add_mutually_exclusive_group ( required = True ) group . add_argument ( "--id" , dest = "cluster_id" , help = "execute on cluster with this id" ) group . add_argument ( "--label" , dest = "label" , help = "execute on cluster with this label" ) if action == "remove" or action == "update" : argparser . add_argument ( "--private_dns" , help = "the private_dns of the machine to be updated/removed" , required = True ) if action == "update" : argparser . add_argument ( "--command" , help = "the update command to be executed" , required = True , choices = [ "replace" ] ) arguments = argparser . parse_args ( args ) return arguments | 251,096 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/cluster.py#L530-L553 | [
"def",
"process_frames_mouth",
"(",
"self",
",",
"frames",
")",
":",
"self",
".",
"face",
"=",
"np",
".",
"array",
"(",
"frames",
")",
"self",
".",
"mouth",
"=",
"np",
".",
"array",
"(",
"frames",
")",
"self",
".",
"set_data",
"(",
"frames",
")"
] |
Parse command line arguments for reassigning label . | def _parse_reassign_label ( cls , args ) : argparser = ArgumentParser ( prog = "cluster reassign_label" ) argparser . add_argument ( "destination_cluster" , metavar = "destination_cluster_id_label" , help = "id/label of the cluster to move the label to" ) argparser . add_argument ( "label" , help = "label to be moved from the source cluster" ) arguments = argparser . parse_args ( args ) return arguments | 251,097 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/cluster.py#L556-L570 | [
"def",
"validate",
"(",
"self",
",",
"data",
")",
":",
"if",
"data",
"is",
"not",
"None",
"and",
"not",
"isinstance",
"(",
"data",
",",
"dict",
")",
":",
"raise",
"serializers",
".",
"ValidationError",
"(",
"\"Invalid data\"",
")",
"try",
":",
"profiles"... |
Reassign a label from one cluster to another . | def reassign_label ( cls , destination_cluster , label ) : conn = Qubole . agent ( version = Cluster . api_version ) data = { "destination_cluster" : destination_cluster , "label" : label } return conn . put ( cls . rest_entity_path + "/reassign-label" , data ) | 251,098 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/cluster.py#L573-L587 | [
"def",
"print_debug",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"WTF_CONFIG_READER",
".",
"get",
"(",
"\"debug\"",
",",
"False",
")",
"==",
"True",
":",
"print",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] |
Parse command line arguments for snapshot command . | def _parse_snapshot_restore_command ( cls , args , action ) : argparser = ArgumentParser ( prog = "cluster %s" % action ) group = argparser . add_mutually_exclusive_group ( required = True ) group . add_argument ( "--id" , dest = "cluster_id" , help = "execute on cluster with this id" ) group . add_argument ( "--label" , dest = "label" , help = "execute on cluster with this label" ) argparser . add_argument ( "--s3_location" , help = "s3_location where backup is stored" , required = True ) if action == "snapshot" : argparser . add_argument ( "--backup_type" , help = "backup_type: full/incremental, default is full" ) elif action == "restore_point" : argparser . add_argument ( "--backup_id" , help = "back_id from which restoration will be done" , required = True ) argparser . add_argument ( "--table_names" , help = "table(s) which are to be restored" , required = True ) argparser . add_argument ( "--no-overwrite" , action = "store_false" , help = "With this option, restore overwrites to the existing table if theres any in restore target" ) argparser . add_argument ( "--no-automatic" , action = "store_false" , help = "With this option, all the dependencies are automatically restored together with this backup image following the correct order" ) arguments = argparser . parse_args ( args ) return arguments | 251,099 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/cluster.py#L598-L625 | [
"def",
"unbind",
"(",
"self",
",",
"devices_to_unbind",
")",
":",
"if",
"self",
".",
"entity_api_key",
"==",
"\"\"",
":",
"return",
"{",
"'status'",
":",
"'failure'",
",",
"'response'",
":",
"'No API key found in request'",
"}",
"url",
"=",
"self",
".",
"bas... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.