idx int64 0 63k | question stringlengths 53 5.28k | target stringlengths 5 805 |
|---|---|---|
22,800 | def wrap ( cls , meth ) : async def inner ( * args , ** kwargs ) : sock = await meth ( * args , ** kwargs ) return cls ( sock ) return inner | Wraps a connection opening method in this class . |
22,801 | async def set ( self , * args , ** kwargs ) : return await _maybe_await ( self . event . set ( * args , ** kwargs ) ) | Sets the value of the event . |
22,802 | def register ( self , library : str , cbl : Callable [ [ '_AsyncLib' ] , None ] ) : self . _handlers [ library ] = cbl | Registers a callable to set up a library . |
22,803 | async def trio_open_connection ( host , port , * , ssl = False , ** kwargs ) : import trio if not ssl : sock = await trio . open_tcp_stream ( host , port ) else : if isinstance ( ssl , bool ) : ssl_context = None else : ssl_context = ssl sock = await trio . open_ssl_over_tcp_stream ( host , port , ssl_context = ssl_context ) await sock . do_handshake ( ) sock . close = sock . aclose return sock | Allows connections to be made that may or may not require ssl . Somewhat surprisingly trio doesn t have an abstraction for this like curio even though it s fairly trivial to write . Down the line hopefully . |
22,804 | def agent ( state , host , server = None , port = None ) : args = [ ] if server : args . append ( '--server=%s' % server ) if port : args . append ( '--masterport=%s' % port ) yield 'puppet agent -t %s' % ' ' . join ( args ) | Run puppet agent |
22,805 | def load_config ( deploy_dir ) : config = Config ( ) config_filename = path . join ( deploy_dir , 'config.py' ) if path . exists ( config_filename ) : extract_file_config ( config_filename , config ) exec_file ( config_filename ) return config | Loads any local config . py file . |
22,806 | def load_deploy_config ( deploy_filename , config = None ) : if not config : config = Config ( ) if not deploy_filename : return if path . exists ( deploy_filename ) : extract_file_config ( deploy_filename , config ) return config | Loads any local config overrides in the deploy file . |
22,807 | def parse_iptables_rule ( line ) : bits = line . split ( ) definition = { } key = None args = [ ] not_arg = False def add_args ( ) : arg_string = ' ' . join ( args ) if key in IPTABLES_ARGS : definition_key = ( 'not_{0}' . format ( IPTABLES_ARGS [ key ] ) if not_arg else IPTABLES_ARGS [ key ] ) definition [ definition_key ] = arg_string else : definition . setdefault ( 'extras' , [ ] ) . extend ( ( key , arg_string ) ) for bit in bits : if bit == '!' : if key : add_args ( ) args = [ ] key = None not_arg = True elif bit . startswith ( '-' ) : if key : add_args ( ) args = [ ] not_arg = False key = bit else : args . append ( bit ) if key : add_args ( ) if 'extras' in definition : definition [ 'extras' ] = set ( definition [ 'extras' ] ) return definition | Parse one iptables rule . Returns a dict where each iptables code argument is mapped to a name using IPTABLES_ARGS . |
22,808 | def add_op ( state , op_func , * args , ** kwargs ) : frameinfo = get_caller_frameinfo ( ) kwargs [ 'frameinfo' ] = frameinfo for host in state . inventory : op_func ( state , host , * args , ** kwargs ) | Prepare & add an operation to pyinfra . state by executing it on all hosts . |
22,809 | def add_deploy ( state , deploy_func , * args , ** kwargs ) : frameinfo = get_caller_frameinfo ( ) kwargs [ 'frameinfo' ] = frameinfo for host in state . inventory : deploy_func ( state , host , * args , ** kwargs ) | Prepare & add an deploy to pyinfra . state by executing it on all hosts . |
22,810 | def setup_arguments ( arguments ) : for key in ( '--parallel' , '--port' , '--fail-percent' ) : if arguments [ key ] : try : arguments [ key ] = int ( arguments [ key ] ) except ValueError : raise CliError ( '{0} is not a valid integer for {1}' . format ( arguments [ key ] , key , ) ) if arguments [ '--run' ] : op , args = setup_op_and_args ( arguments [ '--run' ] , arguments [ 'ARGS' ] ) else : op = args = None if arguments [ 'DEPLOY' ] : if not path . exists ( arguments [ 'DEPLOY' ] ) : raise CliError ( 'Deploy file not found: {0}' . format ( arguments [ 'DEPLOY' ] ) ) if arguments [ '--key' ] : if not path . exists ( arguments [ '--key' ] ) : raise CliError ( 'Private key file not found: {0}' . format ( arguments [ '--key' ] ) ) return { 'inventory' : arguments [ '-i' ] , 'deploy' : arguments [ 'DEPLOY' ] , 'verbose' : arguments [ '-v' ] , 'dry' : arguments [ '--dry' ] , 'serial' : arguments [ '--serial' ] , 'no_wait' : arguments [ '--no-wait' ] , 'debug' : arguments [ '--debug' ] , 'debug_data' : arguments [ '--debug-data' ] , 'debug_state' : arguments [ '--debug-state' ] , 'fact' : arguments [ '--fact' ] , 'limit' : arguments [ '--limit' ] , 'op' : op , 'op_args' : args , 'user' : arguments [ '--user' ] , 'key' : arguments [ '--key' ] , 'key_password' : arguments [ '--key-password' ] , 'password' : arguments [ '--password' ] , 'port' : arguments [ '--port' ] , 'sudo' : arguments [ '--sudo' ] , 'sudo_user' : arguments [ '--sudo-user' ] , 'su_user' : arguments [ '--su-user' ] , 'parallel' : arguments [ '--parallel' ] , 'fail_percent' : arguments [ '--fail-percent' ] , } | Prepares argumnents output by docopt . |
22,811 | def sql ( state , host , sql , database = None , mysql_user = None , mysql_password = None , mysql_host = None , mysql_port = None , ) : yield make_execute_mysql_command ( sql , database = database , user = mysql_user , password = mysql_password , host = mysql_host , port = mysql_port , ) | Execute arbitrary SQL against MySQL . |
22,812 | def dump ( state , host , remote_filename , database = None , mysql_user = None , mysql_password = None , mysql_host = None , mysql_port = None , ) : yield '{0} > {1}' . format ( make_mysql_command ( executable = 'mysqldump' , database = database , user = mysql_user , password = mysql_password , host = mysql_host , port = mysql_port , ) , remote_filename ) | Dump a MySQL database into a . sql file . Requires mysqldump . |
22,813 | def get_host ( self , name , default = NoHostError ) : if name in self . hosts : return self . hosts [ name ] if default is NoHostError : raise NoHostError ( 'No such host: {0}' . format ( name ) ) return default | Get a single host by name . |
22,814 | def get_group ( self , name , default = NoGroupError ) : if name in self . groups : return self . groups [ name ] if default is NoGroupError : raise NoGroupError ( 'No such group: {0}' . format ( name ) ) return default | Get a list of hosts belonging to a group . |
22,815 | def get_groups_data ( self , groups ) : data = { } for group in groups : data . update ( self . get_group_data ( group ) ) return data | Gets aggregated data from a list of groups . Vars are collected in order so for any groups which define the same var twice the last group s value will hold . |
22,816 | def get_deploy_data ( self ) : if self . state and self . state . deploy_data : return self . state . deploy_data return { } | Gets any default data attached to the current deploy if any . |
22,817 | def config ( state , host , key , value , repo = None , ) : existing_config = host . fact . git_config ( repo ) if key not in existing_config or existing_config [ key ] != value : if repo is None : yield 'git config --global {0} "{1}"' . format ( key , value ) else : yield 'cd {0} && git config --local {1} "{2}"' . format ( repo , key , value ) | Manage git config for a repository or globally . |
22,818 | def include ( filename , hosts = False , when = True ) : if not pyinfra . is_cli : raise PyinfraError ( 'local.include is only available in CLI mode.' ) if not when : return if hosts is not False : hosts = ensure_host_list ( hosts , inventory = pseudo_state . inventory ) if pseudo_host not in hosts : return if pseudo_state . deploy_dir : filename = path . join ( pseudo_state . deploy_dir , filename ) frameinfo = get_caller_frameinfo ( ) logger . debug ( 'Including local file: {0}' . format ( filename ) ) try : from pyinfra_cli . config import extract_file_config from pyinfra_cli . util import exec_file config_data = extract_file_config ( filename ) kwargs = { key . lower ( ) : value for key , value in six . iteritems ( config_data ) if key in [ 'SUDO' , 'SUDO_USER' , 'SU_USER' , 'PRESERVE_SUDO_ENV' , 'IGNORE_ERRORS' , ] } with pseudo_state . deploy ( filename , kwargs , None , frameinfo . lineno , in_deploy = False , ) : exec_file ( filename ) except IOError as e : raise PyinfraError ( 'Could not include local file: {0}\n{1}' . format ( filename , e ) , ) | Executes a local python file within the pyinfra . pseudo_state . deploy_dir directory . |
22,819 | def send ( self , request , stem = None ) : if stem is not None : request . url = request . url + "/" + stem . lstrip ( "/" ) prepped = self . session . prepare_request ( request ) settings = self . session . merge_environment_settings ( url = prepped . url , proxies = { } , stream = None , verify = None , cert = None ) return self . session . send ( prepped , ** settings ) | Prepare and send a request |
22,820 | def list ( self ) : request = requests . Request ( 'GET' , 'https://api.github.com/gists' , headers = { 'Accept-Encoding' : 'identity, deflate, compress, gzip' , 'User-Agent' : 'python-requests/1.2.0' , 'Accept' : 'application/vnd.github.v3.base64' , } , params = { 'access_token' : self . token , 'per_page' : 100 , } , ) pattern = re . compile ( r'<([^>]*)>; rel="([^"]*)"' ) gists = [ ] while True : try : response = self . send ( request ) . json ( ) except Exception : break for gist in response : try : gists . append ( GistInfo ( gist [ 'id' ] , gist [ 'public' ] , gist [ 'description' ] , ) ) except KeyError : continue try : link = response . headers [ 'link' ] for result in pattern . finditer ( link ) : url = result . group ( 1 ) rel = result . group ( 2 ) if rel == 'next' : request . url = url break else : return gists except Exception : break return gists | Returns a list of the users gists as GistInfo objects |
22,821 | def create ( self , request , desc , files , public = False ) : request . data = json . dumps ( { "description" : desc , "public" : public , "files" : files , } ) return self . send ( request ) . json ( ) [ 'html_url' ] | Creates a gist |
22,822 | def files ( self , request , id ) : gist = self . send ( request , id ) . json ( ) return gist [ 'files' ] | Returns a list of files in the gist |
22,823 | def content ( self , request , id ) : gist = self . send ( request , id ) . json ( ) def convert ( data ) : return base64 . b64decode ( data ) . decode ( 'utf-8' ) content = { } for name , data in gist [ 'files' ] . items ( ) : content [ name ] = convert ( data [ 'content' ] ) return content | Returns the content of the gist |
22,824 | def archive ( self , request , id ) : gist = self . send ( request , id ) . json ( ) with tarfile . open ( '{}.tar.gz' . format ( id ) , mode = 'w:gz' ) as archive : for name , data in gist [ 'files' ] . items ( ) : with tempfile . NamedTemporaryFile ( 'w+' ) as fp : fp . write ( data [ 'content' ] ) fp . flush ( ) archive . add ( fp . name , arcname = name ) | Create an archive of a gist |
22,825 | def edit ( self , request , id ) : with pushd ( tempfile . gettempdir ( ) ) : try : self . clone ( id ) with pushd ( id ) : files = [ f for f in os . listdir ( '.' ) if os . path . isfile ( f ) ] quoted = [ '"{}"' . format ( f ) for f in files ] os . system ( "{} {}" . format ( self . editor , ' ' . join ( quoted ) ) ) os . system ( 'git commit -av && git push' ) finally : shutil . rmtree ( id ) | Edit a gist |
22,826 | def description ( self , request , id , description ) : request . data = json . dumps ( { "description" : description } ) return self . send ( request , id ) . json ( ) [ 'html_url' ] | Updates the description of a gist |
22,827 | def clone ( self , id , name = None ) : url = 'git@gist.github.com:/{}' . format ( id ) if name is None : os . system ( 'git clone {}' . format ( url ) ) else : os . system ( 'git clone {} {}' . format ( url , name ) ) | Clone a gist |
22,828 | def command ( state , host , hostname , command , ssh_user = None ) : connection_target = hostname if ssh_user : connection_target = '@' . join ( ( ssh_user , hostname ) ) yield 'ssh {0} "{1}"' . format ( connection_target , command ) | Execute commands on other servers over SSH . |
22,829 | def upload ( state , host , hostname , filename , remote_filename = None , use_remote_sudo = False , ssh_keyscan = False , ssh_user = None , ) : remote_filename = remote_filename or filename connection_target = hostname if ssh_user : connection_target = '@' . join ( ( ssh_user , hostname ) ) if ssh_keyscan : yield keyscan ( state , host , hostname ) if not use_remote_sudo : yield 'scp {0} {1}:{2}' . format ( filename , connection_target , remote_filename ) else : temp_remote_filename = state . get_temp_filename ( ) upload_cmd = 'scp {0} {1}:{2}' . format ( filename , connection_target , temp_remote_filename , ) yield upload_cmd yield command ( state , host , connection_target , 'sudo mv {0} {1}' . format ( temp_remote_filename , remote_filename , ) ) | Upload files to other servers using scp . |
22,830 | def download ( state , host , hostname , filename , local_filename = None , force = False , ssh_keyscan = False , ssh_user = None , ) : local_filename = local_filename or filename local_file_info = host . fact . file ( local_filename ) if local_file_info is False : raise OperationError ( 'Local destination {0} already exists and is not a file' . format ( local_filename , ) , ) if local_file_info and not force : return connection_target = hostname if ssh_user : connection_target = '@' . join ( ( ssh_user , hostname ) ) if ssh_keyscan : yield keyscan ( state , host , hostname ) yield 'scp {0}:{1} {2}' . format ( connection_target , filename , local_filename ) | Download files from other servers using scp . |
22,831 | def pop_op_kwargs ( state , kwargs ) : meta_kwargs = state . deploy_kwargs or { } def get_kwarg ( key , default = None ) : return kwargs . pop ( key , meta_kwargs . get ( key , default ) ) env = state . config . ENV . copy ( ) env . update ( get_kwarg ( 'env' , { } ) ) hosts = get_kwarg ( 'hosts' ) hosts = ensure_host_list ( hosts , inventory = state . inventory ) if meta_kwargs . get ( 'hosts' ) is not None : hosts = [ host for host in hosts if host in meta_kwargs [ 'hosts' ] ] return { 'env' : env , 'hosts' : hosts , 'when' : get_kwarg ( 'when' , True ) , 'sudo' : get_kwarg ( 'sudo' , state . config . SUDO ) , 'sudo_user' : get_kwarg ( 'sudo_user' , state . config . SUDO_USER ) , 'su_user' : get_kwarg ( 'su_user' , state . config . SU_USER ) , 'preserve_sudo_env' : get_kwarg ( 'preserve_sudo_env' , state . config . PRESERVE_SUDO_ENV , ) , 'ignore_errors' : get_kwarg ( 'ignore_errors' , state . config . IGNORE_ERRORS , ) , 'timeout' : get_kwarg ( 'timeout' ) , 'get_pty' : get_kwarg ( 'get_pty' , False ) , 'serial' : get_kwarg ( 'serial' , False ) , 'run_once' : get_kwarg ( 'run_once' , False ) , 'parallel' : get_kwarg ( 'parallel' ) , 'on_success' : get_kwarg ( 'on_success' ) , 'on_error' : get_kwarg ( 'on_error' ) , 'op' : get_kwarg ( 'op' ) , } | Pop and return operation global keyword arguments . |
22,832 | def get_template ( filename_or_string , is_string = False ) : cache_key = sha1_hash ( filename_or_string ) if is_string else filename_or_string if cache_key in TEMPLATES : return TEMPLATES [ cache_key ] if is_string : template_string = filename_or_string else : with open ( filename_or_string , 'r' ) as file_io : template_string = file_io . read ( ) TEMPLATES [ cache_key ] = Template ( template_string , keep_trailing_newline = True ) return TEMPLATES [ cache_key ] | Gets a jinja2 Template object for the input filename or string with caching based on the filename of the template or the SHA1 of the input string . |
22,833 | def underscore ( name ) : s1 = re . sub ( '(.)([A-Z][a-z]+)' , r'\1_\2' , name ) return re . sub ( '([a-z0-9])([A-Z])' , r'\1_\2' , s1 ) . lower ( ) | Transform CamelCase - > snake_case . |
22,834 | def sha1_hash ( string ) : hasher = sha1 ( ) hasher . update ( string . encode ( ) ) return hasher . hexdigest ( ) | Return the SHA1 of the input string . |
22,835 | def make_command ( command , env = None , su_user = None , sudo = False , sudo_user = None , preserve_sudo_env = False , ) : debug_meta = { } for key , value in ( ( 'sudo' , sudo ) , ( 'sudo_user' , sudo_user ) , ( 'su_user' , su_user ) , ( 'env' , env ) , ) : if value : debug_meta [ key ] = value logger . debug ( 'Building command ({0}): {1}' . format ( ' ' . join ( '{0}: {1}' . format ( key , value ) for key , value in six . iteritems ( debug_meta ) ) , command ) ) if env : env_string = ' ' . join ( [ '{0}={1}' . format ( key , value ) for key , value in six . iteritems ( env ) ] ) command = 'export {0}; {1}' . format ( env_string , command ) command = shlex_quote ( command ) if su_user : command = 'su {0} -c {1}' . format ( su_user , command ) else : command = 'sh -c {0}' . format ( command ) if sudo : sudo_bits = [ 'sudo' , '-H' ] if preserve_sudo_env : sudo_bits . append ( '-E' ) if sudo_user : sudo_bits . extend ( ( '-u' , sudo_user ) ) command = '{0} {1}' . format ( ' ' . join ( sudo_bits ) , command ) return command | Builds a shell command with various kwargs . |
22,836 | def make_hash ( obj ) : if isinstance ( obj , ( set , tuple , list ) ) : hash_string = '' . join ( [ make_hash ( e ) for e in obj ] ) elif isinstance ( obj , dict ) : hash_string = '' . join ( '' . join ( ( key , make_hash ( value ) ) ) for key , value in six . iteritems ( obj ) ) else : hash_string = ( '_PYINFRA_CONSTANT' if obj in ( True , False , None ) else obj if isinstance ( obj , six . string_types ) else obj . __name__ if hasattr ( obj , '__name__' ) else obj . name if hasattr ( obj , 'name' ) else repr ( obj ) ) return sha1_hash ( hash_string ) | Make a hash from an arbitrary nested dictionary list tuple or set used to generate ID s for operations based on their name & arguments . |
22,837 | def get_file_sha1 ( filename_or_io ) : file_data = get_file_io ( filename_or_io ) cache_key = file_data . cache_key if cache_key and cache_key in FILE_SHAS : return FILE_SHAS [ cache_key ] with file_data as file_io : hasher = sha1 ( ) buff = file_io . read ( BLOCKSIZE ) while len ( buff ) > 0 : if isinstance ( buff , six . text_type ) : buff = buff . encode ( 'utf-8' ) hasher . update ( buff ) buff = file_io . read ( BLOCKSIZE ) digest = hasher . hexdigest ( ) if cache_key : FILE_SHAS [ cache_key ] = digest return digest | Calculates the SHA1 of a file or file object using a buffer to handle larger files . |
22,838 | def read_buffer ( io , print_output = False , print_func = None ) : def _print ( line ) : if print_output : if print_func : formatted_line = print_func ( line ) else : formatted_line = line encoded_line = unicode ( formatted_line ) . encode ( 'utf-8' ) print ( encoded_line ) out = [ ] for line in io : if not isinstance ( line , six . text_type ) : line = line . decode ( 'utf-8' ) line = line . strip ( ) out . append ( line ) _print ( line ) return out | Reads a file - like buffer object into lines and optionally prints the output . |
22,839 | def start ( state , host , ctid , force = False ) : args = [ '{0}' . format ( ctid ) ] if force : args . append ( '--force' ) yield 'vzctl start {0}' . format ( ' ' . join ( args ) ) | Start OpenVZ containers . |
22,840 | def stop ( state , host , ctid ) : args = [ '{0}' . format ( ctid ) ] yield 'vzctl stop {0}' . format ( ' ' . join ( args ) ) | Stop OpenVZ containers . |
22,841 | def restart ( state , host , ctid , force = False ) : yield stop ( state , host , ctid ) yield start ( state , host , ctid , force = force ) | Restart OpenVZ containers . |
22,842 | def create ( state , host , ctid , template = None ) : current_containers = host . fact . openvz_containers if ctid in current_containers : raise OperationError ( 'An OpenVZ container with CTID {0} already exists' . format ( ctid ) , ) args = [ '{0}' . format ( ctid ) ] if template : args . append ( '--ostemplate {0}' . format ( template ) ) yield 'vzctl create {0}' . format ( ' ' . join ( args ) ) | Create OpenVZ containers . |
22,843 | def set ( state , host , ctid , save = True , ** settings ) : args = [ '{0}' . format ( ctid ) ] if save : args . append ( '--save' ) for key , value in six . iteritems ( settings ) : if isinstance ( value , list ) : args . extend ( '--{0} {1}' . format ( key , v ) for v in value ) else : args . append ( '--{0} {1}' . format ( key , value ) ) yield 'vzctl set {0}' . format ( ' ' . join ( args ) ) | Set OpenVZ container details . |
22,844 | def exec_file ( filename , return_locals = False , is_deploy_code = False ) : if filename not in PYTHON_CODES : with open ( filename , 'r' ) as f : code = f . read ( ) code = compile ( code , filename , 'exec' ) PYTHON_CODES [ filename ] = code data = { '__file__' : filename , 'state' : pseudo_state , } exec ( PYTHON_CODES [ filename ] , data ) return data | Execute a Python file and optionally return it s attributes as a dict . |
22,845 | def shell ( state , host , commands , chdir = None ) : if isinstance ( commands , six . string_types ) : commands = [ commands ] for command in commands : if chdir : yield 'cd {0} && ({1})' . format ( chdir , command ) else : yield command | Run raw shell code . |
22,846 | def script ( state , host , filename , chdir = None ) : temp_file = state . get_temp_filename ( filename ) yield files . put ( state , host , filename , temp_file ) yield chmod ( temp_file , '+x' ) if chdir : yield 'cd {0} && {1}' . format ( chdir , temp_file ) else : yield temp_file | Upload and execute a local script on the remote host . |
22,847 | def script_template ( state , host , template_filename , chdir = None , ** data ) : temp_file = state . get_temp_filename ( template_filename ) yield files . template ( state , host , template_filename , temp_file , ** data ) yield chmod ( temp_file , '+x' ) if chdir : yield 'cd {0} && {1}' . format ( chdir , temp_file ) else : yield temp_file | Generate upload and execute a local script template on the remote host . |
22,848 | def hostname ( state , host , hostname , hostname_file = None ) : if hostname_file is None : os = host . fact . os if os == 'Linux' : hostname_file = '/etc/hostname' elif os == 'OpenBSD' : hostname_file = '/etc/myname' current_hostname = host . fact . hostname if current_hostname != hostname : yield 'hostname {0}' . format ( hostname ) if hostname_file : file = six . StringIO ( '{0}\n' . format ( hostname ) ) yield files . put ( state , host , file , hostname_file , ) | Set the system hostname . |
22,849 | def sysctl ( state , host , name , value , persist = False , persist_file = '/etc/sysctl.conf' , ) : string_value = ( ' ' . join ( value ) if isinstance ( value , list ) else value ) existing_value = host . fact . sysctl . get ( name ) if not existing_value or existing_value != value : yield 'sysctl {0}={1}' . format ( name , string_value ) if persist : yield files . line ( state , host , persist_file , '{0}[[:space:]]*=[[:space:]]*{1}' . format ( name , string_value ) , replace = '{0} = {1}' . format ( name , string_value ) , ) | Edit sysctl configuration . |
22,850 | def download ( state , host , source_url , destination , user = None , group = None , mode = None , cache_time = None , force = False , ) : info = host . fact . file ( destination ) if info is False : raise OperationError ( 'Destination {0} already exists and is not a file' . format ( destination ) , ) download = force if info is None : download = True elif cache_time : cache_time = host . fact . date . replace ( tzinfo = None ) - timedelta ( seconds = cache_time ) if info [ 'mtime' ] and info [ 'mtime' ] > cache_time : download = True if download : yield 'wget -q {0} -O {1}' . format ( source_url , destination ) if user or group : yield chown ( destination , user , group ) if mode : yield chmod ( destination , mode ) | Download files from remote locations . |
22,851 | def replace ( state , host , name , match , replace , flags = None ) : yield sed_replace ( name , match , replace , flags = flags ) | A simple shortcut for replacing text in files with sed . |
22,852 | def sync ( state , host , source , destination , user = None , group = None , mode = None , delete = False , exclude = None , exclude_dir = None , add_deploy_dir = True , ) : if not source . endswith ( path . sep ) : source = '{0}{1}' . format ( source , path . sep ) if add_deploy_dir and state . deploy_dir : source = path . join ( state . deploy_dir , source ) if not path . isdir ( source ) : raise IOError ( 'No such directory: {0}' . format ( source ) ) if exclude is not None : if not isinstance ( exclude , ( list , tuple ) ) : exclude = [ exclude ] if exclude_dir is not None : if not isinstance ( exclude_dir , ( list , tuple ) ) : exclude_dir = [ exclude_dir ] put_files = [ ] ensure_dirnames = [ ] for dirname , _ , filenames in walk ( source ) : remote_dirname = dirname . replace ( source , '' ) if exclude_dir and any ( fnmatch ( remote_dirname , match ) for match in exclude_dir ) : continue if remote_dirname : ensure_dirnames . append ( remote_dirname ) for filename in filenames : full_filename = path . join ( dirname , filename ) if exclude and any ( fnmatch ( full_filename , match ) for match in exclude ) : continue put_files . append ( ( full_filename , '/' . join ( item for item in ( destination , remote_dirname , filename ) if item ) , ) ) yield directory ( state , host , destination , user = user , group = group , ) for dirname in ensure_dirnames : yield directory ( state , host , '/' . join ( ( destination , dirname ) ) , user = user , group = group , ) for local_filename , remote_filename in put_files : yield put ( state , host , local_filename , remote_filename , user = user , group = group , mode = mode , add_deploy_dir = False , ) if delete : remote_filenames = set ( host . fact . find_files ( destination ) or [ ] ) wanted_filenames = set ( [ remote_filename for _ , remote_filename in put_files ] ) files_to_delete = remote_filenames - wanted_filenames for filename in files_to_delete : if exclude and any ( fnmatch ( filename , match ) for match in exclude ) : continue yield file ( state , host , filename , present = False ) | Syncs a local directory with a remote one with delete support . Note that delete will remove extra files on the remote side but not extra directories . |
22,853 | def put ( state , host , local_filename , remote_filename , user = None , group = None , mode = None , add_deploy_dir = True , ) : if hasattr ( local_filename , 'read' ) : local_file = local_filename else : if add_deploy_dir and state . deploy_dir : local_filename = path . join ( state . deploy_dir , local_filename ) local_file = local_filename if not path . isfile ( local_file ) : raise IOError ( 'No such file: {0}' . format ( local_file ) ) mode = ensure_mode_int ( mode ) remote_file = host . fact . file ( remote_filename ) if not remote_file : yield ( local_file , remote_filename ) if user or group : yield chown ( remote_filename , user , group ) if mode : yield chmod ( remote_filename , mode ) else : local_sum = get_file_sha1 ( local_filename ) remote_sum = host . fact . sha1_file ( remote_filename ) if local_sum != remote_sum : yield ( local_file , remote_filename ) if user or group : yield chown ( remote_filename , user , group ) if mode : yield chmod ( remote_filename , mode ) else : if mode and remote_file [ 'mode' ] != mode : yield chmod ( remote_filename , mode ) if ( ( user and remote_file [ 'user' ] != user ) or ( group and remote_file [ 'group' ] != group ) ) : yield chown ( remote_filename , user , group ) | Copy a local file to the remote system . |
22,854 | def template ( state , host , template_filename , remote_filename , user = None , group = None , mode = None , ** data ) : if state . deploy_dir : template_filename = path . join ( state . deploy_dir , template_filename ) data [ 'host' ] = host data [ 'inventory' ] = state . inventory try : output = get_template ( template_filename ) . render ( data ) except ( TemplateSyntaxError , UndefinedError ) as e : _ , _ , trace = sys . exc_info ( ) while trace . tb_next : if trace . tb_next . tb_next : trace = trace . tb_next else : break line_number = trace . tb_frame . f_lineno template_lines = open ( template_filename , 'r' ) . readlines ( ) template_lines = [ line . strip ( ) for line in template_lines ] relevant_lines = template_lines [ max ( line_number - 2 , 0 ) : line_number + 1 ] raise OperationError ( 'Error in template: {0} (L{1}): {2}\n...\n{3}\n...' . format ( template_filename , line_number , e , '\n' . join ( relevant_lines ) , ) ) output_file = six . StringIO ( output ) output_file . template = template_filename yield put ( state , host , output_file , remote_filename , user = user , group = group , mode = mode , add_deploy_dir = False , ) | Generate a template and write it to the remote system . |
22,855 | def sql ( state , host , sql , database = None , postgresql_user = None , postgresql_password = None , postgresql_host = None , postgresql_port = None , ) : yield make_execute_psql_command ( sql , database = database , user = postgresql_user , password = postgresql_password , host = postgresql_host , port = postgresql_port , ) | Execute arbitrary SQL against PostgreSQL . |
22,856 | def dump ( state , host , remote_filename , database = None , postgresql_user = None , postgresql_password = None , postgresql_host = None , postgresql_port = None , ) : yield '{0} > {1}' . format ( make_psql_command ( executable = 'pg_dump' , database = database , user = postgresql_user , password = postgresql_password , host = postgresql_host , port = postgresql_port , ) , remote_filename ) | Dump a PostgreSQL database into a . sql file . Requires mysqldump . |
22,857 | def get_fact ( state , host , name ) : if callable ( getattr ( FACTS [ name ] , 'command' , None ) ) : def wrapper ( * args ) : fact_data = get_facts ( state , name , args = args , ensure_hosts = ( host , ) ) return fact_data . get ( host ) return wrapper else : fact_data = get_facts ( state , name , ensure_hosts = ( host , ) ) return fact_data . get ( host ) | Wrapper around get_facts returning facts for one host or a function that does . |
22,858 | def key ( state , host , key = None , keyserver = None , keyid = None ) : if key : if urlparse ( key ) . scheme : yield 'wget -O- {0} | apt-key add -' . format ( key ) else : yield 'apt-key add {0}' . format ( key ) if keyserver and keyid : yield 'apt-key adv --keyserver {0} --recv-keys {1}' . format ( keyserver , keyid ) | Add apt gpg keys with apt - key . |
22,859 | def update ( state , host , cache_time = None , touch_periodic = False ) : if cache_time : cache_info = host . fact . file ( APT_UPDATE_FILENAME ) host_cache_time = host . fact . date . replace ( tzinfo = None ) - timedelta ( seconds = cache_time ) if cache_info and cache_info [ 'mtime' ] and cache_info [ 'mtime' ] > host_cache_time : return yield 'apt-get update' if cache_time : yield 'touch {0}' . format ( APT_UPDATE_FILENAME ) | Updates apt repos . |
22,860 | def run_shell_command ( state , host , command , get_pty = False , timeout = None , print_output = False , ** command_kwargs ) : command = make_command ( command , ** command_kwargs ) logger . debug ( ' . format ( command ) ) if print_output : print ( '{0}>>> {1}' . format ( host . print_prefix , command ) ) process = Popen ( command , shell = True , stdout = PIPE , stderr = PIPE ) stdout_reader = gevent . spawn ( read_buffer , process . stdout , print_output = print_output , print_func = lambda line : '{0}{1}' . format ( host . print_prefix , line ) , ) stderr_reader = gevent . spawn ( read_buffer , process . stderr , print_output = print_output , print_func = lambda line : '{0}{1}' . format ( host . print_prefix , click . style ( line , 'red' ) , ) , ) greenlets = gevent . wait ( ( stdout_reader , stderr_reader ) , timeout = timeout ) if len ( greenlets ) != 2 : stdout_reader . kill ( ) stderr_reader . kill ( ) raise timeout_error ( ) stdout = stdout_reader . get ( ) stderr = stderr_reader . get ( ) logger . debug ( ' ) process . wait ( ) process . stdout . close ( ) logger . debug ( ' . format ( process . returncode ) ) return process . returncode == 0 , stdout , stderr | Execute a command on the local machine . |
22,861 | def upstart ( state , host , name , running = True , restarted = False , reloaded = False , command = None , enabled = None , ) : yield _handle_service_control ( name , host . fact . upstart_status , 'initctl {1} {0}' , running , restarted , reloaded , command , ) if enabled is True : yield files . file ( state , host , '/etc/init/{0}.override' . format ( name ) , present = False , ) elif enabled is False : yield 'echo "manual" > /etc/init/{0}.override' . format ( name ) | Manage the state of upstart managed services . |
22,862 | def service ( state , host , * args , ** kwargs ) : if host . fact . which ( 'systemctl' ) : yield systemd ( state , host , * args , ** kwargs ) return if host . fact . which ( 'initctl' ) : yield upstart ( state , host , * args , ** kwargs ) return if host . fact . directory ( '/etc/init.d' ) : yield d ( state , host , * args , ** kwargs ) return if host . fact . directory ( '/etc/rc.d' ) : yield rc ( state , host , * args , ** kwargs ) return raise OperationError ( ( 'No init system found ' '(no systemctl, initctl, /etc/init.d or /etc/rc.d found)' ) ) | Manage the state of services . This command checks for the presence of all the init systems pyinfra can handle and executes the relevant operation . See init system sepcific operation for arguments . |
22,863 | def connect ( state , host , for_fact = None ) : kwargs = _make_paramiko_kwargs ( state , host ) logger . debug ( 'Connecting to: {0} ({1})' . format ( host . name , kwargs ) ) hostname = kwargs . pop ( 'hostname' , host . data . ssh_hostname or host . name , ) try : client = SSHClient ( ) client . set_missing_host_key_policy ( MissingHostKeyPolicy ( ) ) client . connect ( hostname , ** kwargs ) session = client . get_transport ( ) . open_session ( ) AgentRequestHandler ( session ) log_message = '{0}{1}' . format ( host . print_prefix , click . style ( 'Connected' , 'green' ) , ) if for_fact : log_message = '{0}{1}' . format ( log_message , ' (for {0} fact)' . format ( for_fact ) , ) logger . info ( log_message ) return client except AuthenticationException : auth_kwargs = { } for key , value in kwargs . items ( ) : if key in ( 'username' , 'password' ) : auth_kwargs [ key ] = value continue if key == 'pkey' and value : auth_kwargs [ 'key' ] = host . data . ssh_key auth_args = ', ' . join ( '{0}={1}' . format ( key , value ) for key , value in auth_kwargs . items ( ) ) _log_connect_error ( host , 'Authentication error' , auth_args ) except SSHException as e : _log_connect_error ( host , 'SSH error' , e ) except gaierror : _log_connect_error ( host , 'Could not resolve hostname' , hostname ) except socket_error as e : _log_connect_error ( host , 'Could not connect' , e ) except EOFError as e : _log_connect_error ( host , 'EOF error' , e ) | Connect to a single host . Returns the SSH client if succesful . Stateless by design so can be run in parallel . |
22,864 | def run_shell_command ( state , host , command , get_pty = False , timeout = None , print_output = False , ** command_kwargs ) : command = make_command ( command , ** command_kwargs ) logger . debug ( 'Running command on {0}: (pty={1}) {2}' . format ( host . name , get_pty , command , ) ) if print_output : print ( '{0}>>> {1}' . format ( host . print_prefix , command ) ) _ , stdout_buffer , stderr_buffer = host . connection . exec_command ( command , get_pty = get_pty , ) channel = stdout_buffer . channel stdout_reader = gevent . spawn ( read_buffer , stdout_buffer , print_output = print_output , print_func = lambda line : '{0}{1}' . format ( host . print_prefix , line ) , ) stderr_reader = gevent . spawn ( read_buffer , stderr_buffer , print_output = print_output , print_func = lambda line : '{0}{1}' . format ( host . print_prefix , click . style ( line , 'red' ) , ) , ) greenlets = gevent . wait ( ( stdout_reader , stderr_reader ) , timeout = timeout ) if len ( greenlets ) != 2 : stdout_reader . kill ( ) stderr_reader . kill ( ) raise timeout_error ( ) stdout = stdout_reader . get ( ) stderr = stderr_reader . get ( ) logger . debug ( 'Waiting for exit status...' ) exit_status = channel . recv_exit_status ( ) logger . debug ( 'Command exit status: {0}' . format ( exit_status ) ) return exit_status == 0 , stdout , stderr | Execute a command on the specified host . |
22,865 | def put_file ( state , host , filename_or_io , remote_filename , sudo = False , sudo_user = None , su_user = None , print_output = False , ) : if sudo or su_user : temp_file = state . get_temp_filename ( remote_filename ) _put_file ( host , filename_or_io , temp_file ) if print_output : print ( '{0}file uploaded: {1}' . format ( host . print_prefix , remote_filename ) ) command = 'mv {0} {1}' . format ( temp_file , remote_filename ) if su_user : command = '{0} && chown {1} {2}' . format ( command , su_user , remote_filename ) elif sudo_user : command = '{0} && chown {1} {2}' . format ( command , sudo_user , remote_filename ) status , _ , stderr = run_shell_command ( state , host , command , sudo = sudo , sudo_user = sudo_user , su_user = su_user , print_output = print_output , ) if status is False : logger . error ( 'File error: {0}' . format ( '\n' . join ( stderr ) ) ) return False else : _put_file ( host , filename_or_io , remote_filename ) if print_output : print ( '{0}file uploaded: {1}' . format ( host . print_prefix , remote_filename ) ) return True | Upload file - ios to the specified host using SFTP . Supports uploading files with sudo by uploading to a temporary directory then moving & chowning . |
22,866 | def deploy ( self , name , kwargs , data , line_number , in_deploy = True ) : if self . deploy_name : name = _make_name ( self . deploy_name , name ) old_in_deploy = self . in_deploy old_deploy_name = self . deploy_name old_deploy_kwargs = self . deploy_kwargs old_deploy_data = self . deploy_data old_deploy_line_numbers = self . deploy_line_numbers self . in_deploy = in_deploy if ( old_deploy_kwargs and old_deploy_kwargs . get ( 'hosts' ) is not None ) : if 'hosts' in kwargs : kwargs [ 'hosts' ] = [ host for host in kwargs [ 'hosts' ] if host in old_deploy_kwargs [ 'hosts' ] ] else : kwargs [ 'hosts' ] = old_deploy_kwargs [ 'hosts' ] new_line_numbers = list ( self . deploy_line_numbers or [ ] ) new_line_numbers . append ( line_number ) new_line_numbers = tuple ( new_line_numbers ) self . deploy_name = name self . deploy_kwargs = kwargs self . deploy_data = data self . deploy_line_numbers = new_line_numbers logger . debug ( 'Starting deploy {0} (args={1}, data={2})' . format ( name , kwargs , data , ) ) yield self . in_deploy = old_in_deploy self . deploy_name = old_deploy_name self . deploy_kwargs = old_deploy_kwargs self . deploy_data = old_deploy_data self . deploy_line_numbers = old_deploy_line_numbers logger . debug ( 'Reset deploy to {0} (args={1}, data={2})' . format ( old_deploy_name , old_deploy_kwargs , old_deploy_data , ) ) | Wraps a group of operations as a deploy this should not be used directly instead use pyinfra . api . deploy . deploy . |
22,867 | def activate_host ( self , host ) : logger . debug ( 'Activating host: {0}' . format ( host ) ) self . activated_hosts . add ( host ) self . active_hosts . add ( host ) | Flag a host as active . |
22,868 | def fail_hosts ( self , hosts_to_fail , activated_count = None ) : if not hosts_to_fail : return activated_count = activated_count or len ( self . activated_hosts ) logger . debug ( 'Failing hosts: {0}' . format ( ', ' . join ( ( host . name for host in hosts_to_fail ) , ) ) ) self . active_hosts -= hosts_to_fail active_hosts = self . active_hosts if not active_hosts : raise PyinfraError ( 'No hosts remaining!' ) if self . config . FAIL_PERCENT is not None : percent_failed = ( 1 - len ( active_hosts ) / activated_count ) * 100 if percent_failed > self . config . FAIL_PERCENT : raise PyinfraError ( 'Over {0}% of hosts failed ({1}%)' . format ( self . config . FAIL_PERCENT , int ( round ( percent_failed ) ) , ) ) | Flag a set of hosts as failed error for config . FAIL_PERCENT . |
22,869 | def is_host_in_limit ( self , host ) : limit_hosts = self . limit_hosts if not isinstance ( limit_hosts , list ) : return True return host in limit_hosts | Returns a boolean indicating if the host is within the current state limit . |
22,870 | def get_temp_filename ( self , hash_key = None ) : if not hash_key : hash_key = six . text_type ( uuid4 ( ) ) temp_filename = '{0}/{1}' . format ( self . config . TEMP_DIR , sha1_hash ( hash_key ) , ) return temp_filename | Generate a temporary filename for this deploy . |
22,871 | def _run_server_ops ( state , host , progress = None ) : logger . debug ( 'Running all ops on {0}' . format ( host ) ) for op_hash in state . get_op_order ( ) : op_meta = state . op_meta [ op_hash ] logger . info ( ' . format ( click . style ( ' , 'blue' ) , click . style ( ', ' . join ( op_meta [ 'names' ] ) , bold = True ) , click . style ( host . name , bold = True ) , ) ) result = _run_server_op ( state , host , op_hash ) if progress : progress ( ( host , op_hash ) ) if result is False : raise PyinfraError ( 'Error in operation {0} on {1}' . format ( ', ' . join ( op_meta [ 'names' ] ) , host , ) ) if pyinfra . is_cli : print ( ) | Run all ops for a single server . |
22,872 | def _run_serial_ops ( state ) : for host in list ( state . inventory ) : host_operations = product ( [ host ] , state . get_op_order ( ) ) with progress_spinner ( host_operations ) as progress : try : _run_server_ops ( state , host , progress = progress , ) except PyinfraError : state . fail_hosts ( { host } ) | Run all ops for all servers one server at a time . |
22,873 | def _run_no_wait_ops ( state ) : hosts_operations = product ( state . inventory , state . get_op_order ( ) ) with progress_spinner ( hosts_operations ) as progress : greenlets = [ state . pool . spawn ( _run_server_ops , state , host , progress = progress , ) for host in state . inventory ] gevent . joinall ( greenlets ) | Run all ops for all servers at once . |
22,874 | def _run_single_op ( state , op_hash ) : op_meta = state . op_meta [ op_hash ] op_types = [ ] if op_meta [ 'serial' ] : op_types . append ( 'serial' ) if op_meta [ 'run_once' ] : op_types . append ( 'run once' ) logger . info ( '{0} {1} {2}' . format ( click . style ( ' . format ( ' {0} ' . format ( ', ' . join ( op_types ) ) if op_types else ' ' , ) , 'blue' ) , click . style ( ', ' . join ( op_meta [ 'names' ] ) , bold = True ) , tuple ( op_meta [ 'args' ] ) if op_meta [ 'args' ] else '' , ) ) failed_hosts = set ( ) if op_meta [ 'serial' ] : with progress_spinner ( state . inventory ) as progress : for host in state . inventory : result = _run_server_op ( state , host , op_hash ) progress ( host ) if not result : failed_hosts . add ( host ) else : batches = [ state . inventory ] if op_meta [ 'parallel' ] : parallel = op_meta [ 'parallel' ] hosts = list ( state . inventory ) batches = [ hosts [ i : i + parallel ] for i in range ( 0 , len ( hosts ) , parallel ) ] for batch in batches : with progress_spinner ( batch ) as progress : greenlet_to_host = { state . pool . spawn ( _run_server_op , state , host , op_hash ) : host for host in batch } for greenlet in gevent . iwait ( greenlet_to_host . keys ( ) ) : host = greenlet_to_host [ greenlet ] progress ( host ) for greenlet , host in six . iteritems ( greenlet_to_host ) : if not greenlet . get ( ) : failed_hosts . add ( host ) if not op_meta [ 'ignore_errors' ] : state . fail_hosts ( failed_hosts ) if pyinfra . is_cli : print ( ) | Run a single operation for all servers . Can be configured to run in serial . |
22,875 | def run_ops ( state , serial = False , no_wait = False ) : state . deploying = True if serial : _run_serial_ops ( state ) elif no_wait : _run_no_wait_ops ( state ) for op_hash in state . get_op_order ( ) : _run_single_op ( state , op_hash ) | Runs all operations across all servers in a configurable manner . |
22,876 | def serve ( service_brokers : Union [ List [ ServiceBroker ] , ServiceBroker ] , credentials : Union [ List [ BrokerCredentials ] , BrokerCredentials , None ] , logger : logging . Logger = logging . root , port = 5000 , debug = False ) : from gevent . pywsgi import WSGIServer from flask import Flask app = Flask ( __name__ ) app . debug = debug blueprint = get_blueprint ( service_brokers , credentials , logger ) logger . debug ( "Register openbrokerapi blueprint" ) app . register_blueprint ( blueprint ) logger . info ( "Start Flask on 0.0.0.0:%s" % port ) http_server = WSGIServer ( ( '0.0.0.0' , port ) , app ) http_server . serve_forever ( ) | Starts flask with the given brokers . You can provide a list or just one ServiceBroker |
22,877 | def multi_ping ( dest_addrs , timeout , retry = 0 , ignore_lookup_errors = False ) : retry = int ( retry ) if retry < 0 : retry = 0 timeout = float ( timeout ) if timeout < 0.1 : raise MultiPingError ( "Timeout < 0.1 seconds not allowed" ) retry_timeout = float ( timeout ) / ( retry + 1 ) if retry_timeout < 0.1 : raise MultiPingError ( "Time between ping retries < 0.1 seconds" ) mp = MultiPing ( dest_addrs , ignore_lookup_errors = ignore_lookup_errors ) results = { } retry_count = 0 while retry_count <= retry : mp . send ( ) single_results , no_results = mp . receive ( retry_timeout ) results . update ( single_results ) if not no_results : break retry_count += 1 return results , no_results | Combine send and receive measurement into single function . |
22,878 | def _checksum ( self , msg ) : def carry_around_add ( a , b ) : c = a + b return ( c & 0xffff ) + ( c >> 16 ) s = 0 for i in range ( 0 , len ( msg ) , 2 ) : w = ( msg [ i ] << 8 ) + msg [ i + 1 ] s = carry_around_add ( s , w ) s = ~ s & 0xffff return s | Calculate the checksum of a packet . |
22,879 | def send ( self ) : if not self . _receive_has_been_called : all_addrs = self . _dest_addrs else : all_addrs = [ a for ( i , a ) in list ( self . _id_to_addr . items ( ) ) if i in self . _remaining_ids ] if self . _last_used_id is None : self . _last_used_id = int ( time . time ( ) ) & 0xffff for addr in all_addrs : self . _last_used_id = ( self . _last_used_id + 1 ) & 0xffff self . _id_to_addr [ self . _last_used_id ] = addr self . _send_ping ( addr , payload = struct . pack ( "d" , time . time ( ) ) ) | Send pings to multiple addresses ensuring unique IDs for each request . |
22,880 | def _read_all_from_socket ( self , timeout ) : pkts = [ ] try : self . _sock . settimeout ( timeout ) while True : p = self . _sock . recv ( 64 ) pkts . append ( ( bytearray ( p ) , time . time ( ) ) ) self . _sock . settimeout ( 0 ) except socket . timeout : pass except socket . error as e : if e . errno == errno . EWOULDBLOCK : pass else : raise if self . _ipv6_address_present : try : self . _sock6 . settimeout ( timeout ) while True : p = self . _sock6 . recv ( 128 ) pkts . append ( ( bytearray ( p ) , time . time ( ) ) ) self . _sock6 . settimeout ( 0 ) except socket . timeout : pass except socket . error as e : if e . errno == errno . EWOULDBLOCK : pass else : raise return pkts | Read all packets we currently can on the socket . |
22,881 | async def get ( self ) : try : return self . _parse ( await self . read_registers ( 0 , 16 ) ) except TimeoutError : return { 'ip' : self . ip , 'connected' : False } | Get current state from the Midas gas detector . |
22,882 | def _parse ( self , registers ) : result = { 'ip' : self . ip , 'connected' : True } decoder = BinaryPayloadDecoder . fromRegisters ( registers , byteorder = Endian . Big , wordorder = Endian . Little ) b = [ decoder . decode_bits ( ) , decoder . decode_bits ( ) ] reg_40001 = b [ 1 ] + b [ 0 ] monitor_integer = sum ( 1 << i for i , b in enumerate ( reg_40001 [ : 4 ] ) if b ) result [ 'state' ] = options [ 'monitor state' ] [ monitor_integer ] fault_integer = sum ( 1 << i for i , b in enumerate ( reg_40001 [ 4 : 6 ] ) if b ) result [ 'fault' ] = { 'status' : options [ 'fault status' ] [ fault_integer ] } low , high = reg_40001 [ 6 : 8 ] result [ 'alarm' ] = options [ 'alarm level' ] [ low + high ] decoder . _pointer += 2 result [ 'concentration' ] = decoder . decode_32bit_float ( ) decoder . _pointer += 2 fault_number = decoder . decode_16bit_uint ( ) if fault_number != 0 : code = ( 'm' if fault_number < 30 else 'F' ) + str ( fault_number ) result [ 'fault' ] [ 'code' ] = code result [ 'fault' ] . update ( faults [ code ] ) unit_bit = decoder . decode_bits ( ) . index ( True ) result [ 'units' ] = options [ 'concentration unit' ] [ unit_bit ] decoder . _pointer += 1 result [ 'temperature' ] = decoder . decode_16bit_int ( ) result [ 'life' ] = decoder . decode_16bit_uint ( ) / 24.0 decoder . _pointer += 2 result [ 'flow' ] = decoder . decode_16bit_uint ( ) decoder . _pointer += 2 result [ 'low-alarm threshold' ] = round ( decoder . decode_32bit_float ( ) , 6 ) result [ 'high-alarm threshold' ] = round ( decoder . decode_32bit_float ( ) , 6 ) if result [ 'units' ] == 'ppb' : result [ 'concentration' ] *= 1000 result [ 'low-alarm threshold' ] *= 1000 result [ 'high-alarm threshold' ] *= 1000 return result | Parse the response returning a dictionary . |
22,883 | async def _connect ( self ) : self . waiting = True await self . client . start ( self . ip ) self . waiting = False if self . client . protocol is None : raise IOError ( "Could not connect to '{}'." . format ( self . ip ) ) self . open = True | Start asynchronous reconnect loop . |
22,884 | async def read_registers ( self , address , count ) : registers = [ ] while count > 124 : r = await self . _request ( 'read_holding_registers' , address , 124 ) registers += r . registers address , count = address + 124 , count - 124 r = await self . _request ( 'read_holding_registers' , address , count ) registers += r . registers return registers | Read modbus registers . |
22,885 | async def write_register ( self , address , value , skip_encode = False ) : await self . _request ( 'write_registers' , address , value , skip_encode = skip_encode ) | Write a modbus register . |
22,886 | async def write_registers ( self , address , values , skip_encode = False ) : while len ( values ) > 62 : await self . _request ( 'write_registers' , address , values , skip_encode = skip_encode ) address , values = address + 124 , values [ 62 : ] await self . _request ( 'write_registers' , address , values , skip_encode = skip_encode ) | Write modbus registers . |
22,887 | async def _request ( self , method , * args , ** kwargs ) : if not self . open : await self . _connect ( ) while self . waiting : await asyncio . sleep ( 0.1 ) if self . client . protocol is None or not self . client . protocol . connected : raise TimeoutError ( "Not connected to device." ) try : future = getattr ( self . client . protocol , method ) ( * args , ** kwargs ) except AttributeError : raise TimeoutError ( "Not connected to device." ) self . waiting = True try : return await asyncio . wait_for ( future , timeout = self . timeout ) except asyncio . TimeoutError as e : if self . open : if hasattr ( self , 'modbus' ) : self . client . protocol_lost_connection ( self . modbus ) self . open = False raise TimeoutError ( e ) except pymodbus . exceptions . ConnectionException as e : raise ConnectionError ( e ) finally : self . waiting = False | Send a request to the device and awaits a response . |
22,888 | def _close ( self ) : self . client . stop ( ) self . open = False self . waiting = False | Close the TCP connection . |
22,889 | def command_line ( ) : import argparse import asyncio import json parser = argparse . ArgumentParser ( description = "Read a Honeywell Midas gas " "detector state from the command line." ) parser . add_argument ( 'address' , help = "The IP address of the gas detector." ) args = parser . parse_args ( ) async def get ( ) : async with GasDetector ( args . address ) as detector : print ( json . dumps ( await detector . get ( ) , indent = 4 , sort_keys = True ) ) loop = asyncio . get_event_loop ( ) loop . run_until_complete ( get ( ) ) loop . close ( ) | Command - line tool for Midas gas detector communication . |
22,890 | def build_masked_loss ( loss_function , mask_value ) : def masked_loss_function ( y_true , y_pred ) : mask = K . cast ( K . not_equal ( y_true , mask_value ) , K . floatx ( ) ) return loss_function ( y_true * mask , y_pred * mask ) return masked_loss_function | Builds a loss function that masks based on targets |
22,891 | def Run ( self ) : if not self . executable : logging . error ( 'Could not locate "%s"' % self . long_name ) return 0 finfo = os . stat ( self . executable ) self . date = time . localtime ( finfo [ stat . ST_MTIME ] ) logging . info ( 'Running: %s %s </dev/null 2>&1' % ( self . executable , FLAGS . help_flag ) ) ( child_stdin , child_stdout_and_stderr ) = os . popen4 ( [ self . executable , FLAGS . help_flag ] ) child_stdin . close ( ) self . output = child_stdout_and_stderr . readlines ( ) child_stdout_and_stderr . close ( ) if len ( self . output ) < _MIN_VALID_USAGE_MSG : logging . error ( 'Error: "%s %s" returned only %d lines: %s' % ( self . name , FLAGS . help_flag , len ( self . output ) , self . output ) ) return 0 return 1 | Run it and collect output . |
22,892 | def Parse ( self ) : ( start_line , lang ) = self . ParseDesc ( ) if start_line < 0 : return if 'python' == lang : self . ParsePythonFlags ( start_line ) elif 'c' == lang : self . ParseCFlags ( start_line ) elif 'java' == lang : self . ParseJavaFlags ( start_line ) | Parse program output . |
22,893 | def ParseDesc ( self , start_line = 0 ) : exec_mod_start = self . executable + ':' after_blank = 0 start_line = 0 for start_line in range ( start_line , len ( self . output ) ) : line = self . output [ start_line ] . rstrip ( ) if ( 'flags:' == line and len ( self . output ) > start_line + 1 and '' == self . output [ start_line + 1 ] . rstrip ( ) ) : start_line += 2 logging . debug ( 'Flags start (python): %s' % line ) return ( start_line , 'python' ) if exec_mod_start == line : logging . debug ( 'Flags start (swig): %s' % line ) return ( start_line , 'python' ) if after_blank and line . startswith ( ' Flags from ' ) : logging . debug ( 'Flags start (c): %s' % line ) return ( start_line , 'c' ) if line == 'where flags are' : logging . debug ( 'Flags start (java): %s' % line ) start_line += 2 return ( start_line , 'java' ) logging . debug ( 'Desc: %s' % line ) self . desc . append ( line ) after_blank = ( line == '' ) else : logging . warn ( 'Never found the start of the flags section for "%s"!' % self . long_name ) return ( - 1 , '' ) | Parse the initial description . |
22,894 | def ParseCFlags ( self , start_line = 0 ) : modname = None modlist = [ ] flag = None for line_num in range ( start_line , len ( self . output ) ) : line = self . output [ line_num ] . rstrip ( ) if not line : if flag : modlist . append ( flag ) flag = None continue mobj = self . module_c_re . match ( line ) if mobj : modname = mobj . group ( 1 ) logging . debug ( 'Module: %s' % line ) if flag : modlist . append ( flag ) self . module_list . append ( modname ) self . modules . setdefault ( modname , [ ] ) modlist = self . modules [ modname ] flag = None continue mobj = self . flag_c_re . match ( line ) if mobj : if flag : modlist . append ( flag ) logging . debug ( 'Flag: %s' % line ) flag = Flag ( mobj . group ( 1 ) , mobj . group ( 2 ) ) continue if flag : flag . help += ' ' + line . strip ( ) else : logging . info ( 'Extra: %s' % line ) if flag : modlist . append ( flag ) | Parse C style flags . |
22,895 | def Filter ( self ) : if not self . desc : self . short_desc = '' return for i in range ( len ( self . desc ) ) : if self . desc [ i ] . find ( self . executable ) >= 0 : self . desc [ i ] = self . desc [ i ] . replace ( self . executable , self . name ) self . short_desc = self . desc [ 0 ] word_list = self . short_desc . split ( ' ' ) all_names = [ self . name , self . short_name , ] while word_list and ( word_list [ 0 ] in all_names or word_list [ 0 ] . lower ( ) in all_names ) : del word_list [ 0 ] self . short_desc = '' if not self . short_desc and word_list : self . short_desc = ' ' . join ( word_list ) | Filter parsed data to create derived fields . |
22,896 | def Output ( self ) : self . Open ( ) self . Header ( ) self . Body ( ) self . Footer ( ) | Output all sections of the page . |
22,897 | def GetFlagSuggestions ( attempt , longopt_list ) : if len ( attempt ) <= 2 or not longopt_list : return [ ] option_names = [ v . split ( '=' ) [ 0 ] for v in longopt_list ] distances = [ ( _DamerauLevenshtein ( attempt , option [ 0 : len ( attempt ) ] ) , option ) for option in option_names ] distances . sort ( key = lambda t : t [ 0 ] ) least_errors , _ = distances [ 0 ] if least_errors >= _SUGGESTION_ERROR_RATE_THRESHOLD * len ( attempt ) : return [ ] suggestions = [ ] for errors , name in distances : if errors == least_errors : suggestions . append ( name ) else : break return suggestions | Get helpful similar matches for an invalid flag . |
22,898 | def _DamerauLevenshtein ( a , b ) : memo = { } def Distance ( x , y ) : if ( x , y ) in memo : return memo [ x , y ] if not x : d = len ( y ) elif not y : d = len ( x ) else : d = min ( Distance ( x [ 1 : ] , y ) + 1 , Distance ( x , y [ 1 : ] ) + 1 , Distance ( x [ 1 : ] , y [ 1 : ] ) + ( x [ 0 ] != y [ 0 ] ) ) if len ( x ) >= 2 and len ( y ) >= 2 and x [ 0 ] == y [ 1 ] and x [ 1 ] == y [ 0 ] : t = Distance ( x [ 2 : ] , y [ 2 : ] ) + 1 if d > t : d = t memo [ x , y ] = d return d return Distance ( a , b ) | Damerau - Levenshtein edit distance from a to b . |
22,899 | def FlagDictToArgs ( flag_map ) : for key , value in six . iteritems ( flag_map ) : if value is None : yield '--%s' % key elif isinstance ( value , bool ) : if value : yield '--%s' % key else : yield '--no%s' % key elif isinstance ( value , ( bytes , type ( u'' ) ) ) : yield '--%s=%s' % ( key , value ) else : try : yield '--%s=%s' % ( key , ',' . join ( str ( item ) for item in value ) ) except TypeError : yield '--%s=%s' % ( key , value ) | Convert a dict of values into process call parameters . |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.