idx
int64
0
252k
question
stringlengths
48
5.28k
target
stringlengths
5
1.23k
5,800
def extract ( obj , pointer , bypass_ref = False ) : return Pointer ( pointer ) . extract ( obj , bypass_ref )
Extract member or element of obj according to pointer .
5,801
def aa_counts ( aln , weights = None , gap_chars = '-.' ) : if weights is None : counts = Counter ( ) for rec in aln : seq_counts = Counter ( str ( rec . seq ) ) counts . update ( seq_counts ) else : if weights == True : weights = sequence_weights ( aln ) else : assert len ( weights ) == len ( aln ) , ( "Length mismatc...
Calculate the amino acid frequencies in a set of SeqRecords .
5,802
def aa_frequencies ( aln , weights = None , gap_chars = '-.' ) : counts = aa_counts ( aln , weights , gap_chars ) scale = 1.0 / sum ( counts . values ( ) ) return dict ( ( aa , cnt * scale ) for aa , cnt in counts . iteritems ( ) )
Frequency of each residue type in an alignment .
5,803
def blocks ( aln , threshold = 0.5 , weights = None ) : assert len ( aln ) if weights == False : def pct_nongaps ( col ) : return 1 - ( float ( col . count ( '-' ) ) / len ( col ) ) else : if weights in ( None , True ) : weights = sequence_weights ( aln , 'avg1' ) def pct_nongaps ( col ) : assert len ( col ) == len ( w...
Remove gappy columns from an alignment .
5,804
def col_counts ( col , weights = None , gap_chars = '-.' ) : cnt = defaultdict ( float ) for aa , wt in zip ( col , weights ) : if aa not in gap_chars : cnt [ aa ] += wt return cnt
Absolute counts of each residue type in a single column .
5,805
def remove_empty_cols ( records ) : records = list ( records ) seqstrs = [ str ( rec . seq ) for rec in records ] clean_cols = [ col for col in zip ( * seqstrs ) if not all ( c == '-' for c in col ) ] clean_seqs = [ '' . join ( row ) for row in zip ( * clean_cols ) ] for rec , clean_seq in zip ( records , clean_seqs ) ...
Remove all - gap columns from aligned SeqRecords .
5,806
def sequence_weights ( aln , scaling = 'none' , gap_chars = '-.' ) : expectk = [ 0.0 , 1.0 , 1.953 , 2.861 , 3.705 , 4.524 , 5.304 , 6.026 , 6.724 , 7.397 , 8.04 , 8.622 , 9.191 , 9.739 , 10.264 , 10.758 , 11.194 , 11.635 , 12.049 , 12.468 , 12.806 , 13.185 , 13.539 , 13.863 , 14.177 , 14.466 , 14.737 , 15.005 , 15.245...
Weight aligned sequences to emphasize more divergent members .
5,807
def to_graph ( alnfname , weight_func ) : import networkx G = networkx . Graph ( ) aln = AlignIO . read ( alnfname , 'fasta' ) for i , arec in enumerate ( aln ) : for brec in aln [ i + 1 : ] : ident = weight_func ( str ( arec . seq ) , str ( brec . seq ) ) G . add_edge ( arec . id , brec . id , weight = ident ) return ...
Create a NetworkX graph from a sequence alignment .
5,808
def guidance_UV ( index ) : if 0 < index < 3 : guidance = "Low exposure. No protection required. You can safely stay outside" elif 2 < index < 6 : guidance = "Moderate exposure. Seek shade during midday hours, cover up and wear sunscreen" elif 5 < index < 8 : guidance = "High exposure. Seek shade during midday hours, c...
Return Met Office guidance regarding UV exposure based on UV index
5,809
def parse_sitelist ( sitelist ) : sites = [ ] for site in sitelist [ "Locations" ] [ "Location" ] : try : ident = site [ "id" ] name = site [ "name" ] except KeyError : ident = site [ "@id" ] name = site [ "@name" ] if "latitude" in site : lat = float ( site [ "latitude" ] ) lon = float ( site [ "longitude" ] ) else : ...
Return list of Site instances from retrieved sitelist data
5,810
def _query ( self , data_category , resource_category , field , request , step , isotime = None ) : rest_url = "/" . join ( [ HOST , data_category , resource_category , field , DATA_TYPE , request ] ) query_string = "?" + "&" . join ( [ "res=" + step , "time=" + isotime if isotime is not None else "" , "key=" + self . ...
Request and return data from DataPoint RESTful API .
5,811
def stand_alone_imagery ( self ) : return json . loads ( self . _query ( IMAGE , FORECAST , SURFACE_PRESSURE , CAPABILITIES , "" ) . decode ( errors = "replace" ) )
Returns capabilities data for stand alone imagery and includes URIs for the images .
5,812
def map_overlay_forecast ( self ) : return json . loads ( self . _query ( LAYER , FORECAST , ALL , CAPABILITIES , "" ) . decode ( errors = "replace" ) )
Returns capabilities data for forecast map overlays .
5,813
def map_overlay_obs ( self ) : return json . loads ( self . _query ( LAYER , OBSERVATIONS , ALL , CAPABILITIES , "" ) . decode ( errors = "replace" ) )
Returns capabilities data for observation map overlays .
5,814
def load_and_append ( instrument_dict , instruments = None , raise_errors = False ) : if instruments is None : instruments = { } updated_instruments = { } updated_instruments . update ( instruments ) loaded_failed = { } for instrument_name , instrument_class_name in instrument_dict . items ( ) : instrument_settings = N...
load instrument from instrument_dict and append to instruments
5,815
def fail ( self , reason , obj , pointer = None ) : pointer = pointer_join ( pointer ) err = ValidationError ( reason , obj , pointer ) if self . fail_fast : raise err else : self . errors . append ( err ) return err
Called when validation fails .
5,816
def get_locus ( sequences , kir = False , verbose = False , refdata = None , evalue = 10 ) : if not refdata : refdata = ReferenceData ( ) file_id = str ( randomid ( ) ) input_fasta = file_id + ".fasta" output_xml = file_id + ".xml" SeqIO . write ( sequences , input_fasta , "fasta" ) blastn_cline = NcbiblastnCommandline...
Gets the locus of the sequence by running blastn
5,817
def address ( self ) -> str : return str ( self . _public_key . to_address ( net_query ( self . network ) ) )
generate an address from pubkey
5,818
def sign_transaction ( self , txins : Union [ TxOut ] , tx : MutableTransaction ) -> MutableTransaction : solver = P2pkhSolver ( self . _private_key ) return tx . spend ( txins , [ solver for i in txins ] )
sign the parent txn outputs P2PKH
5,819
def format_name ( net : str ) -> str : if net . startswith ( 't' ) or 'testnet' in net : net = net [ 1 : ] + '-test' else : net = net return net
take care of specifics of cryptoid naming system
5,820
def get_url ( url : str ) -> Union [ dict , int , float , str ] : request = Request ( url , headers = { "User-Agent" : "pypeerassets" } ) response = cast ( HTTPResponse , urlopen ( request ) ) if response . status != 200 : raise Exception ( response . reason ) return json . loads ( response . read ( ) . decode ( ) )
Perform a GET request for the url and return a dictionary parsed from the JSON response .
5,821
def _scan_nodes ( nodelist , context , instance_types , current_block = None , ignore_blocks = None ) : results = [ ] for node in nodelist : if isinstance ( node , instance_types ) : results . append ( node ) elif isinstance ( node , IncludeNode ) : if node . template : if not callable ( getattr ( node . template , 're...
Loop through all nodes of a single scope level .
5,822
def get_node_instances ( nodelist , instances ) : context = _get_main_context ( nodelist ) if TemplateAdapter is not None and isinstance ( nodelist , TemplateAdapter ) : nodelist = nodelist . template return _scan_nodes ( nodelist , context , instances )
Find the nodes of a given instance .
5,823
def get_config_file ( ) : parser = argparse . ArgumentParser ( description = "Read configuration file." ) parser . add_argument ( '-ini' , help = "Full path of configuration file" ) args = parser . parse_args ( ) ini_file = args . ini if not FileClass . is_file_exists ( ini_file ) : print ( "Usage: -ini <full path to t...
Get model configuration file name from argv
5,824
def isnumerical ( x ) : try : xx = float ( x ) except TypeError : return False except ValueError : return False except Exception : return False else : return True
Check the input x is numerical or not .
5,825
def rsquare ( obsvalues , simvalues ) : if len ( obsvalues ) != len ( simvalues ) : raise ValueError ( "The size of observed and simulated values must be " "the same for R-square calculation!" ) if not isinstance ( obsvalues , numpy . ndarray ) : obsvalues = numpy . array ( obsvalues ) if not isinstance ( simvalues , n...
Calculate Coefficient of determination .
5,826
def rmse ( obsvalues , simvalues ) : if len ( obsvalues ) != len ( simvalues ) : raise ValueError ( "The size of observed and simulated values must be " "the same for R-square calculation!" ) if not isinstance ( obsvalues , numpy . ndarray ) : obsvalues = numpy . array ( obsvalues ) if not isinstance ( simvalues , nump...
Calculate RMSE .
5,827
def pbias ( obsvalues , simvalues ) : if len ( obsvalues ) != len ( simvalues ) : raise ValueError ( "The size of observed and simulated values must be" " the same for PBIAS calculation!" ) return sum ( map ( lambda x , y : ( x - y ) * 100 , obsvalues , simvalues ) ) / sum ( obsvalues )
Calculate PBIAS or percent model bias .
5,828
def convert_str2num ( unicode_str ) : if MathClass . isnumerical ( unicode_str ) : unicode_str = float ( unicode_str ) if unicode_str % 1. == 0. : unicode_str = int ( unicode_str ) return unicode_str elif is_string ( unicode_str ) : return str ( unicode_str ) elif isinstance ( unicode_str , tuple ) : return tuple ( Str...
Convert string to string integer or float . Support tuple or list .
5,829
def string_in_list ( tmp_str , strlist ) : new_str_list = strlist [ : ] for i , str_in_list in enumerate ( new_str_list ) : new_str_list [ i ] = str_in_list . lower ( ) return tmp_str . lower ( ) in new_str_list
Is tmp_str in strlist case insensitive .
5,830
def is_file_exists ( filename ) : if filename is None or not os . path . exists ( filename ) or not os . path . isfile ( filename ) : return False else : return True
Check the existence of file path .
5,831
def is_dir_exists ( dirpath ) : if dirpath is None or not os . path . exists ( dirpath ) or not os . path . isdir ( dirpath ) : return False else : return True
Check the existence of folder path .
5,832
def copy_files ( filename , dstfilename ) : FileClass . remove_files ( dstfilename ) dst_prefix = os . path . splitext ( dstfilename ) [ 0 ] pattern = os . path . splitext ( filename ) [ 0 ] + '.*' for f in glob . iglob ( pattern ) : ext = os . path . splitext ( f ) [ 1 ] dst = dst_prefix + ext copy ( f , dst )
Copy files with the same name and different suffixes such as ESRI Shapefile .
5,833
def remove_files ( filename ) : pattern = os . path . splitext ( filename ) [ 0 ] + '.*' for f in glob . iglob ( pattern ) : os . remove ( f )
Delete all files with same root as fileName i . e . regardless of suffix such as ESRI shapefile
5,834
def is_up_to_date ( outfile , basedatetime ) : if os . path . exists ( outfile ) : if os . path . getmtime ( outfile ) >= basedatetime : return True return False
Return true if outfile exists and is no older than base datetime .
5,835
def get_executable_fullpath ( name , dirname = None ) : if name is None : return None if is_string ( name ) : name = str ( name ) else : raise RuntimeError ( 'The input function name or path must be string!' ) if dirname is not None : dirname = os . path . abspath ( dirname ) fpth = dirname + os . sep + name if os . pa...
get the full path of a given executable name
5,836
def get_file_fullpath ( name , dirname = None ) : if name is None : return None if is_string ( name ) : name = str ( name ) else : raise RuntimeError ( 'The input function name or path must be string!' ) for sep in [ '\\' , '/' , os . sep ] : if sep in name : name = os . path . abspath ( name ) return name if dirname i...
Return full path if available .
5,837
def get_filename_by_suffixes ( dir_src , suffixes ) : list_files = os . listdir ( dir_src ) re_files = list ( ) if is_string ( suffixes ) : suffixes = [ suffixes ] if not isinstance ( suffixes , list ) : return None for i , suf in enumerate ( suffixes ) : if len ( suf ) >= 1 and suf [ 0 ] != '.' : suffixes [ i ] = '.' ...
get file names with the given suffixes in the given directory
5,838
def get_full_filename_by_suffixes ( dir_src , suffixes ) : file_names = FileClass . get_filename_by_suffixes ( dir_src , suffixes ) if file_names is None : return None return list ( dir_src + os . sep + name for name in file_names )
get full file names with the given suffixes in the given directory
5,839
def get_core_name_without_suffix ( file_path ) : if '\\' in file_path : file_path = file_path . replace ( '\\' , '/' ) file_name = os . path . basename ( file_path ) core_names = file_name . split ( '.' ) if len ( core_names ) > 1 : core_names = core_names [ : - 1 ] if isinstance ( core_names , list ) : return str ( '....
Return core file name without suffix .
5,840
def add_postfix ( file_path , postfix ) : cur_sep = '' for sep in [ '\\' , '/' , os . sep ] : if sep in file_path : cur_sep = sep break corename = FileClass . get_core_name_without_suffix ( file_path ) tmpspliter = os . path . basename ( file_path ) . split ( '.' ) suffix = '' if len ( tmpspliter ) > 1 : suffix = tmpsp...
Add postfix for a full file path .
5,841
def day_of_year ( dt ) : sec = time . mktime ( dt . timetuple ( ) ) t = time . localtime ( sec ) return t . tm_yday
Day index of year from 1 to 365 or 366
5,842
def run_command ( commands ) : use_shell = False subprocess_flags = 0 startupinfo = None if sysstr == 'Windows' : if isinstance ( commands , list ) : commands = ' ' . join ( str ( c ) for c in commands ) import ctypes SEM_NOGPFAULTERRORBOX = 0x0002 ctypes . windll . kernel32 . SetErrorMode ( SEM_NOGPFAULTERRORBOX ) sub...
Execute external command and return the output lines list . In windows refers to handling - subprocess - crash - in - windows _ .
5,843
def current_path ( local_function ) : from inspect import getsourcefile fpath = getsourcefile ( local_function ) if fpath is None : return None return os . path . dirname ( os . path . abspath ( fpath ) )
Get current path refers to how - do - i - get - the - path - of - the - current - executed - file - in - python _
5,844
def mkdir ( dir_path ) : if not os . path . isdir ( dir_path ) or not os . path . exists ( dir_path ) : os . makedirs ( dir_path )
Make directory if not existed
5,845
def rmmkdir ( dir_path ) : if not os . path . isdir ( dir_path ) or not os . path . exists ( dir_path ) : os . makedirs ( dir_path ) else : rmtree ( dir_path , True ) os . makedirs ( dir_path )
If directory existed then remove and make ; else make it .
5,846
def print_msg ( contentlist ) : if isinstance ( contentlist , list ) or isinstance ( contentlist , tuple ) : return '\n' . join ( contentlist ) else : if len ( contentlist ) > 1 and contentlist [ - 1 ] != '\n' : contentlist += '\n' return contentlist
concatenate message list as single string with line feed .
5,847
def decode_strs_in_dict ( unicode_dict ) : unicode_dict = { StringClass . convert_str2num ( k ) : StringClass . convert_str2num ( v ) for k , v in iteritems ( unicode_dict ) } for k , v in iteritems ( unicode_dict ) : if isinstance ( v , dict ) : unicode_dict [ k ] = UtilClass . decode_strs_in_dict ( v ) return unicode...
Decode strings in dictionary which may contains unicode strings or numeric values .
5,848
def undo ( self ) : self . undo_manager . undo ( ) self . notify_observers ( ) logging . debug ( 'undo_manager undo stack={}' . format ( self . undo_manager . _undo_stack ) )
Rewind the game to the previous state .
5,849
def redo ( self ) : self . undo_manager . redo ( ) self . notify_observers ( ) logging . debug ( 'undo_manager redo stack={}' . format ( self . undo_manager . _redo_stack ) )
Redo the latest undone command .
5,850
def net_query ( name : str ) -> Constants : for net_params in networks : if name in ( net_params . name , net_params . shortname , ) : return net_params raise UnsupportedNetwork
Find the NetworkParams for a network by its long or short name . Raises UnsupportedNetwork if no NetworkParams is found .
5,851
def get_port_at ( self , tile_id , direction ) : for port in self . ports : if port . tile_id == tile_id and port . direction == direction : return port port = Port ( tile_id , direction , PortType . none ) self . ports . append ( port ) return port
If no port is found a new none port is made and added to self . ports .
5,852
def rotate_ports ( self ) : for port in self . ports : port . tile_id = ( ( port . tile_id + 1 ) % len ( hexgrid . coastal_tile_ids ( ) ) ) + 1 port . direction = hexgrid . rotate_direction ( hexgrid . EDGE , port . direction , ccw = True ) self . notify_observers ( )
Rotates the ports 90 degrees . Useful when using the default port setup but the spectator is watching at a rotated angle from true north .
5,853
def intersect_keys ( keys , reffile , cache = False , clean_accs = False ) : index = None if cache : refcache = reffile + '.sqlite' if os . path . exists ( refcache ) : if os . stat ( refcache ) . st_mtime < os . stat ( reffile ) . st_mtime : logging . warn ( "Outdated cache; rebuilding index" ) else : try : index = ( ...
Extract SeqRecords from the index by matching keys .
5,854
def aa_frequencies ( seq , gap_chars = '-.' ) : aa_counts = Counter ( seq ) for gap_char in gap_chars : if gap_char in aa_counts : del aa_counts [ gap_char ] scale = 1.0 / sum ( aa_counts . values ( ) ) return dict ( ( aa , cnt * scale ) for aa , cnt in aa_counts . iteritems ( ) )
Calculate the amino acid frequencies in a sequence set .
5,855
def giving ( self ) : logging . debug ( 'give={}' . format ( self . _give ) ) c = Counter ( self . _give . copy ( ) ) return [ ( n , t ) for t , n in c . items ( ) ]
Returns tuples corresponding to the number and type of each resource in the trade from giver - > getter
5,856
def getting ( self ) : c = Counter ( self . _get . copy ( ) ) return [ ( n , t ) for t , n in c . items ( ) ]
Returns tuples corresponding to the number and type of each resource in the trade from getter - > giver
5,857
def family_check ( self ) : self . logger . info ( "Checking family relations for {0}" . format ( self . family_id ) ) for individual_id in self . individuals : self . logger . debug ( "Checking individual {0}" . format ( individual_id ) ) individual = self . individuals [ individual_id ] self . logger . debug ( "Check...
Check if the family members break the structure of the family . eg . nonexistent parent wrong sex on parent etc . Also extracts all trios found this is of help for many at the moment since GATK can only do phasing of trios and duos .
5,858
def check_parent ( self , parent_id , father = False ) : self . logger . debug ( "Checking parent {0}" . format ( parent_id ) ) if parent_id != '0' : if parent_id not in self . individuals : raise PedigreeError ( self . family_id , parent_id , 'Parent is not in family.' ) if father : if self . individuals [ parent_id ]...
Check if the parent info is correct . If an individual is not present in file raise exeption .
5,859
def to_ped ( self , outfile = None ) : ped_header = [ '#FamilyID' , 'IndividualID' , 'PaternalID' , 'MaternalID' , 'Sex' , 'Phenotype' , ] extra_headers = [ 'InheritanceModel' , 'Proband' , 'Consultand' , 'Alive' ] for individual_id in self . individuals : individual = self . individuals [ individual_id ] for info in i...
Print the individuals of the family in ped format The header will be the original ped header plus all headers found in extra info of the individuals
5,860
def find_deck ( provider : Provider , key : str , version : int , prod : bool = True ) -> Optional [ Deck ] : pa_params = param_query ( provider . network ) if prod : p2th = pa_params . P2TH_addr else : p2th = pa_params . test_P2TH_addr rawtx = provider . getrawtransaction ( key , 1 ) deck = deck_parser ( ( provider , ...
Find specific deck by deck id .
5,861
def deck_spawn ( provider : Provider , deck : Deck , inputs : dict , change_address : str , locktime : int = 0 ) -> Transaction : network_params = net_query ( deck . network ) pa_params = param_query ( deck . network ) if deck . production : p2th_addr = pa_params . P2TH_addr else : p2th_addr = pa_params . test_P2TH_add...
Creates Deck spawn raw transaction .
5,862
def get_card_transfer ( provider : Provider , deck : Deck , txid : str , debug : bool = False ) -> Iterator : rawtx = provider . getrawtransaction ( txid , 1 ) bundle = card_bundler ( provider , deck , rawtx ) return card_bundle_parser ( bundle , debug )
get a single card transfer by it s id
5,863
def find_all_valid_cards ( provider : Provider , deck : Deck ) -> Generator : unfiltered = ( card for batch in get_card_bundles ( provider , deck ) for card in batch ) for card in validate_card_issue_modes ( deck . issue_mode , list ( unfiltered ) ) : yield card
find all the valid cards on this deck filtering out cards which don t play nice with deck issue mode
5,864
def card_transfer ( provider : Provider , card : CardTransfer , inputs : dict , change_address : str , locktime : int = 0 ) -> Transaction : network_params = net_query ( provider . network ) pa_params = param_query ( provider . network ) if card . deck_p2th is None : raise Exception ( "card.deck_p2th required for tx_ou...
Prepare the CardTransfer Transaction object
5,865
def rfc3339_to_datetime ( data ) : try : ts = time . strptime ( data , '%Y-%m-%d' ) return date ( * ts [ : 3 ] ) except ValueError : pass try : dt , _ , tz = data . partition ( 'Z' ) if tz : tz = offset ( tz ) else : tz = offset ( '00:00' ) if '.' in dt and dt . rsplit ( '.' , 1 ) [ - 1 ] . isdigit ( ) : ts = time . st...
convert a rfc3339 date representation into a Python datetime
5,866
def log_config ( verbose = 1 ) : if verbose == 0 : level = logging . WARNING fmt = "%(module)s: %(message)s" elif verbose == 1 : level = logging . INFO fmt = "%(module)s [@%(lineno)s]: %(message)s" else : level = logging . DEBUG fmt = "%(module)s [%(lineno)s]: %(levelname)s: %(message)s" logging . basicConfig ( format ...
Set up logging the way I like it .
5,867
def refresh_instruments ( self ) : def list_access_nested_dict ( dict , somelist ) : return reduce ( operator . getitem , somelist , dict ) def update ( item ) : if item . isExpanded ( ) : for index in range ( item . childCount ( ) ) : child = item . child ( index ) if child . childCount ( ) == 0 : instrument , path_to...
if self . tree_settings has been expanded ask instruments for their actual values
5,868
def update_parameters ( self , treeWidget ) : if treeWidget == self . tree_settings : item = treeWidget . currentItem ( ) instrument , path_to_instrument = item . get_instrument ( ) dictator = item . value for element in path_to_instrument : dictator = { element : dictator } old_value = instrument . settings path_to_in...
updates the internal dictionaries for scripts and instruments with values from the respective trees
5,869
def script_finished ( self ) : script = self . current_script script . updateProgress . disconnect ( self . update_status ) self . script_thread . started . disconnect ( ) script . finished . disconnect ( ) self . current_script = None self . plot_script ( script ) self . progressBar . setValue ( 100 ) self . btn_start...
waits for the script to emit the script_finshed signal
5,870
def update_probes ( self , progress ) : new_values = self . read_probes . probes_values probe_count = len ( self . read_probes . probes ) if probe_count > self . tree_probes . topLevelItemCount ( ) : self . fill_treewidget ( self . tree_probes , new_values ) else : for x in range ( probe_count ) : topLvlItem = self . t...
update the probe tree
5,871
def update_script_from_item ( self , item ) : script , path_to_script , script_item = item . get_script ( ) dictator = list ( script_item . to_dict ( ) . values ( ) ) [ 0 ] for instrument in list ( script . instruments . keys ( ) ) : script . instruments [ instrument ] [ 'settings' ] = dictator [ instrument ] [ 'settin...
updates the script based on the information provided in item
5,872
def message_search ( self , text , on_success , peer = None , min_date = None , max_date = None , max_id = None , offset = 0 , limit = 255 ) : raise TWXUnsupportedMethod ( )
Unsupported in the Bot API
5,873
def remove ( self , pointer ) : doc = deepcopy ( self . document ) parent , obj = None , doc try : for token in Pointer ( pointer ) : parent , obj = obj , token . extract ( obj , bypass_ref = True ) if isinstance ( parent , Mapping ) : del parent [ token ] if isinstance ( parent , MutableSequence ) : parent . pop ( int...
Remove element from sequence member from mapping .
5,874
def _netname ( name : str ) -> dict : try : long = net_query ( name ) . name short = net_query ( name ) . shortname except AttributeError : raise UnsupportedNetwork ( ) return { 'long' : long , 'short' : short }
resolute network name required because some providers use shortnames and other use longnames .
5,875
def sendrawtransaction ( cls , rawtxn : str ) -> str : if cls . is_testnet : url = 'https://testnet-explorer.peercoin.net/api/sendrawtransaction?hex={0}' . format ( rawtxn ) else : url = 'https://explorer.peercoin.net/api/sendrawtransaction?hex={0}' . format ( rawtxn ) resp = urllib . request . urlopen ( url ) return r...
sendrawtransaction remote API
5,876
def validateaddress ( self , address : str ) -> bool : "Returns True if the passed address is valid, False otherwise." try : Address . from_string ( address , self . network_properties ) except InvalidAddress : return False return True
Returns True if the passed address is valid False otherwise .
5,877
def chunker ( l , n ) : for i in ranger ( 0 , len ( l ) , n ) : yield l [ i : i + n ]
Generates n - sized chunks from the list l
5,878
def post ( self , endpoint , data , parallelism = 5 ) : headers = { "Content-Type" : "application/json" , "Accept" : "application/json" , "x-standardize-only" : "true" if self . standardize else "false" , "x-include-invalid" : "true" if self . invalid else "false" , "x-accept-keypair" : "true" if self . accept_keypair ...
Executes most of the request .
5,879
def _cache_init ( self ) : cache_ = cache . get ( self . CACHE_KEY ) if cache_ is None : cache_ = defaultdict ( dict ) self . _cache = cache_
Initializes local cache from Django cache .
5,880
def get_contents_static ( self , block_alias , context ) : if 'request' not in context : return '' current_url = context [ 'request' ] . path try : resolver_match = resolve ( current_url ) namespace = '' if resolver_match . namespaces : namespace = resolver_match . namespaces [ 0 ] resolved_view_name = ':%s:%s' % ( nam...
Returns contents of a static block .
5,881
def get_contents_dynamic ( self , block_alias , context ) : dynamic_block = get_dynamic_blocks ( ) . get ( block_alias , [ ] ) if not dynamic_block : return '' dynamic_block = choice ( dynamic_block ) return dynamic_block ( block_alias = block_alias , block_context = context )
Returns contents of a dynamic block .
5,882
def hash_full_tree ( self , leaves ) : root_hash , hashes = self . _hash_full ( leaves , 0 , len ( leaves ) ) assert len ( hashes ) == count_bits_set ( len ( leaves ) ) assert ( self . _hash_fold ( hashes ) == root_hash if hashes else root_hash == self . hash_empty ( ) ) return root_hash
Hash a set of leaves representing a valid full tree .
5,883
def cal_model_performance ( obsl , siml ) : nse = MathClass . nashcoef ( obsl , siml ) r2 = MathClass . rsquare ( obsl , siml ) rmse = MathClass . rmse ( obsl , siml ) pbias = MathClass . pbias ( obsl , siml ) rsr = MathClass . rsr ( obsl , siml ) print ( 'NSE: %.2f, R-square: %.2f, PBIAS: %.2f%%, RMSE: %.2f, RSR: %.2f...
Calculate model performance indexes .
5,884
def load_features ( self ) : for loc in self . loci : if self . verbose : self . logger . info ( self . logname + "Loading features for " + loc ) self . all_feats . update ( { loc : self . locus_features ( loc ) } ) if self . verbose : self . logger . info ( self . logname + "Finished loading features for " + loc ) if ...
Loads all the known features from the feature service
5,885
def locus_features ( self , locus ) : features = self . api . list_features ( locus = locus ) feat_dict = { ":" . join ( [ a . locus , str ( a . rank ) , a . term , a . sequence ] ) : a . accession for a in features } return feat_dict
Returns all features associated with a locus
5,886
def tarfile_to_pif ( filename , temp_root_dir = '' , verbose = 0 ) : temp_dir = temp_root_dir + str ( uuid . uuid4 ( ) ) os . makedirs ( temp_dir ) try : tar = tarfile . open ( filename , 'r' ) tar . extractall ( path = temp_dir ) tar . close ( ) for i in os . listdir ( temp_dir ) : cur_dir = temp_dir + '/' + i if os ....
Process a tar file that contains DFT data .
5,887
def archive_to_pif ( filename , verbose = 0 ) : if tarfile . is_tarfile ( filename ) : return tarfile_to_pif ( filename , verbose ) raise Exception ( 'Cannot process file type' )
Given a archive file that contains output from a DFT calculation parse the data and return a PIF object .
5,888
def files_to_pif ( files , verbose = 0 , quality_report = True , inline = True ) : found_parser = False for possible_parser in [ PwscfParser , VaspParser ] : try : parser = possible_parser ( files ) found_parser = True break except InvalidIngesterException : pass if not found_parser : raise Exception ( 'Directory is no...
Given a directory that contains output from a DFT calculation parse the data and return a pif object
5,889
def wait_for_confirmation ( provider , transaction_id ) : 'Sleep on a loop until we see a confirmation of the transaction.' while ( True ) : transaction = provider . gettransaction ( transaction_id ) if transaction [ "confirmations" ] > 0 : break time . sleep ( 10 )
Sleep on a loop until we see a confirmation of the transaction .
5,890
def validate_card_issue_modes ( issue_mode : int , cards : list ) -> list : supported_mask = 63 if not bool ( issue_mode & supported_mask ) : return [ ] for i in [ 1 << x for x in range ( len ( IssueMode ) ) ] : if bool ( i & issue_mode ) : try : parser_fn = cast ( Callable [ [ list ] , Optional [ list ] ] , parsers [ ...
validate cards against deck_issue modes
5,891
def p2th_address ( self ) -> Optional [ str ] : if self . id : return Kutil ( network = self . network , privkey = bytearray . fromhex ( self . id ) ) . address else : return None
P2TH address of this deck
5,892
def p2th_wif ( self ) -> Optional [ str ] : if self . id : return Kutil ( network = self . network , privkey = bytearray . fromhex ( self . id ) ) . wif else : return None
P2TH privkey in WIF format
5,893
def metainfo_to_dict ( self ) -> dict : r = { "version" : self . version , "name" : self . name , "number_of_decimals" : self . number_of_decimals , "issue_mode" : self . issue_mode } if self . asset_specific_data : r . update ( { 'asset_specific_data' : self . asset_specific_data } ) return r
encode deck into dictionary
5,894
def to_json ( self ) -> dict : d = self . __dict__ d [ 'p2th_wif' ] = self . p2th_wif return d
export the Deck object to json - ready format
5,895
def metainfo_to_dict ( self ) -> dict : r = { "version" : self . version , "amount" : self . amount , "number_of_decimals" : self . number_of_decimals } if self . asset_specific_data : r . update ( { 'asset_specific_data' : self . asset_specific_data } ) return r
encode card into dictionary
5,896
def _sort_cards ( self , cards : Generator ) -> list : return sorted ( [ card . __dict__ for card in cards ] , key = itemgetter ( 'blocknum' , 'blockseq' , 'cardseq' ) )
sort cards by blocknum and blockseq
5,897
def main ( ) : input_tif = "../tests/data/Jamaica_dem.tif" output_tif = "../tests/data/tmp_results/log_dem.tif" rst = RasterUtilClass . read_raster ( input_tif ) rst_valid = rst . validValues output_data = np . log ( rst_valid ) RasterUtilClass . write_gtiff_file ( output_tif , rst . nRows , rst . nCols , output_data ,...
Read GeoTiff raster data and perform log transformation .
5,898
def val_factory ( val , datatypes ) : exceptions = [ ] for dt in datatypes : try : if isinstance ( val , dt ) : return val return type_handler_object ( val , dt ) except Exception as e : exceptions . append ( str ( e ) ) raise ValueError ( 'val_factory: Unable to instantiate {val} from types {types}. Exceptions: {excs}...
return an instance of val that is of type datatype . keep track of exceptions so we can produce meaningful error messages .
5,899
def handler_for ( obj ) : for handler_type in handlers : if isinstance ( obj , handler_type ) : return handlers [ handler_type ] try : for handler_type in handlers : if issubclass ( obj , handler_type ) : return handlers [ handler_type ] except TypeError : pass
return the handler for the object type