idx int64 0 252k | question stringlengths 48 5.28k | target stringlengths 5 1.23k |
|---|---|---|
5,800 | def extract ( obj , pointer , bypass_ref = False ) : return Pointer ( pointer ) . extract ( obj , bypass_ref ) | Extract member or element of obj according to pointer . |
5,801 | def aa_counts ( aln , weights = None , gap_chars = '-.' ) : if weights is None : counts = Counter ( ) for rec in aln : seq_counts = Counter ( str ( rec . seq ) ) counts . update ( seq_counts ) else : if weights == True : weights = sequence_weights ( aln ) else : assert len ( weights ) == len ( aln ) , ( "Length mismatc... | Calculate the amino acid frequencies in a set of SeqRecords . |
5,802 | def aa_frequencies ( aln , weights = None , gap_chars = '-.' ) : counts = aa_counts ( aln , weights , gap_chars ) scale = 1.0 / sum ( counts . values ( ) ) return dict ( ( aa , cnt * scale ) for aa , cnt in counts . iteritems ( ) ) | Frequency of each residue type in an alignment . |
5,803 | def blocks ( aln , threshold = 0.5 , weights = None ) : assert len ( aln ) if weights == False : def pct_nongaps ( col ) : return 1 - ( float ( col . count ( '-' ) ) / len ( col ) ) else : if weights in ( None , True ) : weights = sequence_weights ( aln , 'avg1' ) def pct_nongaps ( col ) : assert len ( col ) == len ( w... | Remove gappy columns from an alignment . |
5,804 | def col_counts ( col , weights = None , gap_chars = '-.' ) : cnt = defaultdict ( float ) for aa , wt in zip ( col , weights ) : if aa not in gap_chars : cnt [ aa ] += wt return cnt | Absolute counts of each residue type in a single column . |
5,805 | def remove_empty_cols ( records ) : records = list ( records ) seqstrs = [ str ( rec . seq ) for rec in records ] clean_cols = [ col for col in zip ( * seqstrs ) if not all ( c == '-' for c in col ) ] clean_seqs = [ '' . join ( row ) for row in zip ( * clean_cols ) ] for rec , clean_seq in zip ( records , clean_seqs ) ... | Remove all - gap columns from aligned SeqRecords . |
5,806 | def sequence_weights ( aln , scaling = 'none' , gap_chars = '-.' ) : expectk = [ 0.0 , 1.0 , 1.953 , 2.861 , 3.705 , 4.524 , 5.304 , 6.026 , 6.724 , 7.397 , 8.04 , 8.622 , 9.191 , 9.739 , 10.264 , 10.758 , 11.194 , 11.635 , 12.049 , 12.468 , 12.806 , 13.185 , 13.539 , 13.863 , 14.177 , 14.466 , 14.737 , 15.005 , 15.245... | Weight aligned sequences to emphasize more divergent members . |
5,807 | def to_graph ( alnfname , weight_func ) : import networkx G = networkx . Graph ( ) aln = AlignIO . read ( alnfname , 'fasta' ) for i , arec in enumerate ( aln ) : for brec in aln [ i + 1 : ] : ident = weight_func ( str ( arec . seq ) , str ( brec . seq ) ) G . add_edge ( arec . id , brec . id , weight = ident ) return ... | Create a NetworkX graph from a sequence alignment . |
5,808 | def guidance_UV ( index ) : if 0 < index < 3 : guidance = "Low exposure. No protection required. You can safely stay outside" elif 2 < index < 6 : guidance = "Moderate exposure. Seek shade during midday hours, cover up and wear sunscreen" elif 5 < index < 8 : guidance = "High exposure. Seek shade during midday hours, c... | Return Met Office guidance regarding UV exposure based on UV index |
5,809 | def parse_sitelist ( sitelist ) : sites = [ ] for site in sitelist [ "Locations" ] [ "Location" ] : try : ident = site [ "id" ] name = site [ "name" ] except KeyError : ident = site [ "@id" ] name = site [ "@name" ] if "latitude" in site : lat = float ( site [ "latitude" ] ) lon = float ( site [ "longitude" ] ) else : ... | Return list of Site instances from retrieved sitelist data |
5,810 | def _query ( self , data_category , resource_category , field , request , step , isotime = None ) : rest_url = "/" . join ( [ HOST , data_category , resource_category , field , DATA_TYPE , request ] ) query_string = "?" + "&" . join ( [ "res=" + step , "time=" + isotime if isotime is not None else "" , "key=" + self . ... | Request and return data from DataPoint RESTful API . |
5,811 | def stand_alone_imagery ( self ) : return json . loads ( self . _query ( IMAGE , FORECAST , SURFACE_PRESSURE , CAPABILITIES , "" ) . decode ( errors = "replace" ) ) | Returns capabilities data for stand alone imagery and includes URIs for the images . |
5,812 | def map_overlay_forecast ( self ) : return json . loads ( self . _query ( LAYER , FORECAST , ALL , CAPABILITIES , "" ) . decode ( errors = "replace" ) ) | Returns capabilities data for forecast map overlays . |
5,813 | def map_overlay_obs ( self ) : return json . loads ( self . _query ( LAYER , OBSERVATIONS , ALL , CAPABILITIES , "" ) . decode ( errors = "replace" ) ) | Returns capabilities data for observation map overlays . |
5,814 | def load_and_append ( instrument_dict , instruments = None , raise_errors = False ) : if instruments is None : instruments = { } updated_instruments = { } updated_instruments . update ( instruments ) loaded_failed = { } for instrument_name , instrument_class_name in instrument_dict . items ( ) : instrument_settings = N... | load instrument from instrument_dict and append to instruments |
5,815 | def fail ( self , reason , obj , pointer = None ) : pointer = pointer_join ( pointer ) err = ValidationError ( reason , obj , pointer ) if self . fail_fast : raise err else : self . errors . append ( err ) return err | Called when validation fails . |
5,816 | def get_locus ( sequences , kir = False , verbose = False , refdata = None , evalue = 10 ) : if not refdata : refdata = ReferenceData ( ) file_id = str ( randomid ( ) ) input_fasta = file_id + ".fasta" output_xml = file_id + ".xml" SeqIO . write ( sequences , input_fasta , "fasta" ) blastn_cline = NcbiblastnCommandline... | Gets the locus of the sequence by running blastn |
5,817 | def address ( self ) -> str : return str ( self . _public_key . to_address ( net_query ( self . network ) ) ) | generate an address from pubkey |
5,818 | def sign_transaction ( self , txins : Union [ TxOut ] , tx : MutableTransaction ) -> MutableTransaction : solver = P2pkhSolver ( self . _private_key ) return tx . spend ( txins , [ solver for i in txins ] ) | sign the parent txn outputs P2PKH |
5,819 | def format_name ( net : str ) -> str : if net . startswith ( 't' ) or 'testnet' in net : net = net [ 1 : ] + '-test' else : net = net return net | take care of specifics of cryptoid naming system |
5,820 | def get_url ( url : str ) -> Union [ dict , int , float , str ] : request = Request ( url , headers = { "User-Agent" : "pypeerassets" } ) response = cast ( HTTPResponse , urlopen ( request ) ) if response . status != 200 : raise Exception ( response . reason ) return json . loads ( response . read ( ) . decode ( ) ) | Perform a GET request for the url and return a dictionary parsed from the JSON response . |
5,821 | def _scan_nodes ( nodelist , context , instance_types , current_block = None , ignore_blocks = None ) : results = [ ] for node in nodelist : if isinstance ( node , instance_types ) : results . append ( node ) elif isinstance ( node , IncludeNode ) : if node . template : if not callable ( getattr ( node . template , 're... | Loop through all nodes of a single scope level . |
5,822 | def get_node_instances ( nodelist , instances ) : context = _get_main_context ( nodelist ) if TemplateAdapter is not None and isinstance ( nodelist , TemplateAdapter ) : nodelist = nodelist . template return _scan_nodes ( nodelist , context , instances ) | Find the nodes of a given instance . |
5,823 | def get_config_file ( ) : parser = argparse . ArgumentParser ( description = "Read configuration file." ) parser . add_argument ( '-ini' , help = "Full path of configuration file" ) args = parser . parse_args ( ) ini_file = args . ini if not FileClass . is_file_exists ( ini_file ) : print ( "Usage: -ini <full path to t... | Get model configuration file name from argv |
5,824 | def isnumerical ( x ) : try : xx = float ( x ) except TypeError : return False except ValueError : return False except Exception : return False else : return True | Check the input x is numerical or not . |
5,825 | def rsquare ( obsvalues , simvalues ) : if len ( obsvalues ) != len ( simvalues ) : raise ValueError ( "The size of observed and simulated values must be " "the same for R-square calculation!" ) if not isinstance ( obsvalues , numpy . ndarray ) : obsvalues = numpy . array ( obsvalues ) if not isinstance ( simvalues , n... | Calculate Coefficient of determination . |
5,826 | def rmse ( obsvalues , simvalues ) : if len ( obsvalues ) != len ( simvalues ) : raise ValueError ( "The size of observed and simulated values must be " "the same for R-square calculation!" ) if not isinstance ( obsvalues , numpy . ndarray ) : obsvalues = numpy . array ( obsvalues ) if not isinstance ( simvalues , nump... | Calculate RMSE . |
5,827 | def pbias ( obsvalues , simvalues ) : if len ( obsvalues ) != len ( simvalues ) : raise ValueError ( "The size of observed and simulated values must be" " the same for PBIAS calculation!" ) return sum ( map ( lambda x , y : ( x - y ) * 100 , obsvalues , simvalues ) ) / sum ( obsvalues ) | Calculate PBIAS or percent model bias . |
5,828 | def convert_str2num ( unicode_str ) : if MathClass . isnumerical ( unicode_str ) : unicode_str = float ( unicode_str ) if unicode_str % 1. == 0. : unicode_str = int ( unicode_str ) return unicode_str elif is_string ( unicode_str ) : return str ( unicode_str ) elif isinstance ( unicode_str , tuple ) : return tuple ( Str... | Convert string to string integer or float . Support tuple or list . |
5,829 | def string_in_list ( tmp_str , strlist ) : new_str_list = strlist [ : ] for i , str_in_list in enumerate ( new_str_list ) : new_str_list [ i ] = str_in_list . lower ( ) return tmp_str . lower ( ) in new_str_list | Is tmp_str in strlist case insensitive . |
5,830 | def is_file_exists ( filename ) : if filename is None or not os . path . exists ( filename ) or not os . path . isfile ( filename ) : return False else : return True | Check the existence of file path . |
5,831 | def is_dir_exists ( dirpath ) : if dirpath is None or not os . path . exists ( dirpath ) or not os . path . isdir ( dirpath ) : return False else : return True | Check the existence of folder path . |
5,832 | def copy_files ( filename , dstfilename ) : FileClass . remove_files ( dstfilename ) dst_prefix = os . path . splitext ( dstfilename ) [ 0 ] pattern = os . path . splitext ( filename ) [ 0 ] + '.*' for f in glob . iglob ( pattern ) : ext = os . path . splitext ( f ) [ 1 ] dst = dst_prefix + ext copy ( f , dst ) | Copy files with the same name and different suffixes such as ESRI Shapefile . |
5,833 | def remove_files ( filename ) : pattern = os . path . splitext ( filename ) [ 0 ] + '.*' for f in glob . iglob ( pattern ) : os . remove ( f ) | Delete all files with same root as fileName i . e . regardless of suffix such as ESRI shapefile |
5,834 | def is_up_to_date ( outfile , basedatetime ) : if os . path . exists ( outfile ) : if os . path . getmtime ( outfile ) >= basedatetime : return True return False | Return true if outfile exists and is no older than base datetime . |
5,835 | def get_executable_fullpath ( name , dirname = None ) : if name is None : return None if is_string ( name ) : name = str ( name ) else : raise RuntimeError ( 'The input function name or path must be string!' ) if dirname is not None : dirname = os . path . abspath ( dirname ) fpth = dirname + os . sep + name if os . pa... | get the full path of a given executable name |
5,836 | def get_file_fullpath ( name , dirname = None ) : if name is None : return None if is_string ( name ) : name = str ( name ) else : raise RuntimeError ( 'The input function name or path must be string!' ) for sep in [ '\\' , '/' , os . sep ] : if sep in name : name = os . path . abspath ( name ) return name if dirname i... | Return full path if available . |
5,837 | def get_filename_by_suffixes ( dir_src , suffixes ) : list_files = os . listdir ( dir_src ) re_files = list ( ) if is_string ( suffixes ) : suffixes = [ suffixes ] if not isinstance ( suffixes , list ) : return None for i , suf in enumerate ( suffixes ) : if len ( suf ) >= 1 and suf [ 0 ] != '.' : suffixes [ i ] = '.' ... | get file names with the given suffixes in the given directory |
5,838 | def get_full_filename_by_suffixes ( dir_src , suffixes ) : file_names = FileClass . get_filename_by_suffixes ( dir_src , suffixes ) if file_names is None : return None return list ( dir_src + os . sep + name for name in file_names ) | get full file names with the given suffixes in the given directory |
5,839 | def get_core_name_without_suffix ( file_path ) : if '\\' in file_path : file_path = file_path . replace ( '\\' , '/' ) file_name = os . path . basename ( file_path ) core_names = file_name . split ( '.' ) if len ( core_names ) > 1 : core_names = core_names [ : - 1 ] if isinstance ( core_names , list ) : return str ( '.... | Return core file name without suffix . |
5,840 | def add_postfix ( file_path , postfix ) : cur_sep = '' for sep in [ '\\' , '/' , os . sep ] : if sep in file_path : cur_sep = sep break corename = FileClass . get_core_name_without_suffix ( file_path ) tmpspliter = os . path . basename ( file_path ) . split ( '.' ) suffix = '' if len ( tmpspliter ) > 1 : suffix = tmpsp... | Add postfix for a full file path . |
5,841 | def day_of_year ( dt ) : sec = time . mktime ( dt . timetuple ( ) ) t = time . localtime ( sec ) return t . tm_yday | Day index of year from 1 to 365 or 366 |
5,842 | def run_command ( commands ) : use_shell = False subprocess_flags = 0 startupinfo = None if sysstr == 'Windows' : if isinstance ( commands , list ) : commands = ' ' . join ( str ( c ) for c in commands ) import ctypes SEM_NOGPFAULTERRORBOX = 0x0002 ctypes . windll . kernel32 . SetErrorMode ( SEM_NOGPFAULTERRORBOX ) sub... | Execute external command and return the output lines list . In windows refers to handling - subprocess - crash - in - windows _ . |
5,843 | def current_path ( local_function ) : from inspect import getsourcefile fpath = getsourcefile ( local_function ) if fpath is None : return None return os . path . dirname ( os . path . abspath ( fpath ) ) | Get current path refers to how - do - i - get - the - path - of - the - current - executed - file - in - python _ |
5,844 | def mkdir ( dir_path ) : if not os . path . isdir ( dir_path ) or not os . path . exists ( dir_path ) : os . makedirs ( dir_path ) | Make directory if not existed |
5,845 | def rmmkdir ( dir_path ) : if not os . path . isdir ( dir_path ) or not os . path . exists ( dir_path ) : os . makedirs ( dir_path ) else : rmtree ( dir_path , True ) os . makedirs ( dir_path ) | If directory existed then remove and make ; else make it . |
5,846 | def print_msg ( contentlist ) : if isinstance ( contentlist , list ) or isinstance ( contentlist , tuple ) : return '\n' . join ( contentlist ) else : if len ( contentlist ) > 1 and contentlist [ - 1 ] != '\n' : contentlist += '\n' return contentlist | concatenate message list as single string with line feed . |
5,847 | def decode_strs_in_dict ( unicode_dict ) : unicode_dict = { StringClass . convert_str2num ( k ) : StringClass . convert_str2num ( v ) for k , v in iteritems ( unicode_dict ) } for k , v in iteritems ( unicode_dict ) : if isinstance ( v , dict ) : unicode_dict [ k ] = UtilClass . decode_strs_in_dict ( v ) return unicode... | Decode strings in dictionary which may contains unicode strings or numeric values . |
5,848 | def undo ( self ) : self . undo_manager . undo ( ) self . notify_observers ( ) logging . debug ( 'undo_manager undo stack={}' . format ( self . undo_manager . _undo_stack ) ) | Rewind the game to the previous state . |
5,849 | def redo ( self ) : self . undo_manager . redo ( ) self . notify_observers ( ) logging . debug ( 'undo_manager redo stack={}' . format ( self . undo_manager . _redo_stack ) ) | Redo the latest undone command . |
5,850 | def net_query ( name : str ) -> Constants : for net_params in networks : if name in ( net_params . name , net_params . shortname , ) : return net_params raise UnsupportedNetwork | Find the NetworkParams for a network by its long or short name . Raises UnsupportedNetwork if no NetworkParams is found . |
5,851 | def get_port_at ( self , tile_id , direction ) : for port in self . ports : if port . tile_id == tile_id and port . direction == direction : return port port = Port ( tile_id , direction , PortType . none ) self . ports . append ( port ) return port | If no port is found a new none port is made and added to self . ports . |
5,852 | def rotate_ports ( self ) : for port in self . ports : port . tile_id = ( ( port . tile_id + 1 ) % len ( hexgrid . coastal_tile_ids ( ) ) ) + 1 port . direction = hexgrid . rotate_direction ( hexgrid . EDGE , port . direction , ccw = True ) self . notify_observers ( ) | Rotates the ports 90 degrees . Useful when using the default port setup but the spectator is watching at a rotated angle from true north . |
5,853 | def intersect_keys ( keys , reffile , cache = False , clean_accs = False ) : index = None if cache : refcache = reffile + '.sqlite' if os . path . exists ( refcache ) : if os . stat ( refcache ) . st_mtime < os . stat ( reffile ) . st_mtime : logging . warn ( "Outdated cache; rebuilding index" ) else : try : index = ( ... | Extract SeqRecords from the index by matching keys . |
5,854 | def aa_frequencies ( seq , gap_chars = '-.' ) : aa_counts = Counter ( seq ) for gap_char in gap_chars : if gap_char in aa_counts : del aa_counts [ gap_char ] scale = 1.0 / sum ( aa_counts . values ( ) ) return dict ( ( aa , cnt * scale ) for aa , cnt in aa_counts . iteritems ( ) ) | Calculate the amino acid frequencies in a sequence set . |
5,855 | def giving ( self ) : logging . debug ( 'give={}' . format ( self . _give ) ) c = Counter ( self . _give . copy ( ) ) return [ ( n , t ) for t , n in c . items ( ) ] | Returns tuples corresponding to the number and type of each resource in the trade from giver - > getter |
5,856 | def getting ( self ) : c = Counter ( self . _get . copy ( ) ) return [ ( n , t ) for t , n in c . items ( ) ] | Returns tuples corresponding to the number and type of each resource in the trade from getter - > giver |
5,857 | def family_check ( self ) : self . logger . info ( "Checking family relations for {0}" . format ( self . family_id ) ) for individual_id in self . individuals : self . logger . debug ( "Checking individual {0}" . format ( individual_id ) ) individual = self . individuals [ individual_id ] self . logger . debug ( "Check... | Check if the family members break the structure of the family . eg . nonexistent parent wrong sex on parent etc . Also extracts all trios found this is of help for many at the moment since GATK can only do phasing of trios and duos . |
5,858 | def check_parent ( self , parent_id , father = False ) : self . logger . debug ( "Checking parent {0}" . format ( parent_id ) ) if parent_id != '0' : if parent_id not in self . individuals : raise PedigreeError ( self . family_id , parent_id , 'Parent is not in family.' ) if father : if self . individuals [ parent_id ]... | Check if the parent info is correct . If an individual is not present in file raise exeption . |
5,859 | def to_ped ( self , outfile = None ) : ped_header = [ '#FamilyID' , 'IndividualID' , 'PaternalID' , 'MaternalID' , 'Sex' , 'Phenotype' , ] extra_headers = [ 'InheritanceModel' , 'Proband' , 'Consultand' , 'Alive' ] for individual_id in self . individuals : individual = self . individuals [ individual_id ] for info in i... | Print the individuals of the family in ped format The header will be the original ped header plus all headers found in extra info of the individuals |
5,860 | def find_deck ( provider : Provider , key : str , version : int , prod : bool = True ) -> Optional [ Deck ] : pa_params = param_query ( provider . network ) if prod : p2th = pa_params . P2TH_addr else : p2th = pa_params . test_P2TH_addr rawtx = provider . getrawtransaction ( key , 1 ) deck = deck_parser ( ( provider , ... | Find specific deck by deck id . |
5,861 | def deck_spawn ( provider : Provider , deck : Deck , inputs : dict , change_address : str , locktime : int = 0 ) -> Transaction : network_params = net_query ( deck . network ) pa_params = param_query ( deck . network ) if deck . production : p2th_addr = pa_params . P2TH_addr else : p2th_addr = pa_params . test_P2TH_add... | Creates Deck spawn raw transaction . |
5,862 | def get_card_transfer ( provider : Provider , deck : Deck , txid : str , debug : bool = False ) -> Iterator : rawtx = provider . getrawtransaction ( txid , 1 ) bundle = card_bundler ( provider , deck , rawtx ) return card_bundle_parser ( bundle , debug ) | get a single card transfer by it s id |
5,863 | def find_all_valid_cards ( provider : Provider , deck : Deck ) -> Generator : unfiltered = ( card for batch in get_card_bundles ( provider , deck ) for card in batch ) for card in validate_card_issue_modes ( deck . issue_mode , list ( unfiltered ) ) : yield card | find all the valid cards on this deck filtering out cards which don t play nice with deck issue mode |
5,864 | def card_transfer ( provider : Provider , card : CardTransfer , inputs : dict , change_address : str , locktime : int = 0 ) -> Transaction : network_params = net_query ( provider . network ) pa_params = param_query ( provider . network ) if card . deck_p2th is None : raise Exception ( "card.deck_p2th required for tx_ou... | Prepare the CardTransfer Transaction object |
5,865 | def rfc3339_to_datetime ( data ) : try : ts = time . strptime ( data , '%Y-%m-%d' ) return date ( * ts [ : 3 ] ) except ValueError : pass try : dt , _ , tz = data . partition ( 'Z' ) if tz : tz = offset ( tz ) else : tz = offset ( '00:00' ) if '.' in dt and dt . rsplit ( '.' , 1 ) [ - 1 ] . isdigit ( ) : ts = time . st... | convert a rfc3339 date representation into a Python datetime |
5,866 | def log_config ( verbose = 1 ) : if verbose == 0 : level = logging . WARNING fmt = "%(module)s: %(message)s" elif verbose == 1 : level = logging . INFO fmt = "%(module)s [@%(lineno)s]: %(message)s" else : level = logging . DEBUG fmt = "%(module)s [%(lineno)s]: %(levelname)s: %(message)s" logging . basicConfig ( format ... | Set up logging the way I like it . |
5,867 | def refresh_instruments ( self ) : def list_access_nested_dict ( dict , somelist ) : return reduce ( operator . getitem , somelist , dict ) def update ( item ) : if item . isExpanded ( ) : for index in range ( item . childCount ( ) ) : child = item . child ( index ) if child . childCount ( ) == 0 : instrument , path_to... | if self . tree_settings has been expanded ask instruments for their actual values |
5,868 | def update_parameters ( self , treeWidget ) : if treeWidget == self . tree_settings : item = treeWidget . currentItem ( ) instrument , path_to_instrument = item . get_instrument ( ) dictator = item . value for element in path_to_instrument : dictator = { element : dictator } old_value = instrument . settings path_to_in... | updates the internal dictionaries for scripts and instruments with values from the respective trees |
5,869 | def script_finished ( self ) : script = self . current_script script . updateProgress . disconnect ( self . update_status ) self . script_thread . started . disconnect ( ) script . finished . disconnect ( ) self . current_script = None self . plot_script ( script ) self . progressBar . setValue ( 100 ) self . btn_start... | waits for the script to emit the script_finshed signal |
5,870 | def update_probes ( self , progress ) : new_values = self . read_probes . probes_values probe_count = len ( self . read_probes . probes ) if probe_count > self . tree_probes . topLevelItemCount ( ) : self . fill_treewidget ( self . tree_probes , new_values ) else : for x in range ( probe_count ) : topLvlItem = self . t... | update the probe tree |
5,871 | def update_script_from_item ( self , item ) : script , path_to_script , script_item = item . get_script ( ) dictator = list ( script_item . to_dict ( ) . values ( ) ) [ 0 ] for instrument in list ( script . instruments . keys ( ) ) : script . instruments [ instrument ] [ 'settings' ] = dictator [ instrument ] [ 'settin... | updates the script based on the information provided in item |
5,872 | def message_search ( self , text , on_success , peer = None , min_date = None , max_date = None , max_id = None , offset = 0 , limit = 255 ) : raise TWXUnsupportedMethod ( ) | Unsupported in the Bot API |
5,873 | def remove ( self , pointer ) : doc = deepcopy ( self . document ) parent , obj = None , doc try : for token in Pointer ( pointer ) : parent , obj = obj , token . extract ( obj , bypass_ref = True ) if isinstance ( parent , Mapping ) : del parent [ token ] if isinstance ( parent , MutableSequence ) : parent . pop ( int... | Remove element from sequence member from mapping . |
5,874 | def _netname ( name : str ) -> dict : try : long = net_query ( name ) . name short = net_query ( name ) . shortname except AttributeError : raise UnsupportedNetwork ( ) return { 'long' : long , 'short' : short } | resolute network name required because some providers use shortnames and other use longnames . |
5,875 | def sendrawtransaction ( cls , rawtxn : str ) -> str : if cls . is_testnet : url = 'https://testnet-explorer.peercoin.net/api/sendrawtransaction?hex={0}' . format ( rawtxn ) else : url = 'https://explorer.peercoin.net/api/sendrawtransaction?hex={0}' . format ( rawtxn ) resp = urllib . request . urlopen ( url ) return r... | sendrawtransaction remote API |
5,876 | def validateaddress ( self , address : str ) -> bool : "Returns True if the passed address is valid, False otherwise." try : Address . from_string ( address , self . network_properties ) except InvalidAddress : return False return True | Returns True if the passed address is valid False otherwise . |
5,877 | def chunker ( l , n ) : for i in ranger ( 0 , len ( l ) , n ) : yield l [ i : i + n ] | Generates n - sized chunks from the list l |
5,878 | def post ( self , endpoint , data , parallelism = 5 ) : headers = { "Content-Type" : "application/json" , "Accept" : "application/json" , "x-standardize-only" : "true" if self . standardize else "false" , "x-include-invalid" : "true" if self . invalid else "false" , "x-accept-keypair" : "true" if self . accept_keypair ... | Executes most of the request . |
5,879 | def _cache_init ( self ) : cache_ = cache . get ( self . CACHE_KEY ) if cache_ is None : cache_ = defaultdict ( dict ) self . _cache = cache_ | Initializes local cache from Django cache . |
5,880 | def get_contents_static ( self , block_alias , context ) : if 'request' not in context : return '' current_url = context [ 'request' ] . path try : resolver_match = resolve ( current_url ) namespace = '' if resolver_match . namespaces : namespace = resolver_match . namespaces [ 0 ] resolved_view_name = ':%s:%s' % ( nam... | Returns contents of a static block . |
5,881 | def get_contents_dynamic ( self , block_alias , context ) : dynamic_block = get_dynamic_blocks ( ) . get ( block_alias , [ ] ) if not dynamic_block : return '' dynamic_block = choice ( dynamic_block ) return dynamic_block ( block_alias = block_alias , block_context = context ) | Returns contents of a dynamic block . |
5,882 | def hash_full_tree ( self , leaves ) : root_hash , hashes = self . _hash_full ( leaves , 0 , len ( leaves ) ) assert len ( hashes ) == count_bits_set ( len ( leaves ) ) assert ( self . _hash_fold ( hashes ) == root_hash if hashes else root_hash == self . hash_empty ( ) ) return root_hash | Hash a set of leaves representing a valid full tree . |
5,883 | def cal_model_performance ( obsl , siml ) : nse = MathClass . nashcoef ( obsl , siml ) r2 = MathClass . rsquare ( obsl , siml ) rmse = MathClass . rmse ( obsl , siml ) pbias = MathClass . pbias ( obsl , siml ) rsr = MathClass . rsr ( obsl , siml ) print ( 'NSE: %.2f, R-square: %.2f, PBIAS: %.2f%%, RMSE: %.2f, RSR: %.2f... | Calculate model performance indexes . |
5,884 | def load_features ( self ) : for loc in self . loci : if self . verbose : self . logger . info ( self . logname + "Loading features for " + loc ) self . all_feats . update ( { loc : self . locus_features ( loc ) } ) if self . verbose : self . logger . info ( self . logname + "Finished loading features for " + loc ) if ... | Loads all the known features from the feature service |
5,885 | def locus_features ( self , locus ) : features = self . api . list_features ( locus = locus ) feat_dict = { ":" . join ( [ a . locus , str ( a . rank ) , a . term , a . sequence ] ) : a . accession for a in features } return feat_dict | Returns all features associated with a locus |
5,886 | def tarfile_to_pif ( filename , temp_root_dir = '' , verbose = 0 ) : temp_dir = temp_root_dir + str ( uuid . uuid4 ( ) ) os . makedirs ( temp_dir ) try : tar = tarfile . open ( filename , 'r' ) tar . extractall ( path = temp_dir ) tar . close ( ) for i in os . listdir ( temp_dir ) : cur_dir = temp_dir + '/' + i if os .... | Process a tar file that contains DFT data . |
5,887 | def archive_to_pif ( filename , verbose = 0 ) : if tarfile . is_tarfile ( filename ) : return tarfile_to_pif ( filename , verbose ) raise Exception ( 'Cannot process file type' ) | Given a archive file that contains output from a DFT calculation parse the data and return a PIF object . |
5,888 | def files_to_pif ( files , verbose = 0 , quality_report = True , inline = True ) : found_parser = False for possible_parser in [ PwscfParser , VaspParser ] : try : parser = possible_parser ( files ) found_parser = True break except InvalidIngesterException : pass if not found_parser : raise Exception ( 'Directory is no... | Given a directory that contains output from a DFT calculation parse the data and return a pif object |
5,889 | def wait_for_confirmation ( provider , transaction_id ) : 'Sleep on a loop until we see a confirmation of the transaction.' while ( True ) : transaction = provider . gettransaction ( transaction_id ) if transaction [ "confirmations" ] > 0 : break time . sleep ( 10 ) | Sleep on a loop until we see a confirmation of the transaction . |
5,890 | def validate_card_issue_modes ( issue_mode : int , cards : list ) -> list : supported_mask = 63 if not bool ( issue_mode & supported_mask ) : return [ ] for i in [ 1 << x for x in range ( len ( IssueMode ) ) ] : if bool ( i & issue_mode ) : try : parser_fn = cast ( Callable [ [ list ] , Optional [ list ] ] , parsers [ ... | validate cards against deck_issue modes |
5,891 | def p2th_address ( self ) -> Optional [ str ] : if self . id : return Kutil ( network = self . network , privkey = bytearray . fromhex ( self . id ) ) . address else : return None | P2TH address of this deck |
5,892 | def p2th_wif ( self ) -> Optional [ str ] : if self . id : return Kutil ( network = self . network , privkey = bytearray . fromhex ( self . id ) ) . wif else : return None | P2TH privkey in WIF format |
5,893 | def metainfo_to_dict ( self ) -> dict : r = { "version" : self . version , "name" : self . name , "number_of_decimals" : self . number_of_decimals , "issue_mode" : self . issue_mode } if self . asset_specific_data : r . update ( { 'asset_specific_data' : self . asset_specific_data } ) return r | encode deck into dictionary |
5,894 | def to_json ( self ) -> dict : d = self . __dict__ d [ 'p2th_wif' ] = self . p2th_wif return d | export the Deck object to json - ready format |
5,895 | def metainfo_to_dict ( self ) -> dict : r = { "version" : self . version , "amount" : self . amount , "number_of_decimals" : self . number_of_decimals } if self . asset_specific_data : r . update ( { 'asset_specific_data' : self . asset_specific_data } ) return r | encode card into dictionary |
5,896 | def _sort_cards ( self , cards : Generator ) -> list : return sorted ( [ card . __dict__ for card in cards ] , key = itemgetter ( 'blocknum' , 'blockseq' , 'cardseq' ) ) | sort cards by blocknum and blockseq |
5,897 | def main ( ) : input_tif = "../tests/data/Jamaica_dem.tif" output_tif = "../tests/data/tmp_results/log_dem.tif" rst = RasterUtilClass . read_raster ( input_tif ) rst_valid = rst . validValues output_data = np . log ( rst_valid ) RasterUtilClass . write_gtiff_file ( output_tif , rst . nRows , rst . nCols , output_data ,... | Read GeoTiff raster data and perform log transformation . |
5,898 | def val_factory ( val , datatypes ) : exceptions = [ ] for dt in datatypes : try : if isinstance ( val , dt ) : return val return type_handler_object ( val , dt ) except Exception as e : exceptions . append ( str ( e ) ) raise ValueError ( 'val_factory: Unable to instantiate {val} from types {types}. Exceptions: {excs}... | return an instance of val that is of type datatype . keep track of exceptions so we can produce meaningful error messages . |
5,899 | def handler_for ( obj ) : for handler_type in handlers : if isinstance ( obj , handler_type ) : return handlers [ handler_type ] try : for handler_type in handlers : if issubclass ( obj , handler_type ) : return handlers [ handler_type ] except TypeError : pass | return the handler for the object type |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.