idx
int64
0
252k
question
stringlengths
48
5.28k
target
stringlengths
5
1.23k
251,600
def internal_link_sets ( self ) : if not self . __internal_link_sets : self . __internal_link_sets = InternalLinkSets ( self . __connection ) return self . __internal_link_sets
Gets the InternalLinkSets API client .
251,601
def logical_interconnect_groups ( self ) : if not self . __logical_interconnect_groups : self . __logical_interconnect_groups = LogicalInterconnectGroups ( self . __connection ) return self . __logical_interconnect_groups
Gets the LogicalInterconnectGroups API client .
251,602
def sas_logical_interconnects ( self ) : if not self . __sas_logical_interconnects : self . __sas_logical_interconnects = SasLogicalInterconnects ( self . __connection ) return self . __sas_logical_interconnects
Gets the SasLogicalInterconnects API client .
251,603
def logical_downlinks ( self ) : if not self . __logical_downlinks : self . __logical_downlinks = LogicalDownlinks ( self . __connection ) return self . __logical_downlinks
Gets the LogicalDownlinks API client .
251,604
def power_devices ( self ) : if not self . __power_devices : self . __power_devices = PowerDevices ( self . __connection ) return self . __power_devices
Gets the PowerDevices API client .
251,605
def unmanaged_devices ( self ) : if not self . __unmanaged_devices : self . __unmanaged_devices = UnmanagedDevices ( self . __connection ) return self . __unmanaged_devices
Gets the Unmanaged Devices API client .
251,606
def racks ( self ) : if not self . __racks : self . __racks = Racks ( self . __connection ) return self . __racks
Gets the Racks API client .
251,607
def san_managers ( self ) : if not self . __san_managers : self . __san_managers = SanManagers ( self . __connection ) return self . __san_managers
Gets the SanManagers API client .
251,608
def endpoints ( self ) : if not self . __endpoints : self . __endpoints = Endpoints ( self . __connection ) return self . __endpoints
Gets the Endpoints API client .
251,609
def storage_systems ( self ) : if not self . __storage_systems : self . __storage_systems = StorageSystems ( self . __connection ) return self . __storage_systems
Gets the StorageSystems API client .
251,610
def storage_pools ( self ) : if not self . __storage_pools : self . __storage_pools = StoragePools ( self . __connection ) return self . __storage_pools
Gets the StoragePools API client .
251,611
def storage_volume_templates ( self ) : if not self . __storage_volume_templates : self . __storage_volume_templates = StorageVolumeTemplates ( self . __connection ) return self . __storage_volume_templates
Gets the StorageVolumeTemplates API client .
251,612
def storage_volume_attachments ( self ) : if not self . __storage_volume_attachments : self . __storage_volume_attachments = StorageVolumeAttachments ( self . __connection ) return self . __storage_volume_attachments
Gets the StorageVolumeAttachments API client .
251,613
def firmware_drivers ( self ) : if not self . __firmware_drivers : self . __firmware_drivers = FirmwareDrivers ( self . __connection ) return self . __firmware_drivers
Gets the FirmwareDrivers API client .
251,614
def firmware_bundles ( self ) : if not self . __firmware_bundles : self . __firmware_bundles = FirmwareBundles ( self . __connection ) return self . __firmware_bundles
Gets the FirmwareBundles API client .
251,615
def volumes ( self ) : if not self . __volumes : self . __volumes = Volumes ( self . __connection ) return self . __volumes
Gets the Volumes API client .
251,616
def sas_logical_jbod_attachments ( self ) : if not self . __sas_logical_jbod_attachments : self . __sas_logical_jbod_attachments = SasLogicalJbodAttachments ( self . __connection ) return self . __sas_logical_jbod_attachments
Gets the SAS Logical JBOD Attachments client .
251,617
def managed_sans ( self ) : if not self . __managed_sans : self . __managed_sans = ManagedSANs ( self . __connection ) return self . __managed_sans
Gets the Managed SANs API client .
251,618
def migratable_vc_domains ( self ) : if not self . __migratable_vc_domains : self . __migratable_vc_domains = MigratableVcDomains ( self . __connection ) return self . __migratable_vc_domains
Gets the VC Migration Manager API client .
251,619
def sas_interconnects ( self ) : if not self . __sas_interconnects : self . __sas_interconnects = SasInterconnects ( self . __connection ) return self . __sas_interconnects
Gets the SAS Interconnects API client .
251,620
def sas_logical_interconnect_groups ( self ) : if not self . __sas_logical_interconnect_groups : self . __sas_logical_interconnect_groups = SasLogicalInterconnectGroups ( self . __connection ) return self . __sas_logical_interconnect_groups
Gets the SasLogicalInterconnectGroups API client .
251,621
def drive_enclosures ( self ) : if not self . __drive_enclures : self . __drive_enclures = DriveEnclosures ( self . __connection ) return self . __drive_enclures
Gets the Drive Enclosures API client .
251,622
def sas_logical_jbods ( self ) : if not self . __sas_logical_jbods : self . __sas_logical_jbods = SasLogicalJbods ( self . __connection ) return self . __sas_logical_jbods
Gets the SAS Logical JBODs API client .
251,623
def labels ( self ) : if not self . __labels : self . __labels = Labels ( self . __connection ) return self . __labels
Gets the Labels API client .
251,624
def index_resources ( self ) : if not self . __index_resources : self . __index_resources = IndexResources ( self . __connection ) return self . __index_resources
Gets the Index Resources API client .
251,625
def alerts ( self ) : if not self . __alerts : self . __alerts = Alerts ( self . __connection ) return self . __alerts
Gets the Alerts API client .
251,626
def events ( self ) : if not self . __events : self . __events = Events ( self . __connection ) return self . __events
Gets the Events API client .
251,627
def os_deployment_servers ( self ) : if not self . __os_deployment_servers : self . __os_deployment_servers = OsDeploymentServers ( self . __connection ) return self . __os_deployment_servers
Gets the Os Deployment Servers API client .
251,628
def certificate_rabbitmq ( self ) : if not self . __certificate_rabbitmq : self . __certificate_rabbitmq = CertificateRabbitMQ ( self . __connection ) return self . __certificate_rabbitmq
Gets the Certificate RabbitMQ API client .
251,629
def users ( self ) : if not self . __users : self . __users = Users ( self . __connection ) return self . __users
Gets the Users API client .
251,630
def appliance_device_read_community ( self ) : if not self . __appliance_device_read_community : self . __appliance_device_read_community = ApplianceDeviceReadCommunity ( self . __connection ) return self . __appliance_device_read_community
Gets the ApplianceDeviceReadCommunity API client .
251,631
def appliance_device_snmp_v1_trap_destinations ( self ) : if not self . __appliance_device_snmp_v1_trap_destinations : self . __appliance_device_snmp_v1_trap_destinations = ApplianceDeviceSNMPv1TrapDestinations ( self . __connection ) return self . __appliance_device_snmp_v1_trap_destinations
Gets the ApplianceDeviceSNMPv1TrapDestinations API client .
251,632
def appliance_device_snmp_v3_trap_destinations ( self ) : if not self . __appliance_device_snmp_v3_trap_destinations : self . __appliance_device_snmp_v3_trap_destinations = ApplianceDeviceSNMPv3TrapDestinations ( self . __connection ) return self . __appliance_device_snmp_v3_trap_destinations
Gets the ApplianceDeviceSNMPv3TrapDestinations API client .
251,633
def appliance_device_snmp_v3_users ( self ) : if not self . __appliance_device_snmp_v3_users : self . __appliance_device_snmp_v3_users = ApplianceDeviceSNMPv3Users ( self . __connection ) return self . __appliance_device_snmp_v3_users
Gets the ApplianceDeviceSNMPv3Users API client .
251,634
def appliance_node_information ( self ) : if not self . __appliance_node_information : self . __appliance_node_information = ApplianceNodeInformation ( self . __connection ) return self . __appliance_node_information
Gets the ApplianceNodeInformation API client .
251,635
def appliance_time_and_locale_configuration ( self ) : if not self . __appliance_time_and_locale_configuration : self . __appliance_time_and_locale_configuration = ApplianceTimeAndLocaleConfiguration ( self . __connection ) return self . __appliance_time_and_locale_configuration
Gets the ApplianceTimeAndLocaleConfiguration API client .
251,636
def versions ( self ) : if not self . __versions : self . __versions = Versions ( self . __connection ) return self . __versions
Gets the Version API client .
251,637
def backups ( self ) : if not self . __backups : self . __backups = Backups ( self . __connection ) return self . __backups
Gets the Backup API client .
251,638
def login_details ( self ) : if not self . __login_details : self . __login_details = LoginDetails ( self . __connection ) return self . __login_details
Gets the login details
251,639
def get_available_networks ( self , ** kwargs ) : query_string = '&' . join ( '{}={}' . format ( key , value ) for key , value in kwargs . items ( ) if value ) uri = self . URI + "{}?{}" . format ( "/available-networks" , query_string ) return self . _helper . do_get ( uri )
Retrieves the list of Ethernet networks Fibre Channel networks and network sets that are available to a server profile template along with their respective ports . The scopeUris serverHardwareTypeUri and enclosureGroupUri parameters should be specified to get the available networks for a new server profile template . The serverHardwareTypeUri enclosureGroupUri and profileTemplateUri should be specified to get available networks for an existing server profile template . The scopeUris parameter is ignored when the profileTemplateUri is specified .
251,640
def get_all_without_ethernet ( self , start = 0 , count = - 1 , filter = '' , sort = '' ) : without_ethernet_client = ResourceClient ( self . _connection , "/rest/logical-downlinks/withoutEthernet" ) return without_ethernet_client . get_all ( start , count , filter = filter , sort = sort )
Gets a paginated collection of logical downlinks without ethernet . The collection is based on optional sorting and filtering and is constrained by start and count parameters .
251,641
def get_without_ethernet ( self , id_or_uri ) : uri = self . _client . build_uri ( id_or_uri ) + "/withoutEthernet" return self . _client . get ( uri )
Gets the logical downlink with the specified ID without ethernet .
251,642
def update_firmware ( self , firmware_information , force = False ) : firmware_uri = "{}/firmware" . format ( self . data [ "uri" ] ) result = self . _helper . update ( firmware_information , firmware_uri , force = force ) self . refresh ( ) return result
Installs firmware to the member interconnects of a SAS Logical Interconnect .
251,643
def get_firmware ( self ) : firmware_uri = "{}/firmware" . format ( self . data [ "uri" ] ) return self . _helper . do_get ( firmware_uri )
Gets baseline firmware information for a SAS Logical Interconnect .
251,644
def update_compliance_all ( self , information , timeout = - 1 ) : uri = self . URI + "/compliance" result = self . _helper . update ( information , uri , timeout = timeout ) return result
Returns SAS Logical Interconnects to a consistent state . The current SAS Logical Interconnect state is compared to the associated SAS Logical Interconnect group .
251,645
def replace_drive_enclosure ( self , information ) : uri = "{}/replaceDriveEnclosure" . format ( self . data [ "uri" ] ) result = self . _helper . create ( information , uri ) self . refresh ( ) return result
When a drive enclosure has been physically replaced initiate the replacement operation that enables the new drive enclosure to take over as a replacement for the prior drive enclosure . The request requires specification of both the serial numbers of the original drive enclosure and its replacement to be provided .
251,646
def update_configuration ( self ) : uri = "{}/configuration" . format ( self . data [ "uri" ] ) result = self . _helper . update ( { } , uri ) self . refresh ( ) return result
Asynchronously applies or re - applies the SAS Logical Interconnect configuration to all managed interconnects of a SAS Logical Interconnect .
251,647
def mutationhash ( strings , nedit ) : maxlen = max ( [ len ( string ) for string in strings ] ) indexes = generate_idx ( maxlen , nedit ) muthash = defaultdict ( set ) for string in strings : muthash [ string ] . update ( [ string ] ) for x in substitution_set ( string , indexes ) : muthash [ x ] . update ( [ string ] ) return muthash
produce a hash with each key a nedit distance substitution for a set of strings . values of the hash is the set of strings the substitution could have come from
251,648
def substitution_set ( string , indexes ) : strlen = len ( string ) return { mutate_string ( string , x ) for x in indexes if valid_substitution ( strlen , x ) }
for a string return a set of all possible substitutions
251,649
def valid_substitution ( strlen , index ) : values = index [ 0 ] return all ( [ strlen > i for i in values ] )
skip performing substitutions that are outside the bounds of the string
251,650
def acgt_match ( string ) : search = re . compile ( r'[^ACGT]' ) . search return not bool ( search ( string ) )
returns True if sting consist of only A C G T
251,651
def stream_fastq ( file_handler ) : next_element = '' for i , line in enumerate ( file_handler ) : next_element += line if i % 4 == 3 : yield next_element next_element = ''
Generator which gives all four lines if a fastq read as one string
251,652
def read_fastq ( filename ) : if not filename : return itertools . cycle ( ( None , ) ) if filename == "-" : filename_fh = sys . stdin elif filename . endswith ( 'gz' ) : if is_python3 : filename_fh = gzip . open ( filename , mode = 'rt' ) else : filename_fh = BufferedReader ( gzip . open ( filename , mode = 'rt' ) ) else : filename_fh = open ( filename ) return stream_fastq ( filename_fh )
return a stream of FASTQ entries handling gzipped and empty files
251,653
def write_fastq ( filename ) : if filename : if filename . endswith ( 'gz' ) : filename_fh = gzip . open ( filename , mode = 'wb' ) else : filename_fh = open ( filename , mode = 'w' ) else : filename_fh = None return filename_fh
return a handle for FASTQ writing handling gzipped files
251,654
def detect_alignment_annotations ( queryalignment , tags = False ) : annotations = set ( ) for k , v in BARCODEINFO . items ( ) : if tags : if queryalignment . has_tag ( v . bamtag ) : annotations . add ( k ) else : if v . readprefix in queryalignment . qname : annotations . add ( k ) return annotations
detects the annotations present in a SAM file inspecting either the tags or the query names and returns a set of annotations present
251,655
def detect_fastq_annotations ( fastq_file ) : annotations = set ( ) queryread = tz . first ( read_fastq ( fastq_file ) ) for k , v in BARCODEINFO . items ( ) : if v . readprefix in queryread : annotations . add ( k ) return annotations
detects annotations preesent in a FASTQ file by examining the first read
251,656
def construct_transformed_regex ( annotations ) : re_string = '.*' if "cellular" in annotations : re_string += ":CELL_(?P<CB>.*)" if "molecular" in annotations : re_string += ":UMI_(?P<MB>\w*)" if "sample" in annotations : re_string += ":SAMPLE_(?P<SB>\w*)" if re_string == ".*" : logger . error ( "No annotation present on this file, aborting." ) sys . exit ( 1 ) return re_string
construct a regex that matches possible fields in a transformed file annotations is a set of which keys in BARCODEINFO are present in the file
251,657
def _infer_transform_options ( transform ) : TransformOptions = collections . namedtuple ( "TransformOptions" , [ 'CB' , 'dual_index' , 'triple_index' , 'MB' , 'SB' ] ) CB = False SB = False MB = False dual_index = False triple_index = False for rx in transform . values ( ) : if not rx : continue if "CB1" in rx : if "CB3" in rx : triple_index = True else : dual_index = True if "SB" in rx : SB = True if "CB" in rx : CB = True if "MB" in rx : MB = True return TransformOptions ( CB = CB , dual_index = dual_index , triple_index = triple_index , MB = MB , SB = SB )
figure out what transform options should be by examining the provided regexes for keywords
251,658
def _extract_readnum ( read_dict ) : pat = re . compile ( r"(?P<readnum>/\d+)$" ) parts = pat . split ( read_dict [ "name" ] ) if len ( parts ) == 3 : name , readnum , endofline = parts read_dict [ "name" ] = name read_dict [ "readnum" ] = readnum else : read_dict [ "readnum" ] = "" return read_dict
Extract read numbers from old - style fastqs .
251,659
def sparse ( csv , sparse ) : import pandas as pd df = pd . read_csv ( csv , index_col = 0 , header = 0 ) pd . Series ( df . index ) . to_csv ( sparse + ".rownames" , index = False ) pd . Series ( df . columns . values ) . to_csv ( sparse + ".colnames" , index = False ) with open ( sparse , "w+b" ) as out_handle : scipy . io . mmwrite ( out_handle , scipy . sparse . csr_matrix ( df ) )
Convert a CSV file to a sparse matrix with rows and column names saved as companion files .
251,660
def cb_histogram ( fastq , umi_histogram ) : annotations = detect_fastq_annotations ( fastq ) re_string = construct_transformed_regex ( annotations ) parser_re = re . compile ( re_string ) cb_counter = collections . Counter ( ) umi_counter = collections . Counter ( ) for read in read_fastq ( fastq ) : match = parser_re . search ( read ) . groupdict ( ) cb = match [ 'CB' ] cb_counter [ cb ] += 1 if umi_histogram : umi = match [ 'MB' ] umi_counter [ ( cb , umi ) ] += 1 for bc , count in cb_counter . most_common ( ) : sys . stdout . write ( '{}\t{}\n' . format ( bc , count ) ) if umi_histogram : with open ( umi_histogram , "w" ) as umi_handle : for cbumi , count in umi_counter . most_common ( ) : umi_handle . write ( '{}\t{}\t{}\n' . format ( cbumi [ 0 ] , cbumi [ 1 ] , count ) )
Counts the number of reads for each cellular barcode
251,661
def umi_histogram ( fastq ) : annotations = detect_fastq_annotations ( fastq ) re_string = construct_transformed_regex ( annotations ) parser_re = re . compile ( re_string ) counter = collections . Counter ( ) for read in read_fastq ( fastq ) : match = parser_re . search ( read ) . groupdict ( ) counter [ match [ 'MB' ] ] += 1 for bc , count in counter . most_common ( ) : sys . stdout . write ( '{}\t{}\n' . format ( bc , count ) )
Counts the number of reads for each UMI
251,662
def get_cb_depth_set ( cb_histogram , cb_cutoff ) : cb_keep_set = set ( ) if not cb_histogram : return cb_keep_set with read_cbhistogram ( cb_histogram ) as fh : cb_map = dict ( p . strip ( ) . split ( ) for p in fh ) cb_keep_set = set ( [ k for k , v in cb_map . items ( ) if int ( v ) > cb_cutoff ] ) logger . info ( 'Keeping %d out of %d cellular barcodes.' % ( len ( cb_keep_set ) , len ( cb_map ) ) ) return cb_keep_set
Returns a set of barcodes with a minimum number of reads
251,663
def guess_depth_cutoff ( cb_histogram ) : with read_cbhistogram ( cb_histogram ) as fh : cb_vals = [ int ( p . strip ( ) . split ( ) [ 1 ] ) for p in fh ] histo = np . histogram ( np . log10 ( cb_vals ) , bins = 50 ) vals = histo [ 0 ] edges = histo [ 1 ] mids = np . array ( [ ( edges [ i ] + edges [ i + 1 ] ) / 2 for i in range ( edges . size - 1 ) ] ) wdensity = vals * ( 10 ** mids ) / sum ( vals * ( 10 ** mids ) ) baseline = np . median ( wdensity ) wdensity = list ( wdensity ) peak = wdensity . index ( max ( wdensity [ len ( wdensity ) / 2 : ] ) ) cutoff = None for index , dens in reversed ( list ( enumerate ( wdensity [ 1 : peak ] ) ) ) : if dens < 2 * baseline : cutoff = index break if not cutoff : return None else : cutoff = 10 ** mids [ cutoff ] logger . info ( 'Setting barcode cutoff to %d' % cutoff ) return cutoff
Guesses at an appropriate barcode cutoff
251,664
def cb_filter ( fastq , bc1 , bc2 , bc3 , cores , nedit ) : with open_gzipsafe ( bc1 ) as bc1_fh : bc1 = set ( cb . strip ( ) for cb in bc1_fh ) if bc2 : with open_gzipsafe ( bc2 ) as bc2_fh : bc2 = set ( cb . strip ( ) for cb in bc2_fh ) if bc3 : with open_gzipsafe ( bc3 ) as bc3_fh : bc3 = set ( cb . strip ( ) for cb in bc3_fh ) annotations = detect_fastq_annotations ( fastq ) re_string = construct_transformed_regex ( annotations ) if nedit == 0 : filter_cb = partial ( exact_barcode_filter , bc1 = bc1 , bc2 = bc2 , bc3 = bc3 , re_string = re_string ) else : bc1hash = MutationHash ( bc1 , nedit ) bc2hash = None bc3hash = None if bc2 : bc2hash = MutationHash ( bc2 , nedit ) if bc3 : bc3hash = MutationHash ( bc3 , nedit ) filter_cb = partial ( correcting_barcode_filter , bc1hash = bc1hash , bc2hash = bc2hash , bc3hash = bc3hash , re_string = re_string ) p = multiprocessing . Pool ( cores ) chunks = tz . partition_all ( 10000 , read_fastq ( fastq ) ) bigchunks = tz . partition_all ( cores , chunks ) for bigchunk in bigchunks : for chunk in p . map ( filter_cb , list ( bigchunk ) ) : for read in chunk : sys . stdout . write ( read )
Filters reads with non - matching barcodes Expects formatted fastq files .
251,665
def sb_filter ( fastq , bc , cores , nedit ) : barcodes = set ( sb . strip ( ) for sb in bc ) if nedit == 0 : filter_sb = partial ( exact_sample_filter2 , barcodes = barcodes ) else : barcodehash = MutationHash ( barcodes , nedit ) filter_sb = partial ( correcting_sample_filter2 , barcodehash = barcodehash ) p = multiprocessing . Pool ( cores ) chunks = tz . partition_all ( 10000 , read_fastq ( fastq ) ) bigchunks = tz . partition_all ( cores , chunks ) for bigchunk in bigchunks : for chunk in p . map ( filter_sb , list ( bigchunk ) ) : for read in chunk : sys . stdout . write ( read )
Filters reads with non - matching sample barcodes Expects formatted fastq files .
251,666
def mb_filter ( fastq , cores ) : filter_mb = partial ( umi_filter ) p = multiprocessing . Pool ( cores ) chunks = tz . partition_all ( 10000 , read_fastq ( fastq ) ) bigchunks = tz . partition_all ( cores , chunks ) for bigchunk in bigchunks : for chunk in p . map ( filter_mb , list ( bigchunk ) ) : for read in chunk : sys . stdout . write ( read )
Filters umis with non - ACGT bases Expects formatted fastq files .
251,667
def kallisto ( fastq , out_dir , cb_histogram , cb_cutoff ) : parser_re = re . compile ( '(.*):CELL_(?<CB>.*):UMI_(?P<UMI>.*)\\n(.*)\\n\\+\\n(.*)\\n' ) if fastq . endswith ( 'gz' ) : fastq_fh = gzip . GzipFile ( fileobj = open ( fastq ) ) elif fastq == "-" : fastq_fh = sys . stdin else : fastq_fh = open ( fastq ) cb_depth_set = get_cb_depth_set ( cb_histogram , cb_cutoff ) cb_set = set ( ) cb_batch = collections . defaultdict ( list ) parsed = 0 for read in stream_fastq ( fastq_fh ) : match = parser_re . search ( read ) . groupdict ( ) umi = match [ 'UMI' ] cb = match [ 'CB' ] if cb_depth_set and cb not in cb_depth_set : continue parsed += 1 cb_set . add ( cb ) cb_batch [ cb ] . append ( ( read , umi ) ) if not parsed % 10000000 : for cb , chunk in cb_batch . items ( ) : write_kallisto_chunk ( out_dir , cb , chunk ) cb_batch = collections . defaultdict ( list ) for cb , chunk in cb_batch . items ( ) : write_kallisto_chunk ( out_dir , cb , chunk ) with open ( os . path . join ( out_dir , "barcodes.batch" ) , "w" ) as out_handle : out_handle . write ( "#id umi-file file-1\n" ) batchformat = "{cb} {cb}.umi {cb}.fq\n" for cb in cb_set : out_handle . write ( batchformat . format ( ** locals ( ) ) )
Convert fastqtransformed file to output format compatible with kallisto .
251,668
def demultiplex_samples ( fastq , out_dir , nedit , barcodes ) : annotations = detect_fastq_annotations ( fastq ) re_string = construct_transformed_regex ( annotations ) parser_re = re . compile ( re_string ) if barcodes : barcodes = set ( barcode . strip ( ) for barcode in barcodes ) else : barcodes = set ( ) if nedit == 0 : filter_bc = partial ( exact_sample_filter , barcodes = barcodes ) else : barcodehash = MutationHash ( barcodes , nedit ) filter_bc = partial ( correcting_sample_filter , barcodehash = barcodehash ) sample_set = set ( ) batch = collections . defaultdict ( list ) parsed = 0 safe_makedir ( out_dir ) for read in read_fastq ( fastq ) : parsed += 1 read = filter_bc ( read ) if not read : continue match = parser_re . search ( read ) . groupdict ( ) sample = match [ 'SB' ] sample_set . add ( sample ) batch [ sample ] . append ( read ) if not parsed % 10000000 : for sample , reads in batch . items ( ) : out_file = os . path . join ( out_dir , sample + ".fq" ) with open ( out_file , "a" ) as out_handle : for read in reads : fixed = filter_bc ( read ) if fixed : out_handle . write ( fixed ) batch = collections . defaultdict ( list ) for sample , reads in batch . items ( ) : out_file = os . path . join ( out_dir , sample + ".fq" ) with open ( out_file , "a" ) as out_handle : for read in reads : fixed = filter_bc ( read ) if fixed : out_handle . write ( read )
Demultiplex a fastqtransformed FASTQ file into a FASTQ file for each sample .
251,669
def demultiplex_cells ( fastq , out_dir , readnumber , prefix , cb_histogram , cb_cutoff ) : annotations = detect_fastq_annotations ( fastq ) re_string = construct_transformed_regex ( annotations ) parser_re = re . compile ( re_string ) readstring = "" if not readnumber else "_R{}" . format ( readnumber ) filestring = "{prefix}{sample}{readstring}.fq" cb_set = set ( ) if cb_histogram : cb_set = get_cb_depth_set ( cb_histogram , cb_cutoff ) sample_set = set ( ) batch = collections . defaultdict ( list ) parsed = 0 safe_makedir ( out_dir ) for read in read_fastq ( fastq ) : parsed += 1 match = parser_re . search ( read ) . groupdict ( ) sample = match [ 'CB' ] if cb_set and sample not in cb_set : continue sample_set . add ( sample ) batch [ sample ] . append ( read ) if not parsed % 10000000 : for sample , reads in batch . items ( ) : out_file = os . path . join ( out_dir , filestring . format ( ** locals ( ) ) ) with open ( out_file , "a" ) as out_handle : for read in reads : out_handle . write ( read ) batch = collections . defaultdict ( list ) for sample , reads in batch . items ( ) : out_file = os . path . join ( out_dir , filestring . format ( ** locals ( ) ) ) with open ( out_file , "a" ) as out_handle : for read in reads : out_handle . write ( read )
Demultiplex a fastqtransformed FASTQ file into a FASTQ file for each cell .
251,670
def array_type ( data_types , field ) : from sqlalchemy . dialects import postgresql internal_type = field . base_field . get_internal_type ( ) if internal_type in data_types and internal_type != 'ArrayField' : sub_type = data_types [ internal_type ] ( field ) if not isinstance ( sub_type , ( list , tuple ) ) : sub_type = [ sub_type ] else : raise RuntimeError ( 'Unsupported array element type' ) return postgresql . ARRAY ( sub_type )
Allows conversion of Django ArrayField to SQLAlchemy Array . Takes care of mapping the type of the array element .
251,671
def set_verbose_logger_handlers ( ) : global _REGISTERED_LOGGER_HANDLERS formatter = logging . Formatter ( '%(asctime)s %(levelname)s %(name)s:%(funcName)s:%(lineno)d ' '%(message)s' ) formatter . default_msec_format = '%s.%03d' for handler in _REGISTERED_LOGGER_HANDLERS : handler . setFormatter ( formatter )
Set logger handler formatters to more detail
251,672
def download ( ctx ) : settings . add_cli_options ( ctx . cli_options , settings . TransferAction . Download ) ctx . initialize ( settings . TransferAction . Download ) specs = settings . create_download_specifications ( ctx . cli_options , ctx . config ) del ctx . cli_options for spec in specs : blobxfer . api . Downloader ( ctx . general_options , ctx . credentials , spec ) . start ( )
Download blobs or files from Azure Storage
251,673
def synccopy ( ctx ) : settings . add_cli_options ( ctx . cli_options , settings . TransferAction . Synccopy ) ctx . initialize ( settings . TransferAction . Synccopy ) specs = settings . create_synccopy_specifications ( ctx . cli_options , ctx . config ) del ctx . cli_options for spec in specs : blobxfer . api . SyncCopy ( ctx . general_options , ctx . credentials , spec ) . start ( )
Synchronously copy blobs or files between Azure Storage accounts
251,674
def upload ( ctx ) : settings . add_cli_options ( ctx . cli_options , settings . TransferAction . Upload ) ctx . initialize ( settings . TransferAction . Upload ) specs = settings . create_upload_specifications ( ctx . cli_options , ctx . config ) del ctx . cli_options for spec in specs : blobxfer . api . Uploader ( ctx . general_options , ctx . credentials , spec ) . start ( )
Upload files to Azure Storage
251,675
def get_idp_sso_supported_bindings ( idp_entity_id = None , config = None ) : if config is None : from djangosaml2 . conf import get_config config = get_config ( ) meta = getattr ( config , 'metadata' , { } ) if idp_entity_id is None : try : idp_entity_id = list ( available_idps ( config ) . keys ( ) ) [ 0 ] except IndexError : raise ImproperlyConfigured ( "No IdP configured!" ) try : return meta . service ( idp_entity_id , 'idpsso_descriptor' , 'single_sign_on_service' ) . keys ( ) except UnknownSystemEntity : return [ ]
Returns the list of bindings supported by an IDP This is not clear in the pysaml2 code so wrapping it in a util
251,676
def fail_acs_response ( request , * args , ** kwargs ) : failure_function = import_string ( get_custom_setting ( 'SAML_ACS_FAILURE_RESPONSE_FUNCTION' , 'djangosaml2.acs_failures.template_failure' ) ) return failure_function ( request , * args , ** kwargs )
Serves as a common mechanism for ending ACS in case of any SAML related failure . Handling can be configured by setting the SAML_ACS_FAILURE_RESPONSE_FUNCTION as suitable for the project .
251,677
def echo_attributes ( request , config_loader_path = None , template = 'djangosaml2/echo_attributes.html' ) : state = StateCache ( request . session ) conf = get_config ( config_loader_path , request ) client = Saml2Client ( conf , state_cache = state , identity_cache = IdentityCache ( request . session ) ) subject_id = _get_subject_id ( request . session ) try : identity = client . users . get_identity ( subject_id , check_not_on_or_after = False ) except AttributeError : return HttpResponse ( "No active SAML identity found. Are you sure you have logged in via SAML?" ) return render ( request , template , { 'attributes' : identity [ 0 ] } )
Example view that echo the SAML attributes of an user
251,678
def logout ( request , config_loader_path = None ) : state = StateCache ( request . session ) conf = get_config ( config_loader_path , request ) client = Saml2Client ( conf , state_cache = state , identity_cache = IdentityCache ( request . session ) ) subject_id = _get_subject_id ( request . session ) if subject_id is None : logger . warning ( 'The session does not contain the subject id for user %s' , request . user ) result = client . global_logout ( subject_id ) state . sync ( ) if not result : logger . error ( "Looks like the user %s is not logged in any IdP/AA" , subject_id ) return HttpResponseBadRequest ( "You are not logged in any IdP/AA" ) if len ( result ) > 1 : logger . error ( 'Sorry, I do not know how to logout from several sources. I will logout just from the first one' ) for entityid , logout_info in result . items ( ) : if isinstance ( logout_info , tuple ) : binding , http_info = logout_info if binding == BINDING_HTTP_POST : logger . debug ( 'Returning form to the IdP to continue the logout process' ) body = '' . join ( http_info [ 'data' ] ) return HttpResponse ( body ) elif binding == BINDING_HTTP_REDIRECT : logger . debug ( 'Redirecting to the IdP to continue the logout process' ) return HttpResponseRedirect ( get_location ( http_info ) ) else : logger . error ( 'Unknown binding: %s' , binding ) return HttpResponseServerError ( 'Failed to log out' ) else : return finish_logout ( request , logout_info ) logger . error ( 'Could not logout because there only the HTTP_REDIRECT is supported' ) return HttpResponseServerError ( 'Logout Binding not supported' )
SAML Logout Request initiator
251,679
def do_logout_service ( request , data , binding , config_loader_path = None , next_page = None , logout_error_template = 'djangosaml2/logout_error.html' ) : logger . debug ( 'Logout service started' ) conf = get_config ( config_loader_path , request ) state = StateCache ( request . session ) client = Saml2Client ( conf , state_cache = state , identity_cache = IdentityCache ( request . session ) ) if 'SAMLResponse' in data : logger . debug ( 'Receiving a logout response from the IdP' ) response = client . parse_logout_request_response ( data [ 'SAMLResponse' ] , binding ) state . sync ( ) return finish_logout ( request , response , next_page = next_page ) elif 'SAMLRequest' in data : logger . debug ( 'Receiving a logout request from the IdP' ) subject_id = _get_subject_id ( request . session ) if subject_id is None : logger . warning ( 'The session does not contain the subject id for user %s. Performing local logout' , request . user ) auth . logout ( request ) return render ( request , logout_error_template , status = 403 ) else : http_info = client . handle_logout_request ( data [ 'SAMLRequest' ] , subject_id , binding , relay_state = data . get ( 'RelayState' , '' ) ) state . sync ( ) auth . logout ( request ) return HttpResponseRedirect ( get_location ( http_info ) ) else : logger . error ( 'No SAMLResponse or SAMLRequest parameter found' ) raise Http404 ( 'No SAMLResponse or SAMLRequest parameter found' )
SAML Logout Response endpoint
251,680
def metadata ( request , config_loader_path = None , valid_for = None ) : conf = get_config ( config_loader_path , request ) metadata = entity_descriptor ( conf ) return HttpResponse ( content = text_type ( metadata ) . encode ( 'utf-8' ) , content_type = "text/xml; charset=utf8" )
Returns an XML with the SAML 2 . 0 metadata for this SP as configured in the settings . py file .
251,681
def configure_user ( self , user , attributes , attribute_mapping ) : user . set_unusable_password ( ) return self . update_user ( user , attributes , attribute_mapping , force_save = True )
Configures a user after creation and returns the updated user .
251,682
def update_user ( self , user , attributes , attribute_mapping , force_save = False ) : if not attribute_mapping : return user user_modified = False for saml_attr , django_attrs in attribute_mapping . items ( ) : attr_value_list = attributes . get ( saml_attr ) if not attr_value_list : logger . debug ( 'Could not find value for "%s", not updating fields "%s"' , saml_attr , django_attrs ) continue for attr in django_attrs : if hasattr ( user , attr ) : user_attr = getattr ( user , attr ) if callable ( user_attr ) : modified = user_attr ( attr_value_list ) else : modified = self . _set_attribute ( user , attr , attr_value_list [ 0 ] ) user_modified = user_modified or modified else : logger . debug ( 'Could not find attribute "%s" on user "%s"' , attr , user ) logger . debug ( 'Sending the pre_save signal' ) signal_modified = any ( [ response for receiver , response in pre_user_save . send_robust ( sender = user . __class__ , instance = user , attributes = attributes , user_modified = user_modified ) ] ) if user_modified or signal_modified or force_save : user . save ( ) return user
Update a user with a set of attributes and returns the updated user .
251,683
def _set_attribute ( self , obj , attr , value ) : field = obj . _meta . get_field ( attr ) if field . max_length is not None and len ( value ) > field . max_length : cleaned_value = value [ : field . max_length ] logger . warn ( 'The attribute "%s" was trimmed from "%s" to "%s"' , attr , value , cleaned_value ) else : cleaned_value = value old_value = getattr ( obj , attr ) if cleaned_value != old_value : setattr ( obj , attr , cleaned_value ) return True return False
Set an attribute of an object to a specific value .
251,684
def config_settings_loader ( request = None ) : conf = SPConfig ( ) conf . load ( copy . deepcopy ( settings . SAML_CONFIG ) ) return conf
Utility function to load the pysaml2 configuration .
251,685
def mkpath ( * segments , ** query ) : segments = [ bytes_to_str ( s ) for s in segments if s is not None ] pathstring = '/' . join ( segments ) pathstring = re . sub ( '/+' , '/' , pathstring ) _query = { } for key in query : if query [ key ] in [ False , True ] : _query [ key ] = str ( query [ key ] ) . lower ( ) elif query [ key ] is not None : if PY2 and isinstance ( query [ key ] , unicode ) : _query [ key ] = query [ key ] . encode ( 'utf-8' ) else : _query [ key ] = query [ key ] if len ( _query ) > 0 : pathstring += "?" + urlencode ( _query ) if not pathstring . startswith ( '/' ) : pathstring = '/' + pathstring return pathstring
Constructs the path & query portion of a URI from path segments and a dict .
251,686
def search_index_path ( self , index = None , ** options ) : if not self . yz_wm_index : raise RiakError ( "Yokozuna search is unsupported by this Riak node" ) if index : quote_plus ( index ) return mkpath ( self . yz_wm_index , "index" , index , ** options )
Builds a Yokozuna search index URL .
251,687
def search_schema_path ( self , index , ** options ) : if not self . yz_wm_schema : raise RiakError ( "Yokozuna search is unsupported by this Riak node" ) return mkpath ( self . yz_wm_schema , "schema" , quote_plus ( index ) , ** options )
Builds a Yokozuna search Solr schema URL .
251,688
def to_op ( self ) : if not self . _adds : return None changes = { } if self . _adds : changes [ 'adds' ] = list ( self . _adds ) return changes
Extracts the modification operation from the Hll .
251,689
def add ( self , element ) : if not isinstance ( element , six . string_types ) : raise TypeError ( "Hll elements can only be strings" ) self . _adds . add ( element )
Adds an element to the HyperLogLog . Datatype cardinality will be updated when the object is saved .
251,690
def ping ( self ) : status , _ , body = self . _request ( 'GET' , self . ping_path ( ) ) return ( status is not None ) and ( bytes_to_str ( body ) == 'OK' )
Check server is alive over HTTP
251,691
def stats ( self ) : status , _ , body = self . _request ( 'GET' , self . stats_path ( ) , { 'Accept' : 'application/json' } ) if status == 200 : return json . loads ( bytes_to_str ( body ) ) else : return None
Gets performance statistics and server information
251,692
def get_keys ( self , bucket , timeout = None ) : bucket_type = self . _get_bucket_type ( bucket . bucket_type ) url = self . key_list_path ( bucket . name , bucket_type = bucket_type , timeout = timeout ) status , _ , body = self . _request ( 'GET' , url ) if status == 200 : props = json . loads ( bytes_to_str ( body ) ) return props [ 'keys' ] else : raise RiakError ( 'Error listing keys.' )
Fetch a list of keys for the bucket
251,693
def get_buckets ( self , bucket_type = None , timeout = None ) : bucket_type = self . _get_bucket_type ( bucket_type ) url = self . bucket_list_path ( bucket_type = bucket_type , timeout = timeout ) status , headers , body = self . _request ( 'GET' , url ) if status == 200 : props = json . loads ( bytes_to_str ( body ) ) return props [ 'buckets' ] else : raise RiakError ( 'Error getting buckets.' )
Fetch a list of all buckets
251,694
def get_bucket_props ( self , bucket ) : bucket_type = self . _get_bucket_type ( bucket . bucket_type ) url = self . bucket_properties_path ( bucket . name , bucket_type = bucket_type ) status , headers , body = self . _request ( 'GET' , url ) if status == 200 : props = json . loads ( bytes_to_str ( body ) ) return props [ 'props' ] else : raise RiakError ( 'Error getting bucket properties.' )
Get properties for a bucket
251,695
def set_bucket_props ( self , bucket , props ) : bucket_type = self . _get_bucket_type ( bucket . bucket_type ) url = self . bucket_properties_path ( bucket . name , bucket_type = bucket_type ) headers = { 'Content-Type' : 'application/json' } content = json . dumps ( { 'props' : props } ) status , _ , body = self . _request ( 'PUT' , url , headers , content ) if status == 401 : raise SecurityError ( 'Not authorized to set bucket properties.' ) elif status != 204 : raise RiakError ( 'Error setting bucket properties.' ) return True
Set the properties on the bucket object given
251,696
def clear_bucket_props ( self , bucket ) : bucket_type = self . _get_bucket_type ( bucket . bucket_type ) url = self . bucket_properties_path ( bucket . name , bucket_type = bucket_type ) url = self . bucket_properties_path ( bucket . name ) headers = { 'Content-Type' : 'application/json' } status , _ , _ = self . _request ( 'DELETE' , url , headers , None ) if status == 204 : return True elif status == 405 : return False else : raise RiakError ( 'Error %s clearing bucket properties.' % status )
reset the properties on the bucket object given
251,697
def get_bucket_type_props ( self , bucket_type ) : self . _check_bucket_types ( bucket_type ) url = self . bucket_type_properties_path ( bucket_type . name ) status , headers , body = self . _request ( 'GET' , url ) if status == 200 : props = json . loads ( bytes_to_str ( body ) ) return props [ 'props' ] else : raise RiakError ( 'Error getting bucket-type properties.' )
Get properties for a bucket - type
251,698
def set_bucket_type_props ( self , bucket_type , props ) : self . _check_bucket_types ( bucket_type ) url = self . bucket_type_properties_path ( bucket_type . name ) headers = { 'Content-Type' : 'application/json' } content = json . dumps ( { 'props' : props } ) status , _ , _ = self . _request ( 'PUT' , url , headers , content ) if status != 204 : raise RiakError ( 'Error setting bucket-type properties.' ) return True
Set the properties on the bucket - type
251,699
def mapred ( self , inputs , query , timeout = None ) : content = self . _construct_mapred_json ( inputs , query , timeout ) url = self . mapred_path ( ) headers = { 'Content-Type' : 'application/json' } status , headers , body = self . _request ( 'POST' , url , headers , content ) if status != 200 : raise RiakError ( 'Error running MapReduce operation. Headers: %s Body: %s' % ( repr ( headers ) , repr ( body ) ) ) result = json . loads ( bytes_to_str ( body ) ) return result
Run a MapReduce query .