bugged
stringlengths
4
228k
fixed
stringlengths
0
96.3M
__index_level_0__
int64
0
481k
def update_veto_lists(self, timeoffset, veto_definer = None):
def update_veto_lists(self, timeoffset, veto_definer = None):
479,500
def fromsegmentxml(file): """ Read a segmentlist from the file object file containing an xml segment table. """ xmldoc,digest = ligolw_utils.load_fileobj(file) seg_table = table.get_table(xmldoc,lsctables.SegmentTable.tableName) segs = segmentlist() for seg in seg_table: segs.append(segment(seg.start_time,seg.end_tim...
def fromsegmentxml(file): """ Read a segmentlist from the file object file containing an xml segment table. """ xmldoc,digest = utils.load_fileobj(file) seg_table = table.get_table(xmldoc,lsctables.SegmentTable.tableName) segs = segmentlist() for seg in seg_table: segs.append(segment(seg.start_time,seg.end_time)) seg...
479,501
def fromsegmentcsvCSV(csvfile): """ Read a segmentlist from the file object file containing a comma separated list of segments. """ def CSVLineToSeg(line): tstart, tend = map(int, line.split(',')) return segment(tstart, tend) segs = segmentlist([CSVLineToSeg(line) for line in csvfile]) return segs.coalesce()
def fromsegmentcsv(csvfile): """ Read a segmentlist from the file object file containing a comma separated list of segments. """ def CSVLineToSeg(line): tstart, tend = map(int, line.split(',')) return segment(tstart, tend) segs = segmentlist([CSVLineToSeg(line) for line in csvfile]) return segs.coalesce()
479,502
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ exe = make_external_call('which ligolw_segment_query')[0] #== construct segment query segment_cmd = ' '.join([exe,'--query-segments',\ '--database','--include-segments',fl...
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ exe = make_external_call('which ligolw_segment_query')[0] #== construct segment query segment_cmd = ' '.join([exe,'--query-segments',\ '--database','--include-segments',fl...
479,503
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ exe = make_external_call('which ligolw_segment_query')[0] #== construct segment query segment_cmd = ' '.join([exe,'--query-segments',\ '--database','--include-segments',fl...
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ exe = make_external_call('which ligolw_segment_query')[0] #== construct segment query segment_cmd = ' '.join([exe,'--query-segments',\ '--database','--include-segments',fl...
479,504
def write_abstract_dag(self): """ Write all the nodes in the workflow to the DAX file. """ if not self.__dax_file_path: # this workflow is not dax-compatible, so don't write a dax return try: dagfile = open( self.__dax_file_path, 'w' ) except: raise CondorDAGError, "Cannot open file " + self.__dag_file_path
def write_abstract_dag(self): """ Write all the nodes in the workflow to the DAX file. """ if not self.__dax_file_path: # this workflow is not dax-compatible, so don't write a dax return try: dagfile = open( self.__dax_file_path, 'w' ) except: raise CondorDAGError, "Cannot open file " + self.__dag_file_path
479,505
def add_dq(self, page): page.add_section("DQ", "Data Quality for %s" % (self.coinctime,)) page.sections["DQ"].div("This section gives vetoes and flags that were on")
def add_dq(self, page): page.add_section("DQ", "Data Quality for %s" % (self.coinctime,)) page.sections["DQ"].div("This section gives vetoes and flags that were on")
479,506
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarseg...
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarsegr...
479,507
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarseg...
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarseg...
479,508
def get_unique_filename(name): """ use this to avoid name collisions """ counter = 1 base_name, ext = os.path.splitext(name) while os.path.isfile(base_name): base_name = base_name + '_' + str(counter) + ext counter += 1 return base_name + ext
def get_unique_filename(name): """ use this to avoid name collisions """ counter = 1 base_name, ext = os.path.splitext(name) while os.path.isfile(name): name = base_name + '_' + str(counter) + ext counter += 1 return base_name + ext
479,509
def get_unique_filename(name): """ use this to avoid name collisions """ counter = 1 base_name, ext = os.path.splitext(name) while os.path.isfile(base_name): base_name = base_name + '_' + str(counter) + ext counter += 1 return base_name + ext
def get_unique_filename(name): """ use this to avoid name collisions """ counter = 1 base_name, ext = os.path.splitext(name) while os.path.isfile(base_name): base_name = base_name + '_' + str(counter) + ext counter += 1 return name
479,510
def convert_duration( duration ): return sqlutils.convert_duration( duration, convert_durations )
def convert_duration( duration ): return sqlutils.convert_duration( duration, convert_durations )
479,511
def get_pyvalue(self): return generic_get_pyvalue(self)
def get_pyvalue(self): return generic_get_pyvalue(self)
479,512
def get_decisive_distance( *args ): return sorted(args)[1]
def get_decisive_distance( *args ): return sorted(args)[1]
479,513
def grab_data(start,end,channel,type,\ nds=False,verbose=False,dmt=False): """ This function will return the frame data for the given channel of the given type in the given [start,end] time range and will construct a gps time vector to go with it. The nds option is not yet supported, and the dmt option will return data...
def grab_data(start,end,channel,type,\ nds=False,verbose=False,dmt=False): """ This function will return the frame data for the given channel of the given type in the given [start,end] time range and will construct a gps time vector to go with it. The nds option is not yet supported, and the dmt option will return data...
479,514
def generate_cache(start_time,end_time,ifos,types,return_files=False): """ This function will return a cache of files as found by ligo_data_find, given start and end time, and lists of ifos and types. If the return_files option is given as 'True' the function will return a list of frames with absolute paths, otherwise ...
def generate_cache(start_time,end_time,ifos,types,return_files=False): """ This function will return a cache of files as found by ligo_data_find, given start and end time, and lists of ifos and types. If the return_files option is given as 'True' the function will return a list of frames with absolute paths, otherwise ...
479,515
def generate_cache(start_time,end_time,ifos,types,return_files=False): """ This function will return a cache of files as found by ligo_data_find, given start and end time, and lists of ifos and types. If the return_files option is given as 'True' the function will return a list of frames with absolute paths, otherwise ...
def generate_cache(start_time,end_time,ifos,types,return_files=False): """ This function will return a cache of files as found by ligo_data_find, given start and end time, and lists of ifos and types. If the return_files option is given as 'True' the function will return a list of frames with absolute paths, otherwise ...
479,516
def generate_cache(start_time,end_time,ifos,types,return_files=False): """ This function will return a cache of files as found by ligo_data_find, given start and end time, and lists of ifos and types. If the return_files option is given as 'True' the function will return a list of frames with absolute paths, otherwise ...
def generate_cache(start_time,end_time,ifos,types,return_files=False): """ This function will return a cache of files as found by ligo_data_find, given start and end time, and lists of ifos and types. If the return_files option is given as 'True' the function will return a list of frames with absolute paths, otherwise ...
479,517
def find_types(types,search='standard'): """ This function will return a valid list of LIGO frame types given the list of type strings. The search option defines the breadth of the search, to speed up the search, the following search options are supported: 'standard','short','full'. The 'R', 'T', and 'M' (raw, raw sec...
def find_types(types,search='standard'): """ This function will return a valid list of LIGO frame types given the list of type strings. The search option defines the breadth of the search, to speed up the search, the following search options are supported: 'standard','short','full'. The 'R', 'T', and 'M' (raw, raw sec...
479,518
def find_types(types,search='standard'): """ This function will return a valid list of LIGO frame types given the list of type strings. The search option defines the breadth of the search, to speed up the search, the following search options are supported: 'standard','short','full'. The 'R', 'T', and 'M' (raw, raw sec...
def find_types(types,search='standard'): """ This function will return a valid list of LIGO frame types given the list of type strings. The search option defines the breadth of the search, to speed up the search, the following search options are supported: 'standard','short','full'. The 'R', 'T', and 'M' (raw, raw sec...
479,519
def find_types(types,search='standard'): """ This function will return a valid list of LIGO frame types given the list of type strings. The search option defines the breadth of the search, to speed up the search, the following search options are supported: 'standard','short','full'. The 'R', 'T', and 'M' (raw, raw sec...
def find_types(types,search='standard'): """ This function will return a valid list of LIGO frame types given the list of type strings. The search option defines the breadth of the search, to speed up the search, the following search options are supported: 'standard','short','full'. The 'R', 'T', and 'M' (raw, raw sec...
479,520
def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_...
def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_...
479,521
def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_...
def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_...
479,522
def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_...
def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_...
479,523
def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_...
def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_...
479,524
def detector_thresholds(min_threshold, ifos, RA, dec, gps_time, sensitivities=None): """ Return a dictionary of sensitivity thresholds for each detector, based on a minimum threshold of min_threshold in the least sensitive one, for a source at position (RA,dec) specified in radians at time gps_time. Specifying a dictio...
def detector_thresholds(min_threshold, ifos, RA, dec, gps_time, sensitivities=None): """ Return a dictionary of sensitivity thresholds for each detector, based on a minimum threshold of min_threshold in the least sensitive one, for a source at position (RA,dec) specified in radians at time gps_time. Specifying a dictio...
479,525
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # # Check to see if wiki file with name already exists # maxCount=0 while os.pat...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # # Check to see if wiki file with name already exists # maxCount=0 while os.pat...
479,526
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # # Check to see if wiki file with name already exists # maxCount=0 while os.pat...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # # Check to see if wiki file with name already exists # maxCount=0 while os.pat...
479,527
def greedyBin2(posterior_array,par_bins,confidence_levels,par_names=None,injection=None): if par_names: par1_name,par2_name=par_names else: par1_name="Parameter 1" par2_name="Parameter 2" par1pos=posterior_array[:,0] par2pos=posterior_array[:,1] par1_bin,par2_bin=par_bins if injection: par1_injvalue,par2_injvalue=i...
def greedyBin2(posterior_array,par_bins,confidence_levels,par_names=None,injection=None): if par_names: par1_name,par2_name=par_names else: par1_name="Parameter 1" par2_name="Parameter 2" par1pos=posterior_array[:,0] par2pos=posterior_array[:,1] par1_bin,par2_bin=par_bins if injection: par1_injvalue,par2_injvalue=i...
479,528
coldef = create_col(coldict['nolong'])
coldef = create_col(coldict['nolong'])
479,529
def run(self): # remove the automatically generated user env scripts for script in ["pylal-user-env.sh", "pylal-user-env.csh"]: log.info("removing " + script ) try: os.unlink(os.path.join("etc", script)) except: pass
def run(self): # remove the automatically generated user env scripts for script in ["pylal-user-env.sh", "pylal-user-env.csh"]: log.info("removing " + script ) try: os.unlink(os.path.join("etc", script)) except: pass
479,530
def add_content(self, data, label="_nolabel_"): """ Stub. Replace with a method that appends values or lists of values to self.data_sets and appends labels to self.data_labels. Feel free to accept complicated inputs, but try to store only the raw numbers that will enter the plot. """ raise NotImplemented
def add_content(self, data, label="_nolabel_"): """ Stub. Replace with a method that appends values or lists of values to self.data_sets and appends labels to self.data_labels. Feel free to accept complicated inputs, but try to store only the raw numbers that will enter the plot. """ raise NotImplemented
479,531
def finalize(self): """ Stub. Replace with a function that creates and makes your plot pretty. Do not do I/O here. """ raise NotImplemented
def finalize(self): """ Stub. Replace with a function that creates and makes your plot pretty. Do not do I/O here. """ raise NotImplemented
479,532
def __init__(self, tag, title="", secnum="1", pagenum="1", level=2): markup.page.__init__(self, mode="strict_html") self.pagenum = pagenum self.secnum = secnum self._title = title self.sections = {} self.section_ids = [] self.level = level self.tag = tag self.id = tag + self.secnum self.tables = 0 self.add('<div class=...
def __init__(self, tag, title="", secnum="1", pagenum="1", level=2, open_by_default=False): markup.page.__init__(self, mode="strict_html") self.pagenum = pagenum self.secnum = secnum self._title = title self.sections = {} self.section_ids = [] self.level = level self.tag = tag self.id = tag + self.secnum self.tables = ...
479,533
def __init__(self, tag, title="", secnum="1", pagenum="1", level=2): markup.page.__init__(self, mode="strict_html") self.pagenum = pagenum self.secnum = secnum self._title = title self.sections = {} self.section_ids = [] self.level = level self.tag = tag self.id = tag + self.secnum self.tables = 0 self.add('<div class=...
def __init__(self, tag, title="", secnum="1", pagenum="1", level=2): markup.page.__init__(self, mode="strict_html") self.pagenum = pagenum self.secnum = secnum self._title = title self.sections = {} self.section_ids = [] self.level = level self.tag = tag self.id = tag + self.secnum self.tables = 0 self.add('<div class=...
479,534
def add_section(self, tag, title=""): secnum = "%s.%d" % (self.secnum, len(self.sections.values())+1) self.sections[tag] = _section(tag, title=title, secnum=secnum, pagenum=self.pagenum, level=self.level+1) self.section_ids.append([len(self.sections.values()), tag]) return self.sections[tag]
def add_section(self, tag, title=""): secnum = "%s.%d" % (self.secnum, len(self.sections.values())+1) self.sections[tag] = _section(tag, title=title, secnum=secnum, pagenum=self.pagenum, level=self.level+1, open_by_default=open_by_default) self.section_ids.append([len(self.sections.values()), tag]) return self.sections...
479,535
def add_section(self, tag, title="", level=2): """ """ secnum = len(self.sections.values()) + 1 self.section_ids.append([secnum, tag]) self.sections[tag] = _section(title=title, tag=tag, secnum=str(secnum), pagenum=str(self.pagenum), level=level) return self.sections[tag]
def add_section(self, tag, title="", level=2, open_by_default=False): """ """ secnum = len(self.sections.values()) + 1 self.section_ids.append([secnum, tag]) self.sections[tag] = _section(title=title, tag=tag, secnum=str(secnum), pagenum=str(self.pagenum), level=level) return self.sections[tag]
479,536
def add_section(self, tag, title="", level=2): """ """ secnum = len(self.sections.values()) + 1 self.section_ids.append([secnum, tag]) self.sections[tag] = _section(title=title, tag=tag, secnum=str(secnum), pagenum=str(self.pagenum), level=level) return self.sections[tag]
def add_section(self, tag, title="", level=2): """ """ secnum = len(self.sections.values()) + 1 self.section_ids.append([secnum, tag]) self.sections[tag] = _section(title=title, tag=tag, secnum=str(secnum), pagenum=str(self.pagenum), level=level, open_by_default=open_by_default) return self.sections[tag]
479,537
def run(self): # remove the automatically generated user env scripts for script in ["pylal-user-env.sh", "pylal-user-env.csh"]: log.info("removing " + script ) try: os.unlink(os.path.join("etc", script)) except: pass
def run(self): # remove the automatically generated user env scripts for script in ["pylal-user-env.sh", "pylal-user-env.csh"]: log.info("removing " + script ) try: os.unlink(os.path.join("etc", script)) except: pass
479,538
def run(self): # remove the automatically generated user env scripts for script in [ 'glue-user-env.sh', 'glue-user-env.csh' ]: log.info( 'removing ' + script ) try: os.unlink(os.path.join('etc',script)) except: pass
def run(self): # remove the automatically generated user env scripts for script in [ 'glue-user-env.sh', 'glue-user-env.csh' ]: log.info( 'removing ' + script ) try: os.unlink(os.path.join('etc',script)) except: pass
479,539
def finish(self, filters = {}, verbose = False): default_filter = rate.gaussian_window(21) # normalizing each array so that its sum is 1 has the # effect of making the integral of P(x) dx equal 1 after # the array is transformed to an array of densities (which # is done by dividing each bin by dx). N = len(self.zero_la...
def finish(self, filters = {}, verbose = False): default_filter = rate.gaussian_window(21) # normalizing each array so that its sum is 1 has the # effect of making the integral of P(x) dx equal 1 after # the array is transformed to an array of densities (which # is done by dividing each bin by dx). N = len(self.zero_la...
479,540
def load_fileobj(fileobj, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file object fileobj, and return the contents as a LIGO Light Weight document tree. The file object does not need to be seekable. The file is gzip decompressed while reading if gz is set to True. If the optional...
def load_fileobj(fileobj, gz = None, xmldoc = None, contenthandler = None): """ Parse the contents of the file object fileobj, and return the contents as a LIGO Light Weight document tree. The file object does not need to be seekable. The file is gzip decompressed while reading if gz is set to True. If the optional ...
479,541
def load_fileobj(fileobj, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file object fileobj, and return the contents as a LIGO Light Weight document tree. The file object does not need to be seekable. The file is gzip decompressed while reading if gz is set to True. If the optional...
def load_fileobj(fileobj, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file object fileobj, and return the contents as a LIGO Light Weight document tree. The file object does not need to be seekable. If the gz parameter is None (the default) then gzip compressed data will be automa...
479,542
def load_fileobj(fileobj, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file object fileobj, and return the contents as a LIGO Light Weight document tree. The file object does not need to be seekable. The file is gzip decompressed while reading if gz is set to True. If the optional...
def load_fileobj(fileobj, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file object fileobj, and return the contents as a LIGO Light Weight document tree. The file object does not need to be seekable. The file is gzip decompressed while reading if gz is set to True. If the optional...
479,543
def load_fileobj(fileobj, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file object fileobj, and return the contents as a LIGO Light Weight document tree. The file object does not need to be seekable. The file is gzip decompressed while reading if gz is set to True. If the optional...
def load_fileobj(fileobj, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file object fileobj, and return the contents as a LIGO Light Weight document tree. The file object does not need to be seekable. The file is gzip decompressed while reading if gz is set to True. If the optional...
479,544
def load_filename(filename, verbose = False, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file identified by filename, and return the contents as a LIGO Light Weight document tree. Helpful verbosity messages are printed to stderr if verbose is True, and the file is gzip decompressed...
def load_filename(filename, verbose = False, gz = None, xmldoc = None, contenthandler = None): """ Parse the contents of the file identified by filename, and return the contents as a LIGO Light Weight document tree. Helpful verbosity messages are printed to stderr if verbose is True, and the file is gzip decompressed ...
479,545
def load_filename(filename, verbose = False, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file identified by filename, and return the contents as a LIGO Light Weight document tree. Helpful verbosity messages are printed to stderr if verbose is True, and the file is gzip decompressed...
def load_filename(filename, verbose = False, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file identified by filename, and return the contents as a LIGO Light Weight document tree. Helpful verbosity messages are printed to stderr if verbose is True. All other parameters are passed ...
479,546
def load_filename(filename, verbose = False, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file identified by filename, and return the contents as a LIGO Light Weight document tree. Helpful verbosity messages are printed to stderr if verbose is True, and the file is gzip decompressed...
def load_filename(filename, verbose = False, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file identified by filename, and return the contents as a LIGO Light Weight document tree. Helpful verbosity messages are printed to stderr if verbose is True, and the file is gzip decompressed...
479,547
def load_filename(filename, verbose = False, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file identified by filename, and return the contents as a LIGO Light Weight document tree. Helpful verbosity messages are printed to stderr if verbose is True, and the file is gzip decompressed...
def load_filename(filename, verbose = False, gz = False, xmldoc = None, contenthandler = None): """ Parse the contents of the file identified by filename, and return the contents as a LIGO Light Weight document tree. Helpful verbosity messages are printed to stderr if verbose is True, and the file is gzip decompressed...
479,548
def load_url(url, verbose = False, gz = False, xmldoc = None, contenthandler = None): """ This function has the same behaviour as load_filename() but accepts a URL instead of a filename. Any source from which Python's urllib2 library can read data is acceptable. stdin is parsed if the URL is None. If the optional xm...
def load_url(url, verbose = False, gz = None, xmldoc = None, contenthandler = None): """ This function has the same behaviour as load_filename() but accepts a URL instead of a filename. Any source from which Python's urllib2 library can read data is acceptable. stdin is parsed if the URL is None. If the optional xml...
479,549
def load_url(url, verbose = False, gz = False, xmldoc = None, contenthandler = None): """ This function has the same behaviour as load_filename() but accepts a URL instead of a filename. Any source from which Python's urllib2 library can read data is acceptable. stdin is parsed if the URL is None. If the optional xm...
def load_url(url, verbose = False, gz = False, xmldoc = None, contenthandler = None): """ This function has the same behaviour as load_filename() but accepts a URL instead of a filename. Any source from which Python's urllib2 library can read data is acceptable. stdin is parsed if the URL is None. Example: >>> from...
479,550
def __init__(self, configfile=None): cp = ConfigParser.ConfigParser() self.cp = cp self.time_now = "_".join([str(i) for i in time_method.gmtime()[0:6]]) self.ini_file=self.time_now + ".ini" home_base = home_dirs() # CONDOR SECTION NEEDED BY THINGS IN INSPIRAL.PY cp.add_section("condor") cp.set("condor","datafind",self...
def __init__(self, configfile=None): cp = ConfigParser.ConfigParser() self.cp = cp self.time_now = "_".join([str(i) for i in time_method.gmtime()[0:6]]) self.ini_file=self.time_now + ".ini" home_base = home_dirs() # CONDOR SECTION NEEDED BY THINGS IN INSPIRAL.PY cp.add_section("condor") cp.set("condor","datafind",self...
479,551
def figure_out_site(ifo): siteDico = {"H1":"LHO","H2":"LHO","L1":"LLO","V1":"Virgo"} if ifo in siteDico: return siteDico[ifo] else: print >> sys.stderr, "ifo " + ifo + "is not defined in siteDico dictionary" sys.exit(1)
479,552
def __init__(self, dag, job, cp, opts, time, ifo, frame_cache, p_nodes=[], type="ht", variety="fg"): """ """ pipeline.CondorDAGNode.__init__(self,job)
def __init__(self, dag, job, cp, opts, time, ifo, frame_cache, p_nodes=[], type="ht", variety="fg"): """ """ pipeline.CondorDAGNode.__init__(self,job)
479,553
def __init__(self, dag, job, cp, opts, time, ifo, p_nodes=[], type="ht", variety="fg"):
def __init__(self, dag, job, cp, opts, time, ifo, p_nodes=[], type="ht", variety="fg"):
479,554
def latexnumber(s): """ Convert a string of the form "d.dddde-dd" to "d.dddd \times 10^{-dd}" """ return re.sub(r"([+-]?[.0-9]+)[Ee]?([+-]?[0-9]+)", r"\1 \\times 10^{\2}", s)
def latexnumber(s): """ Convert a string of the form "d.dddde-dd" to "d.dddd \times 10^{-dd}" """ m, e = floatpattern.match(s).groups() return r"%s \\times 10^{%d}" % (m, int(e))
479,555
def set_temp_store_directory(connection, temp_store_directory, verbose = False): """ Sets the temp_store_directory parameter in sqlite. """ if verbose: print >>sys.stderr, "setting the temp_store_directory to %s ..." % temp_store_directory, cursor = connection.cursor() cursor.execute("PRAGMA temp_store_directory = '%s'...
def set_temp_store_directory(connection, temp_store_directory, verbose = False): """ Sets the temp_store_directory parameter in sqlite. """ if verbose: print >>sys.stderr, "setting the temp_store_directory to %s ..." % temp_store_directory, cursor = connection.cursor() cursor.execute("PRAGMA temp_store_directory = '%s'...
479,556
def newsigterm(signum, frame): global __llwapp_write_filename_got_sig __llwapp_write_filename_got_sig.append(signum)
def newsigterm(signum, frame): global __llwapp_write_filename_got_sig __llwapp_write_filename_got_sig.append(signum)
479,557
def newsigterm(signum, frame): global __llwapp_write_filename_got_sig __llwapp_write_filename_got_sig.append(signum)
def newsigterm(signum, frame): global __llwapp_write_filename_got_sig __llwapp_write_filename_got_sig.append(signum)
479,558
def background_livetime_ring_by_slide(connection, live_time_program, seglists, veto_segments, verbose = False): background_livetime = {} instruments = frozenset(seglists.keys()) offset_vectors = db_thinca_rings.get_background_offset_vectors(connection) # first work out time slide live time for on_instruments, livetimes...
def background_livetime_ring_by_slide(connection, live_time_program, seglists, veto_segments, verbose = False): background_livetime = {} instruments = frozenset(seglists.keys()) offset_vectors = db_thinca_rings.get_background_offset_vectors(connection) # first work out time slide live time for on_instruments, livetimes...
479,559
def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ # # loop overall the bins in cafepacker.bins. we pop items out of # cafepacker.bins and append new ones to the end so need a while loop # checking the extent of each bin in cafep...
def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ # # loop overall the bins in cafepacker.bins. we pop items out of # cafepacker.bins and append new ones to the end so need a while loop # checking the extent of each bin in cafep...
479,560
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
479,561
def make_external_call(command, show_stdout=False, show_command=False): """ Run a program on the shell and print informative messages on failure. """ if show_command: print command stdin, out, err = os.popen3(command) pid, status = os.wait() if status != 0: print >>sys.stderr, "External call failed." print >>sys.stde...
def make_external_call(command, show_stdout=False, show_command=False): """ Run a program on the shell and print informative messages on failure. """ if show_command: print command p = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=isinstance(command, str)) out, err = p.communicate() ...
479,562
def make_external_call(command, show_stdout=False, show_command=False): """ Run a program on the shell and print informative messages on failure. """ if show_command: print command stdin, out, err = os.popen3(command) pid, status = os.wait() if status != 0: print >>sys.stderr, "External call failed." print >>sys.stde...
def make_external_call(command, show_stdout=False, show_command=False): """ Run a program on the shell and print informative messages on failure. """ if show_command: print command stdin, out, err = os.popen3(command) pid, status = os.wait() if status != 0: print >>sys.stderr, "External call failed." print >>sys.stde...
479,563
def make_external_call(command, show_stdout=False, show_command=False): """ Run a program on the shell and print informative messages on failure. """ if show_command: print command stdin, out, err = os.popen3(command) pid, status = os.wait() if status != 0: print >>sys.stderr, "External call failed." print >>sys.stde...
def make_external_call(command, show_stdout=False, show_command=False): """ Run a program on the shell and print informative messages on failure. """ if show_command: print command stdin, out, err = os.popen3(command) pid, status = os.wait() if status != 0: print >>sys.stderr, "External call failed." print >>sys.stde...
479,564
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [(latitude-0.5*pi, longitude)] whi...
def gridsky(resolution,shifted=False): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [(latitude-0.5*pi, l...
479,565
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [(latitude-0.5*pi, longitude)] whi...
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = 0.0 ds = pi*resolution/180.0 points = [(latitude-0.5*pi, longitude)] wh...
479,566
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [(latitude-0.5*pi, longitude)] whi...
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [] if shifted: dlat = 0.0 else: dl...
479,567
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [(latitude-0.5*pi, longitude)] whi...
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [(latitude-0.5*pi, longitude)] whi...
479,568
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [(latitude-0.5*pi, longitude)] whi...
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [(latitude-0.5*pi, longitude)] whi...
479,569
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [(latitude-0.5*pi, longitude)] whi...
def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*resolution/180.0 points = [(latitude-0.5*pi, longitude)] whi...
479,570
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarsegr...
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns (1) a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coar...
479,571
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarsegr...
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values and (2) (for debugging purposes) returns a list of points in the fine grid that didn...
479,572
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarsegr...
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarsegr...
479,573
def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarsegr...
deffirst_col = [pt for pt in coarsegrid if pt[1] == 0.0] for cpt in first_col: flist = [] for fpt in fgtemp: if (cpt[0]-fpt[0])*(cpt[0]-fpt[0])-ds*ds/4.0 <= epsilon and \ (2*pi-fpt[1])*(2*pi-fpt[1])*sin(cpt[0])*sin(cpt[0])-ds*ds/4.0 <= epsilon: coarsedict[cpt].append(fpt) flist.append(fpt) for rpt in flist: fgtemp.remo...
479,574
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ #== construct segment query segment_cmd = "ligolw_segment_query --query-segments"+\ " --segment_url https://segdb.ligo.caltech.edu"+\ " --include-segments "+flag+\ " --gp...
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ #== construct segment query segment_cmd = "ligolw_segment_query --query-segments"+\ " --segment-url https://segdb.ligo.caltech.edu"+\ " --include-segments "+flag+\ " --gp...
479,575
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ #== construct segment query segment_cmd = "ligolw_segment_query --query-segments"+\ " --segment_url https://segdb.ligo.caltech.edu"+\ " --include-segments "+flag+\ " --gp...
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ #== construct segment query segment_cmd = "ligolw_segment_query --query-segments"+\ " --segment_url https://segdb.ligo.caltech.edu"+\ " --include-segments "+flag+\ " --gp...
479,576
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ #== construct segment query segment_cmd = "ligolw_segment_query --query-segments"+\ " --segment_url https://segdb.ligo.caltech.edu"+\ " --include-segments "+flag+\ " --gp...
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ #== construct segment query segment_cmd = "ligolw_segment_query --query-segments"+\ " --segment_url https://segdb.ligo.caltech.edu"+\ " --include-segments "+flag+\ " --gp...
479,577
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ #== construct segment query segment_cmd = "ligolw_segment_query --query-segments"+\ " --segment_url https://segdb.ligo.caltech.edu"+\ " --include-segments "+flag+\ " --gp...
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ #== construct segment query segment_cmd = "ligolw_segment_query --query-segments"+\ " --segment_url https://segdb.ligo.caltech.edu"+\ " --include-segments "+flag+\ " --gp...
479,578
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ #== construct segment query segment_cmd = "ligolw_segment_query --query-segments"+\ " --segment_url https://segdb.ligo.caltech.edu"+\ " --include-segments "+flag+\ " --gp...
def grab_segments(start,end,flag): """ Returns a segmentlist containing the segments during which the given flag was active in the given period. """ #== construct segment query segment_cmd = "ligolw_segment_query --query-segments"+\ " --segment_url https://segdb.ligo.caltech.edu"+\ " --include-segments "+flag+\ " --gp...
479,579
def createDQbackground(self,ifoEpochList=list(),pickleLocale=None): """ Two inputs a list of tuples (ifo,epochname) for each instrument. Also a place to save the potential pickle to for quick access later. """ if type(ifoEpochList) != type(list()): raise Exception, \ "Invalid input argument ifoEpochList,%s type(%s)"...
def createDQbackground(self,ifoEpochList=list(),pickleLocale=None): """ Two inputs a list of tuples (ifo,epochname) for each instrument. Also a place to save the potential pickle to for quick access later. """ if type(ifoEpochList) != type(list()): raise Exception, \ "Invalid input argument ifoEpochList,%s type(%s)"...
479,580
def createDQbackground(self,ifoEpochList=list(),pickleLocale=None): """ Two inputs a list of tuples (ifo,epochname) for each instrument. Also a place to save the potential pickle to for quick access later. """ if type(ifoEpochList) != type(list()): raise Exception, \ "Invalid input argument ifoEpochList,%s type(%s)"...
def createDQbackground(self,ifoEpochList=list(),pickleLocale=None): """ Two inputs a list of tuples (ifo,epochname) for each instrument. Also a place to save the potential pickle to for quick access later. """ if type(ifoEpochList) != type(list()): raise Exception, \ "Invalid input argument ifoEpochList,%s type(%s)"...
479,581
def createDQbackground(self,ifoEpochList=list(),pickleLocale=None): """ Two inputs a list of tuples (ifo,epochname) for each instrument. Also a place to save the potential pickle to for quick access later. """ if type(ifoEpochList) != type(list()): raise Exception, \ "Invalid input argument ifoEpochList,%s type(%s)"...
def createDQbackground(self,ifoEpochList=list(),pickleLocale=None): """ Two inputs a list of tuples (ifo,epochname) for each instrument. Also a place to save the potential pickle to for quick access later. """ if type(ifoEpochList) != type(list()): raise Exception, \ "Invalid input argument ifoEpochList,%s type(%s)"...
479,582
def insert(connection, urls, preserve_ids = False, verbose = False): """ Iterate over a sequence of URLs and parse and insert each one into the database the dbtables.DBTable class is currently connected to. """ if not preserve_ids: # enable ID remapping dbtables.idmap_create(connection) dbtables.DBTable.append = dbtabl...
def insert_from_urls(connection, urls, preserve_ids = False, verbose = False): """ Iterate over a sequence of URLs and parse and insert each one into the database the dbtables.DBTable class is currently connected to. """ if not preserve_ids: # enable ID remapping dbtables.idmap_create(connection) dbtables.DBTable.appen...
479,583
def insert_from_xmldoc(connection, xmldoc, preserve_ids = False, verbose = False): """ Insert the tables from an in-ram XML document into the database at the given connection. """ if not preserve_ids: # enable ID remapping dbtables.idmap_create(connection) dbtables.DBTable.append = dbtables.DBTable._remapping_append el...
def insert_from_xmldoc(connection, xmldoc, preserve_ids = False, verbose = False): """ Insert the tables from an in-ram XML document into the database at the given connection. """ if not preserve_ids: # enable ID remapping dbtables.idmap_create(connection) dbtables.DBTable.append = dbtables.DBTable._remapping_append el...
479,584
def ligolw_burca2(database, likelihood_ratio, params_func, verbose = False, params_func_extra_args = ()): """ Assigns likelihood ratio values to excess power coincidences. database is pylal.SnglBurstUtils.CoincDatabase instance, and likelihood_ratio is a LikelihoodRatio class instance. """ # # Find document parts. # t...
def ligolw_burca2(database, likelihood_ratio, params_func, verbose = False, params_func_extra_args = ()): """ Assigns likelihood ratio values to excess power coincidences. database is pylal.SnglBurstUtils.CoincDatabase instance, and likelihood_ratio is a LikelihoodRatio class instance. """ # # Find document parts. # t...
479,585
def ligolw_burca2(database, likelihood_ratio, params_func, verbose = False, params_func_extra_args = ()): """ Assigns likelihood ratio values to excess power coincidences. database is pylal.SnglBurstUtils.CoincDatabase instance, and likelihood_ratio is a LikelihoodRatio class instance. """ # # Find document parts. # t...
def ligolw_burca2(database, likelihood_ratio, params_func, verbose = False, params_func_extra_args = ()): """ Assigns likelihood ratio values to excess power coincidences. database is pylal.SnglBurstUtils.CoincDatabase instance, and likelihood_ratio is a LikelihoodRatio class instance. """ # # Find document parts. # t...
479,586
def ligolw_burca2(database, likelihood_ratio, params_func, verbose = False, params_func_extra_args = ()): """ Assigns likelihood ratio values to excess power coincidences. database is pylal.SnglBurstUtils.CoincDatabase instance, and likelihood_ratio is a LikelihoodRatio class instance. """ # # Find document parts. # t...
def ligolw_burca2(database, likelihood_ratio, params_func, verbose = False, params_func_extra_args = ()): """ Assigns likelihood ratio values to excess power coincidences. database is pylal.SnglBurstUtils.CoincDatabase instance, and likelihood_ratio is a LikelihoodRatio class instance. """ # # Find document parts. # t...
479,587
def __filenameToChannelList__(self,filenameList=[]): """ This method attempts to construct a set of simplified channel names based of a list of image filenames. """ #Parsed filename channel list specialChannelList=[ "ISI-OMC_DISPPF_H1_IN1_DAQ",\ "PEM-LVEA2_V2",\ "OMC-ASC_POS_X_IN1_DAQ",\ "OMC-ASC_POS_Y_IN1_DAQ",\ "OMC-...
def __filenameToChannelList__(self,filenameList=[]): """ This method attempts to construct a set of simplified channel names based of a list of image filenames. """ #Parsed filename channel list specialChannelList=[ "XX:ISI-OMC_DISPPF_H1_IN1_DAQ",\ "XX:PEM-LVEA2_V2",\ "XX:OMC-ASC_POS_X_IN1_DAQ",\ "XX:OMC-ASC_POS_Y_IN1_...
479,588
def __filenameToChannelList__(self,filenameList=[]): """ This method attempts to construct a set of simplified channel names based of a list of image filenames. """ #Parsed filename channel list specialChannelList=[ "ISI-OMC_DISPPF_H1_IN1_DAQ",\ "PEM-LVEA2_V2",\ "OMC-ASC_POS_X_IN1_DAQ",\ "OMC-ASC_POS_Y_IN1_DAQ",\ "OMC-...
def __filenameToChannelList__(self,filenameList=[]): """ This method attempts to construct a set of simplified channel names based of a list of image filenames. """ #Parsed filename channel list specialChannelList=[ "ISI-OMC_DISPPF_H1_IN1_DAQ",\ "PEM-LVEA2_V2",\ "OMC-ASC_POS_X_IN1_DAQ",\ "OMC-ASC_POS_Y_IN1_DAQ",\ "OMC-...
479,589
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
479,590
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
479,591
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
479,592
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
479,593
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
479,594
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
479,595
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
479,596
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
defprepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None):"""Methodtoprepareachecklistwheredataproductsareisolatedindirectory."""endOfS5=int(875232014)wikiFileFinder=findFileType(wikiTree,wikiCoinc)#ChecktoseeifwikifilewithnamealreadyexistsmaxCount=0whileos.path.exists(wikiFilename)andmaxCount<1...
479,597
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
479,598
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex...
479,599