rem stringlengths 1 322k | add stringlengths 0 2.05M | context stringlengths 4 228k | meta stringlengths 156 215 |
|---|---|---|---|
imageDictAQ[sngl.ifo]=list() indexDictAQ[sngl.ifo]=list() thumbDictAQ[sngl.ifo]=list() | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py | |
for myFile in fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*html"%(sngl.ifo,timeString)): indexDictAQ[sngl.ifo].append(myFile) zValueFiles=fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*txt"%(sngl.ifo,timeString)) zValueDictAQ[sngl.ifo]=list() if len(zValueFiles)>0: for zFile in z... | for myFile in fnmatch.filter(filesAnalyze,\ | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py |
for myFile in fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ | for myFile in fnmatch.filter(filesAnalyze,\ | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py |
imageDict=dict() indexDict=dict() thumbDict=dict() zValueDict=dict() imageDictAQ=dict() indexDictAQ=dict() thumbDictAQ=dict() zValueDictAQ=dict() | imageDict,indexDict,thumbDict,zValueDict = dict(),dict(),dict(),dict() imageDictAQ,indexDictAQ,thumbDictAQ,zValueDictAQ = dict(),dict(),dict(),dict() | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py |
imageDict[sngl.ifo]=list() indexDict[sngl.ifo]=list() thumbDict[sngl.ifo]=list() for myFile in fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ | indexDict[sngl.ifo],imageDict[sngl.ifo],thumbDict[sngl.ifo],zValueDict[sngl.ifo]=list(),list(),list(),list() indexDictAQ[sngl.ifo],imageDictAQ[sngl.ifo],thumbDictAQ[sngl.ifo],zValueDictAQ[sngl.ifo]=list(),list(),list(),list() for myFile in fnmatch.filter(filesOmega,\ | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py |
for myFile in fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ | for myFile in fnmatch.filter(filesOmega,\ | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py |
for myFile in fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ | for myFile in fnmatch.filter(filesOmega,\ | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py |
zValueFiles=fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ "*/%s_RDS_*/%s/*summary.txt"%(sngl.ifo,sngl.time)) | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py | |
if len(zValueFiles)>0: for zFile in zValueFiles: zValueDict[sngl.ifo].extend(wikiFileFinder.__readSummary__(zFile)) tmpList=list() for chan in zValueDict[sngl.ifo]: | for zFile in fnmatch.filter(filesOmega,\ "*/%s_RDS_*/%s/*summary.txt"%(sngl.ifo,sngl.time)): for chan in wikiFileFinder.__readSummary__(zFile): | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py |
tmpList.append(chan) zValueDict[sngl.ifo]=tmpList else: sys.stdout.write("Omega scan summary file not for for %s. ...skipping...\n"%sngl.ifo) | zValueDict[sngl.ifo].append(chan) if len(zValueDict[sngl.ifo]) == 0: sys.stdout.write("Omega scan summary file not or empty for %s. ...continuing...\n"%sngl.ifo) | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py |
imageDictAQ[sngl.ifo]=list() indexDictAQ[sngl.ifo]=list() thumbDictAQ[sngl.ifo]=list() | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py | |
for myFile in fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*html"%(sngl.ifo,timeString)): indexDictAQ[sngl.ifo].append(myFile) zValueFiles=fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*txt"%(sngl.ifo,timeString)) zValueDictAQ[sngl.ifo]=list() if len(zValueFiles)>0: for zFile in... | for myFile in fnmatch.filter(filesAnalyze,\ | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py |
for myFile in fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ | for myFile in fnmatch.filter(filesAnalyze,\ | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py |
for myFile in fnmatch.filter(filesAnalyze,\ "*%s-*_%s_*html"%(sngl.ifo,timeString)): indexDictAQ[sngl.ifo].append(myFile) zValueDictAQ[sngl.ifo]=list() for zFile in fnmatch.filter(filesAnalyze,\ "*%s-*_%s_*txt"%(sngl.ifo,timeString)): for chan in wikiFileFinder.__readSummary__(zFile): if not "PEM" in chan[0] or not "SE... | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.ex... | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py | |
sys.stdout.write("Available via browser at %s\n"%(mapFileURL.convert(myDestPath+"/"+myChecklistFilename))) | def __init__(self,type=None,ifo=None,time=None,snr=None,chisqr=None,mass1=None,mass2=None,mchirp=None): """ """ self.type=str(type) self.ifo=str(ifo) self.time=float(time) self.snr=float(snr) self.chisqr=float(chisqr) self.mass1=float(mass1) self.mass2=float(mass2) self.mchirp=float(mchirp) | 481b556f895e5b0b4caf0acaf35400d2993db8e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/481b556f895e5b0b4caf0acaf35400d2993db8e5/makeCheckListWiki.py | |
""" | """ | def skyhist_cart(skycarts,samples): """ Histogram the list of samples into bins defined by Cartesian vectors in skycarts """ dot=numpy.dot N=len(skycarts) print 'operating on %d sky points'%(N) bins=zeros(N) for sample in samples: sampcart=pol2cart(sample[RAdim],sample[decdim]) maxdx=max(xrange(0,N),key=lambda i:dot(sa... | 44fefd219acc9ff3c7b6c89412a08f1d90bf0eff /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/44fefd219acc9ff3c7b6c89412a08f1d90bf0eff/cbcBayesSkyRes.py |
i=list(nonzero(np.asarray(toppoints)[:,2]==injbin))[0] | i=list(np.nonzero(np.asarray(toppoints)[:,2]==injbin))[0] | def plotSkyMap(skypos,skyres,sky_injpoint): from pylal import skylocutils from mpl_toolkits.basemap import Basemap skypoints=array(skylocutils.gridsky(float(skyres))) skycarts=map(lambda s: pol2cart(s[1],s[0]),skypoints) skyinjectionconfidence=None shist=bayespputils.skyhist_cart(array(skycarts),skypos) #shist=skyhi... | 7c44282f4e4a859b6cb5db1816d07de6910a2141 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/7c44282f4e4a859b6cb5db1816d07de6910a2141/cbcBayesSkyRes.py |
self.cp.set("fu-condor","qscan",home_dirs()+"/rgouaty/opt/omega/omega_r3270_glnxa64_binary/bin/wpipeline") | self.cp.set("fu-condor","qscan",stfu_pipe.home_dirs()+"/rgouaty/opt/omega/omega_r3270_glnxa64_binary/bin/wpipeline") | def set_qscan_executable(self): host = stfu_pipe.get_hostname() if 'phy.syr.edu' in host: self.cp.set("fu-condor","qscan",home_dirs()+"/rgouaty/opt/omega/omega_r3270_glnxa64_binary/bin/wpipeline") else: self.cp.set("fu-condor","qscan",home_dirs()+"/romain/opt/omega/omega_r3270_glnxa64_binary/bin/wpipeline") | 73e6f0c892381f08ec6f8865c22409d910afc731 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/73e6f0c892381f08ec6f8865c22409d910afc731/WOD_Bologna.py |
self.cp.set("fu-condor","qscan",home_dirs()+"/romain/opt/omega/omega_r3270_glnxa64_binary/bin/wpipeline") | self.cp.set("fu-condor","qscan",stfu_pipe.home_dirs()+"/romain/opt/omega/omega_r3270_glnxa64_binary/bin/wpipeline") def get_cp(self): return self.cp def write(self): self.get_cp().write(open(self.ini_file,"w")) | def set_qscan_executable(self): host = stfu_pipe.get_hostname() if 'phy.syr.edu' in host: self.cp.set("fu-condor","qscan",home_dirs()+"/rgouaty/opt/omega/omega_r3270_glnxa64_binary/bin/wpipeline") else: self.cp.set("fu-condor","qscan",home_dirs()+"/romain/opt/omega/omega_r3270_glnxa64_binary/bin/wpipeline") | 73e6f0c892381f08ec6f8865c22409d910afc731 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/73e6f0c892381f08ec6f8865c22409d910afc731/WOD_Bologna.py |
parser.add_option("-f", "--config-file", default="followup_pipe.ini", help="the config file, default looks for stfu_pipe.ini in path, if none is found it makes one from your environment (only provide a config file if you know you must override something)") parser.add_option("-g", "--gps-times", default='', help="Specif... | parser.add_option("-f", "--config-file", default="WOD_Bologna.ini", help="the config file, default looks for stfu_pipe.ini in path, if none is found it makes one from your environment (only provide a config file if you know you must override something)") parser.add_option("-g", "--gps-times", default='', help="Specify ... | def parse_command_line(): parser = OptionParser( version = "%prog", description = "Pipeline to setup Remote Wscans On Demand" ) parser.add_option("-v", "--verbose", action = "store_true", help = "Be verbose.") parser.add_option("-f", "--config-file", default="followup_pipe.ini", help="the config file, default looks for... | 73e6f0c892381f08ec6f8865c22409d910afc731 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/73e6f0c892381f08ec6f8865c22409d910afc731/WOD_Bologna.py |
default_cp = stfu_pipe.create_default_config_wod(options.config_file) | default_cp = create_default_config_wod(options.config_file) | def parse_command_line(): parser = OptionParser( version = "%prog", description = "Pipeline to setup Remote Wscans On Demand" ) parser.add_option("-v", "--verbose", action = "store_true", help = "Be verbose.") parser.add_option("-f", "--config-file", default="followup_pipe.ini", help="the config file, default looks for... | 73e6f0c892381f08ec6f8865c22409d910afc731 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/73e6f0c892381f08ec6f8865c22409d910afc731/WOD_Bologna.py |
gpsevents = time_only_events(options.gps_times) | if options.gps_times: gpsevents = time_only_events(options.gps_times) elif options.input_file: gpsevents = extractTimesFromFile(options.input_file) else: print >> sys.stderr, "an argument is missing in the command:\n You need to use one of the options --gps-times or --input-file" sys.exit(1) | def parse_command_line(): parser = OptionParser( version = "%prog", description = "Pipeline to setup Remote Wscans On Demand" ) parser.add_option("-v", "--verbose", action = "store_true", help = "Be verbose.") parser.add_option("-f", "--config-file", default="followup_pipe.ini", help="the config file, default looks for... | 73e6f0c892381f08ec6f8865c22409d910afc731 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/73e6f0c892381f08ec6f8865c22409d910afc731/WOD_Bologna.py |
print >>sys.stderr, "following up %s @ %s" % (ifo, event.time) | print >>sys.stdout, "following up %s @ %s" % (ifo, event.time) | def parse_command_line(): parser = OptionParser( version = "%prog", description = "Pipeline to setup Remote Wscans On Demand" ) parser.add_option("-v", "--verbose", action = "store_true", help = "Be verbose.") parser.add_option("-f", "--config-file", default="followup_pipe.ini", help="the config file, default looks for... | 73e6f0c892381f08ec6f8865c22409d910afc731 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/73e6f0c892381f08ec6f8865c22409d910afc731/WOD_Bologna.py |
injected_cols.extend(['injected_end_time', 'injected_end_time_ns', 'injected_end_time_utc__Px_click_for_daily_ihope_xP_']) | injected_cols.extend(['injected_decisive_distance','injected_end_time', 'injected_end_time_ns', 'injected_end_time_utc__Px_click_for_daily_ihope_xP_']) | def convert_duration( duration ): return sqlutils.convert_duration( duration, convert_durations ) | 3052e11e800db376830bb9bfb701aad8dd50346c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/3052e11e800db376830bb9bfb701aad8dd50346c/printutils.py |
from glue.lal import LIGOTimeGPS | def get_pyvalue(self): return generic_get_pyvalue(self) | 3052e11e800db376830bb9bfb701aad8dd50346c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/3052e11e800db376830bb9bfb701aad8dd50346c/printutils.py | |
AND rank(""", decisive_distance, """) <= """, str(limit), """ | %s""" % (limit is not None and ''.join(['AND rank(', decisive_distance, ') <= ', str(limit)]) or ''), """ | def get_decisive_distance( *args ): return sorted(args)[1] | 3052e11e800db376830bb9bfb701aad8dd50346c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/3052e11e800db376830bb9bfb701aad8dd50346c/printutils.py |
dag.add_node(self) | if opts.do_makeCheckList: dag.add_node(self) | def __init__(self,dag,job,cp,opts): pipeline.CondorDAGNode.__init__(self,job) #Specify pipe location self.add_var_opt('followup-directory',cp.get("makeCheckListWiki", "location").strip()) #Specify pipe ini file self.add_var_opt('ini-file',cp.get("makeCheckListWiki", "ini-file").strip()) if not opts.disable_dag_categori... | 9f410b765542e07ca53b0567c39690f246423bb9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/9f410b765542e07ca53b0567c39690f246423bb9/stfu_pipe.py |
twoDGreedyCL['ra_sb,dec_sb']=skyinjectionconfidence if min_sky_area_containing_injection: twoDGreedyInj['ra_sb,dec_sb']=min_sky_area_containing_injection | twoDGreedyInj['ra_sb,dec_sb']={} twoDGreedyInj['ra_sb,dec_sb']['confidence']=min_sky_area_containing_injection if min_sky_area_containing_injection: twoDGreedyInj['ra_sb,dec_sb']['area']=min_sky_area_containing_injection | def cbcBayesSkyRes(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None,bayesfactornoise=None,bayesfactorcoherent=None): if eventnum is not None and injfile is None: print "You specified an event number but no injection file. Ignoring!" if data is None: prin... | de61da130b7db9e84280bb363fed90835f99068a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/de61da130b7db9e84280bb363fed90835f99068a/cbcBayesSkyRes.py |
while True: | while interval<len(np.asarray(toppoints)[:,3]): | def greedyBin1(par_samps,par_bin,confidence_levels,par_injvalue=None): oneDGreedyCL={} oneDGreedyInj={} parpos_min=min(par_samps) parpos_max=max(par_samps) par_point=parpos_min parpos_Nbins= int(ceil((parpos_max - parpos_min)/par_bin))+1 greedyPoints=np.zeros((parpos_Nbins,2)) #2D so it can be put through same con... | e548f7ac41a9a9e36ff61ebfa00c8afd13d3de17 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/e548f7ac41a9a9e36ff61ebfa00c8afd13d3de17/bayespputils.py |
galpt = fbins[sbin(fbins,inj_pt,fine_res)] | galpt = fbins[skylocutils.sbin(fbins,inj_pt,fine_res)] | def get_unique_filename(name): """ use this to avoid name collisions """ counter = 1 base_name, ext = os.path.splitext(name) while os.path.isfile(name): name = base_name + '_' + str(counter) + ext counter += 1 return name | 9292b77f771568d3044bb051ce6a4d89cf8a514b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/9292b77f771568d3044bb051ce6a4d89cf8a514b/run_skypoints.py |
fileListing.append(entry) | finalList.append(entry) | def __readCache__(self,cacheListing=list()): """ Simple mehtod to read in a cache or list of cache files and return a list of files or an empty list if nothing found. It uses the pathing information from the files passed via cacheListing to aid in our filesystem search. """ #Open the cache entry and search for those en... | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py |
def __readCache__(self,cacheListing=list()): """ Simple mehtod to read in a cache or list of cache files and return a list of files or an empty list if nothing found. It uses the pathing information from the files passed via cacheListing to aid in our filesystem search. """ #Open the cache entry and search for those en... | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py | ||
def get_findVetos(self): tmpList=list() #H1,H2,L1-findFlags_H1,H2,L1_831695156.714.wiki #instrument,ifos ifoString="" for i in range(0,len(self.coinc.ifos)/2):ifoString=ifoString+"%s,"%self.coinc.ifos[2*i:2*i+2] ifoString=ifoString.rstrip(",") insString="" for i in range(0,len(self.coinc.instruments)/2):insString=insSt... | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py | ||
myMask="*%s*%s-findVetos_%s_%s.wiki"%\ | myMask="*%s/*%s-findVetos_%s_%s.wiki"%\ | def get_findVetos(self): tmpList=list() #H1,H2,L1-findFlags_H1,H2,L1_831695156.714.wiki #instrument,ifos ifoString="" for i in range(0,len(self.coinc.ifos)/2):ifoString=ifoString+"%s,"%self.coinc.ifos[2*i:2*i+2] ifoString=ifoString.rstrip(",") insString="" for i in range(0,len(self.coinc.instruments)/2):insString=insSt... | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py |
myMask="*%s*%s-findFlags_%s_%s.wiki"%\ | myMask="*%s/*%s-findFlags_%s_%s.wiki"%\ | def get_findFlags(self): """ """ tmpList=list() #H1,H2,L1-findFlags_H1,H2,L1_831695156.714.wiki #instrument,ifos ifoString="" for i in range(0,len(self.coinc.ifos)/2):ifoString=ifoString+"%s,"%self.coinc.ifos[2*i:2*i+2] ifoString=ifoString.rstrip(",") insString="" for i in range(0,len(self.coinc.instruments)/2):insStri... | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py |
(self.coint.type,sngl.ifo,sngl.ifo,timeString) | (self.coinc.type,sngl.ifo,sngl.ifo,timeString) | def get_analyzeQscan_RDS(self): """ """ #analyseQscan.py_FG_RDS_full_data/H1-analyseQscan_H1_931176926_116_rds-unspecified-gpstime.cache cacheList=list() cacheFiles=list() for sngl in self.coinc.sngls: timeString=str(float(sngl.time)).replace(".","_") myCacheMask="*%s*/%s-analyseQscan_%s_%s_rds*.cache"%\ (self.coint.ty... | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py |
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # # Check to see if wiki file with name already exists # maxCount=0 while os.pat... | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py | ||
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # # Check to see if wiki file with name already exists # maxCount=0 while os.pat... | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py | ||
while os.path.exists(wikiFilename) and maxCount < 10: | while os.path.exists(wikiFilename) and maxCount < 15: | def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # # Check to see if wiki file with name already exists # maxCount=0 while os.pat... | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py |
sys.stdout.write("Found: %s\n",publication_directory) | sys.stdout.write("Found: %s\n"%publication_directory) | def __init__(self,type=None,ifo=None,time=None,snr=None,chisqr=None,mass1=None,mass2=None,mchirp=None): """ """ self.type=str(type) self.ifo=str(ifo) self.time=float(time) self.snr=float(snr) self.chisqr=float(chisqr) self.mass1=float(mass1) self.mass2=float(mass2) self.mchirp=float(mchirp) | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py |
sys.stdout.write("Found: %s\n",publication_url) | sys.stdout.write("Found: %s\n"%publication_url) | def __init__(self,type=None,ifo=None,time=None,snr=None,chisqr=None,mass1=None,mass2=None,mchirp=None): """ """ self.type=str(type) self.ifo=str(ifo) self.time=float(time) self.snr=float(snr) self.chisqr=float(chisqr) self.mass1=float(mass1) self.mass2=float(mass2) self.mchirp=float(mchirp) | d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/d3f37f3c912ccbf137aa3634e2e95a271c2cfd9f/makeCheckListWiki.py |
self.offset_vectors = offset_vectors | self.offset_vectors = list(offset_vectors) self.offset_vectors.sort(key = lambda offset_vector: sorted(offset_vector.items())) | def set_offset_vectors(self, offset_vectors): """ Set the list of offset vectors to be considered when deciding the bins in which each file belongs. Must be called before packing any files. The input is a list of dictionaries, each mapping instruments to offsets. """ self.offset_vectors = offset_vectors min_offset = ... | b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed/ligolw_cafe.py |
def set_offset_vectors(self, offset_vectors): """ Set the list of offset vectors to be considered when deciding the bins in which each file belongs. Must be called before packing any files. The input is a list of dictionaries, each mapping instruments to offsets. """ self.offset_vectors = offset_vectors min_offset = ... | b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed/ligolw_cafe.py | ||
def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ | def split_bins(cafepacker, extentlimit, verbose = False): """ Split bins in CafePacker so that each bin has an extent no longer than extentlimit. """ | def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ # # loop overall the bins in cafepacker.bins. we pop items out of # cafepacker.bins and append new ones to the end so need a while loop # checking the extent of each bin in cafep... | b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed/ligolw_cafe.py |
if abs(cafepacker.bins[idx].extent) <= extentlimit: | origbin = cafepacker.bins[idx] n = int(math.ceil(float(abs(origbin.extent)) / extentlimit)) if n <= 1: | def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ # # loop overall the bins in cafepacker.bins. we pop items out of # cafepacker.bins and append new ones to the end so need a while loop # checking the extent of each bin in cafep... | b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed/ligolw_cafe.py |
bigbin = cafepacker.bins.pop(idx) splittime = lsctables.LIGOTimeGPS(bigbin.extent[0] + (bigbin.extent[1] - bigbin.extent[0])/2) splitseglistdict = segments.segmentlistdict() for key in bigbin.size.keys(): splitseglistdict[key] = segments.segmentlist([segments.segment(-segments.infinity(),splittime)]) ... | splits = [-segments.infinity()] + [lsctables.LIGOTimeGPS(origbin.extent[0] + i * float(origbin.extent[1] - origbin.extent[0]) / n) for i in range(1, n)] + [+segments.infinity()] if verbose: print >>sys.stderr, "\tsplitting cache spanning %s at %s" % (str(origbin.extent), ", ".join(str(split) for split in splits[1:-... | def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ # # loop overall the bins in cafepacker.bins. we pop items out of # cafepacker.bins and append new ones to the end so need a while loop # checking the extent of each bin in cafep... | b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed/ligolw_cafe.py |
def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ # # loop overall the bins in cafepacker.bins. we pop items out of # cafepacker.bins and append new ones to the end so need a while loop # checking the extent of each bin in cafep... | b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed/ligolw_cafe.py | ||
bin1.size.offsets.update(offset_vector) bin2.size.offsets.update(offset_vector) thisseglistdict.offsets.update(offset_vector) if not coinc1 and bin1.size.is_coincident(thisseglistdict, keys = offset_vector.keys()): | if cache_entry.segment.protract(cafepacker.max_gap).disjoint(bin.extent): continue cache_entry_segs = cache_entry.to_segmentlistdict() for offset_vector in cafepacker.offset_vectors: cache_entry_segs.offsets.update(offset_vector) | def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ # # loop overall the bins in cafepacker.bins. we pop items out of # cafepacker.bins and append new ones to the end so need a while loop # checking the extent of each bin in cafep... | b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed/ligolw_cafe.py |
def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ # # loop overall the bins in cafepacker.bins. we pop items out of # cafepacker.bins and append new ones to the end so need a while loop # checking the extent of each bin in cafep... | b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed/ligolw_cafe.py | ||
coinc1 = 1 bin1.objects.append(cache) if not coinc2 and bin2.size.is_coincident(thisseglistdict, keys = offset_vector.keys()): coinc2 = 1 bin2.objects.append(cache) if coinc1 and coinc2: break thisseglistdict.offsets.clear() bin1.size.offsets.clear() bin2.size.offsets.clear() cafepacker.bins.ap... | if cache_entry_segs.intersects_segment(bin.extent): bin.objects.append(cache_entry) break cafepacker.bins[idx:idx+1] = newbins idx += len(newbins) | def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ # # loop overall the bins in cafepacker.bins. we pop items out of # cafepacker.bins and append new ones to the end so need a while loop # checking the extent of each bin in cafep... | b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed/ligolw_cafe.py |
pattern = "%%s%%0%dd.cache" % int(log10(len(bins)) + 1) | pattern = "%%s%%0%dd.cache" % int(math.log10(len(bins)) + 1) | def write_caches(base, bins, instruments, verbose = False): filenames = [] if len(bins): pattern = "%%s%%0%dd.cache" % int(log10(len(bins)) + 1) for n, bin in enumerate(bins): filename = pattern % (base, n) filenames.append(filename) if verbose: print >>sys.stderr, "writing %s ..." % filename f = file(filename, "w") fo... | b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b726e6b93ab3eca6e0cab29241e6a54f9ac3e9ed/ligolw_cafe.py |
raise ValueError r'param-name cannot have "\n","\t", "DROP", or "DELETE" in it' | raise ValueError, r'param-name cannot have "\n","\t", "DROP", or "DELETE" in it' | def __init__( self, table_name, table_param, param_ranges_opt, verbose = False ): """ Parse --param-ranges option. Creates self.param which is the table_name and the table_param appended together (with a '.') and self.param_ranges, which is a list of tuples that give the lower parameter value, whether it is an open or ... | 0fba21943e16a39333702aa510a6df64a01f91df /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/0fba21943e16a39333702aa510a6df64a01f91df/ligolw_sqlutils.py |
os.path.join("bin", "OddsPostProc.py"), | def run(self): # remove the automatically generated user env scripts for script in ["pylal-user-env.sh", "pylal-user-env.csh"]: log.info("removing " + script ) try: os.unlink(os.path.join("etc", script)) except: pass | 2b7075f5e9f2bb452d1f7a6d230ff2777e08c222 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/2b7075f5e9f2bb452d1f7a6d230ff2777e08c222/setup.py | |
def detector_thresholds(min_threshold, ifos, RA, dec, gps_time, sensitivities=None): | def detector_thresholds(ifos, RA, dec, gps_time, sensitivities=None, min_threshold=4.5, max_threshold=7.5): | def detector_thresholds(min_threshold, ifos, RA, dec, gps_time, sensitivities=None): """ Return a dictionary of sensitivity thresholds for each detector, based on a minimum threshold of min_threshold in the least sensitive one, for a source at position (RA,dec) specified in radians at time gps_time. Specifying a dictio... | 53a79e706267554633da0f19dc24a8fc82ee14b9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53a79e706267554633da0f19dc24a8fc82ee14b9/grbsummary.py |
return map(lambda (a,b,c): detector_thresholds(min_threshold,ifos,a,b,c,sensitivities), zip(RA,dec,gps_time)) | return map(lambda (a,b,c): detector_thresholds(ifos,a,b,c,sensitivities,min_threshold=min_threshold,max_threshold=max_threshold), zip(RA,dec,gps_time)) | def detector_thresholds(min_threshold, ifos, RA, dec, gps_time, sensitivities=None): """ Return a dictionary of sensitivity thresholds for each detector, based on a minimum threshold of min_threshold in the least sensitive one, for a source at position (RA,dec) specified in radians at time gps_time. Specifying a dictio... | 53a79e706267554633da0f19dc24a8fc82ee14b9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53a79e706267554633da0f19dc24a8fc82ee14b9/grbsummary.py |
sensitivies[det]=1.0 | sensitivities[det]=1.0 | def detector_thresholds(min_threshold, ifos, RA, dec, gps_time, sensitivities=None): """ Return a dictionary of sensitivity thresholds for each detector, based on a minimum threshold of min_threshold in the least sensitive one, for a source at position (RA,dec) specified in radians at time gps_time. Specifying a dictio... | 53a79e706267554633da0f19dc24a8fc82ee14b9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53a79e706267554633da0f19dc24a8fc82ee14b9/grbsummary.py |
assert len(ifos)==len(sensitivites) | assert len(ifos)==len(sensitivities) | def detector_thresholds(min_threshold, ifos, RA, dec, gps_time, sensitivities=None): """ Return a dictionary of sensitivity thresholds for each detector, based on a minimum threshold of min_threshold in the least sensitive one, for a source at position (RA,dec) specified in radians at time gps_time. Specifying a dictio... | 53a79e706267554633da0f19dc24a8fc82ee14b9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53a79e706267554633da0f19dc24a8fc82ee14b9/grbsummary.py |
resps[det]=antenna.response(gps_time,RA,dec,0,0,'radians',det)[2] | resps[det]=antenna.response(gps_time,RA,dec,0,0,'radians',det)[2]*sensitivities[det] | def detector_thresholds(min_threshold, ifos, RA, dec, gps_time, sensitivities=None): """ Return a dictionary of sensitivity thresholds for each detector, based on a minimum threshold of min_threshold in the least sensitive one, for a source at position (RA,dec) specified in radians at time gps_time. Specifying a dictio... | 53a79e706267554633da0f19dc24a8fc82ee14b9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53a79e706267554633da0f19dc24a8fc82ee14b9/grbsummary.py |
threshs[det]=min_threshold*(resps[det]/worst_resp)*sensitivities[det] | threshs[det]=min_threshold*(resps[det]/worst_resp) if threshs[det]>max_threshold: threshs[det]=max_threshold | def detector_thresholds(min_threshold, ifos, RA, dec, gps_time, sensitivities=None): """ Return a dictionary of sensitivity thresholds for each detector, based on a minimum threshold of min_threshold in the least sensitive one, for a source at position (RA,dec) specified in radians at time gps_time. Specifying a dictio... | 53a79e706267554633da0f19dc24a8fc82ee14b9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53a79e706267554633da0f19dc24a8fc82ee14b9/grbsummary.py |
if triple_coinc is True: for ifo_3 in ifos: if ifos.index(ifo_3)>ifos.index(ifo_2): triples.append(ifo_1+ifo_2+ifo_3) | for ifo_3 in ifos: if ifos.index(ifo_3)>ifos.index(ifo_2): triples.append(ifo_1+ifo_2+ifo_3) | def coinc_segments(start,end,ifos): #== first, construct doubles and triples lists doubles=[] triples=[] for ifo_1 in ifos: for ifo_2 in ifos: if ifos.index(ifo_2)>ifos.index(ifo_1): doubles.append(ifo_1+ifo_2) if triple_coinc is True: for ifo_3 in ifos: if ifos.index(ifo_3)>ifos.index(ifo_2): triples.append(ifo_1+ifo_... | 5b9508dadb38fe84da42a62a617834158cce6467 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5b9508dadb38fe84da42a62a617834158cce6467/dqSegmentUtils.py |
segments[ifo] = grab_segments(gps_start,\ gps_end,\ | segments[ifo] = grab_segments(start,\ end,\ | def coinc_segments(start,end,ifos): #== first, construct doubles and triples lists doubles=[] triples=[] for ifo_1 in ifos: for ifo_2 in ifos: if ifos.index(ifo_2)>ifos.index(ifo_1): doubles.append(ifo_1+ifo_2) if triple_coinc is True: for ifo_3 in ifos: if ifos.index(ifo_3)>ifos.index(ifo_2): triples.append(ifo_1+ifo_... | 5b9508dadb38fe84da42a62a617834158cce6467 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5b9508dadb38fe84da42a62a617834158cce6467/dqSegmentUtils.py |
maxdx=max(xrange(0,N),key=lambda i:dot(sampcart,skycarts[i])) | maxdot=0 for i in range(0,N): thisdot=dot(sampcart,skycarts[i]) if thisdot>maxdot: maxdot=thisdot maxdx=i | def skyhist_cart(skycarts,samples): """ Histogram the list of samples into bins defined by Cartesian vectors in skycarts """ dot=numpy.dot N=len(skycarts) print 'operating on %d sky points'%(N) bins=zeros(N) for sample in samples: sampcart=pol2cart(sample[RAdim],sample[decdim]) maxdx=max(xrange(0,N),key=lambda i:dot(sa... | 16deb0cf66e821e380f9fcad62275215a2a6c4ee /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/16deb0cf66e821e380f9fcad62275215a2a6c4ee/cbcBayesSkyRes.py |
from pylal import antenna | from pylal import antenna | def directional_horizon(ifos, RA, dec, gps_time, horizons=None): """ Return a dictionary of sensitivity numbers for each detector, based on a known sky location and an optional input dictionary of inspiral horizon distances for a reference source of the user's choice. If the horizons dictionary is specified, the return... | 71005aa7c5c0e51e741c51b585aa70ac798ca4f6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/71005aa7c5c0e51e741c51b585aa70ac798ca4f6/grbsummary.py |
threshs[det]=min_threshold*(horizons[det]/worst_horizon) | threshs[det]=min_threshold*(horizons[det]/worst_horizon) | def detector_thresholds(horizons,min_threshold,max_threshold=7.5): """ Return a set of detector thresholds adjusted for a particular set of inspiral horizon distances (calculated with directional_horizon). The min_threshold specified the minimum threshold which will be set for all detectors less sensitive than the best... | 71005aa7c5c0e51e741c51b585aa70ac798ca4f6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/71005aa7c5c0e51e741c51b585aa70ac798ca4f6/grbsummary.py |
segtest = Popen('FrCheck -i '+frame,shell=True,stdout=PIPE) if os.waitpid(segtest.pid,0)[1]==11: continue | segtest = subprocess.Popen([frcheck,"-i",frame],stdout=subprocess.PIPE) if os.waitpid(segtest.pid,0)[1]==11: print >>sys.stderr, "Warning. Segmentation fault detected with command:" print >>sys.stderr, "FrCheck -i "+frame continue | def grab_data(start,end,channel,type,\ nds=False,verbose=False,dmt=False): """ This function will return the frame data for the given channel of the given type in the given [start,end] time range and will construct a gps time vector to go with it. The nds option is not yet supported, and the dmt option will return data... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py |
ldf_exe,sourced = GetCommandOutput('which ligo_data_find') if ldf_exe.find('ligo_data_find')==-1: print "Cannot find ligo_data_find. Please ensure, lscsoft is sourced." sys.exit() | def generate_cache(start_time,end_time,ifos,types,return_files=False): """ This function will return a cache of files as found by ligo_data_find, given start and end time, and lists of ifos and types. If the return_files option is given as 'True' the function will return a list of frames with absolute paths, otherwise ... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py | |
data_find_cmd = ldf_exe.replace('\n','')+\ | data_find_cmd = ldf+\ | def generate_cache(start_time,end_time,ifos,types,return_files=False): """ This function will return a cache of files as found by ligo_data_find, given start and end time, and lists of ifos and types. If the return_files option is given as 'True' the function will return a list of frames with absolute paths, otherwise ... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py |
cache_out = Popen(data_find_cmd,shell=True,stdout=PIPE) | cache_out = subprocess.Popen(data_find_cmd,shell=True,\ stdout=subprocess.PIPE) | def generate_cache(start_time,end_time,ifos,types,return_files=False): """ This function will return a cache of files as found by ligo_data_find, given start and end time, and lists of ifos and types. If the return_files option is given as 'True' the function will return a list of frames with absolute paths, otherwise ... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py |
ldf_exe='ligo_data_find' ldf,ldf_status = GetCommandOutput('which '+ldf_exe) if ldf_status != 0: print >>sys.stderr, \ "Error: ligo_data_find not found. Please ensure lscsoftrc is sourced" sys.exit() | p = subprocess.Popen(["which", "ligo_data_find"], stdout=subprocess.PIPE) ldf = p.communicate()[0].replace('\n','') if p.returncode != 0: raise ValueError, "ligo_data_find" p.stdout.close() ldf = os.path.realpath(ldf) | def find_types(types,search='standard'): """ This function will return a valid list of LIGO frame types given the list of type strings. The search option defines the breadth of the search, to speed up the search, the following search options are supported: 'standard','short','full'. The 'R', 'T', and 'M' (raw, raw sec... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py |
find_cmd = ldf_exe+" -y | egrep " | find_cmd = ldf+" -y | egrep " | def find_types(types,search='standard'): """ This function will return a valid list of LIGO frame types given the list of type strings. The search option defines the breadth of the search, to speed up the search, the following search options are supported: 'standard','short','full'. The 'R', 'T', and 'M' (raw, raw sec... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py |
if find_cmd != ldf_exe+" -y | egrep '": found_types_out = Popen(find_cmd,shell=True,stdout=PIPE) | if find_cmd != ldf+" -y | egrep '": found_types_out = subprocess.Popen(find_cmd,shell=True,\ stdout=subprocess.PIPE) | def find_types(types,search='standard'): """ This function will return a valid list of LIGO frame types given the list of type strings. The search option defines the breadth of the search, to speed up the search, the following search options are supported: 'standard','short','full'. The 'R', 'T', and 'M' (raw, raw sec... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py |
ldf_exe='ligo_data_find' ldf_status = GetCommandOutput('which '+ldf_exe)[1] if ldf_status != 0: print >>sys.stderr, "Error: ligo_data_find not found. "+\ "Please ensure lscsoftrc is sourced" sys.exit() | p = subprocess.Popen(["which", "ligo_data_find"], stdout=subprocess.PIPE) ldf = p.communicate()[0].replace('\n','') if p.returncode != 0: raise ValueError, "ligo_data_find" p.stdout.close() ldf = os.path.realpath(ldf) | def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py |
frame_cmd = ldf_exe+''' --observatory '''+ifo[0:1]+\ | frame_cmd = ldf+''' --observatory '''+ifo[0:1]+\ | def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py |
frame_out = Popen(frame_cmd,shell=True,stdout=PIPE,stderr=PIPE) | frame_out = subprocess.Popen(frame_cmd,shell=True,stdout=subprocess.PIPE,\ stderr=subprocess.PIPE) | def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py |
channel_list_out = Popen(channel_find_cmd,shell=True,stdout=PIPE) | channel_list_out = subprocess.Popen(channel_find_cmd,shell=True,\ stdout=subprocess.PIPE) | def find_channels(channels=None,\ types=None,\ ifos=None,\ ex_channels=None,\ ignore=[],\ match=False,\ time=None,\ unique=False,\ verbose=False): """ This function will use FrChannels to return all LIGO data channels matching the given list of 'channels' strings, whilst exluding the 'ex_channels' strings. Using find_... | 25494bc1b1cc42ac72937fd2b2d61f6c554aadd6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25494bc1b1cc42ac72937fd2b2d61f6c554aadd6/dqFrameUtils.py |
for (frac,size) in skyreses: htmlfile.write('<tr><td>%f</td>%f</td></tr>'%(frac,size)) | for (frac,skysize) in skyreses: htmlfile.write('<tr><td>%f</td>%f</td></tr>'%(frac,skysize)) | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | c14ba4df2f17594fac21d95132f9de75c1b20c93 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/c14ba4df2f17594fac21d95132f9de75c1b20c93/OddsPostProc.py |
connection.commit() | database.connection.commit() | def get_likelihood_ratio(coinc_event_id, time_slide_id, row_from_cols = database.sngl_burst_table.row_from_cols, cursor = database.connection.cursor(), offset_vectors = offset_vectors, params_func = params_func, params_func_extra_args = params_func_extra_args): events = map(row_from_cols, cursor.execute(""" | 51d60c1566dbbe2ac5a81fe88baa8e8f96a17239 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/51d60c1566dbbe2ac5a81fe88baa8e8f96a17239/ligolw_burca2.py |
sngl_burst.* time_slide.offset, | sngl_burst.*, time_slide.offset | def add_noninjections(self, param_func, database, *args): # iterate over burst<-->burst coincs cursor = database.connection.cursor() for coinc_event_id, time_slide_id in database.connection.cursor().execute(""" | fbb01c74546b6b69d24a1f8d4c68e6cdf929abaa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/fbb01c74546b6b69d24a1f8d4c68e6cdf929abaa/ligolw_burca_tailor.py |
self.veto_definer = os.path.basename(definer_file) | def update_veto_lists(self, timeoffset, veto_definer = None): | 320d8cf2b8034ac9ade2033bc56c27b4c7711a7d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/320d8cf2b8034ac9ade2033bc56c27b4c7711a7d/pylal_exttrig_llutils.py | |
injectionconfidence=0 | def loadDataFile(filename): print filename infile=open(filename,'r') formatstr=infile.readline().lstrip() header=formatstr.split() llines=[] import re dec=re.compile(r'[^\d.-]+') for line in infile: sline=line.split() proceed=True if len(sline)<1: print 'Ignoring empty line in input file: %s'%(sline) proceed=False for ... | 4bd45aa67dea28fca44510afdf1f7ed78973a6e1 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/4bd45aa67dea28fca44510afdf1f7ed78973a6e1/cbcBayesSkyRes.py | |
if injectionconfidence: | if injectionconfidence!=0: | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | 4bd45aa67dea28fca44510afdf1f7ed78973a6e1 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/4bd45aa67dea28fca44510afdf1f7ed78973a6e1/cbcBayesSkyRes.py |
self.generations[n] = tuple(TimeSlideGraphNode(offset_vector) for offset_vector in ligolw_tisi.time_slide_component_vectors((node.offset_vector for node in self.head), n)) | self.generations[n] = tuple(TimeSlideGraphNode(offset_vector) for offset_vector in ligolw_tisi.time_slide_component_vectors((node.offset_vector for node in self.head if len(node.offset_vector) == n), n)) | def __init__(self, offset_vector_dict, verbose = False): if verbose: print >>sys.stderr, "constructing coincidence assembly graph for %d target offset vectors ..." % len(offset_vector_dict) | aadef6db4e4e8b776ea67897539cb82959742b0f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/aadef6db4e4e8b776ea67897539cb82959742b0f/snglcoinc.py |
for component in node.components: component_name = vectorstring(component.offset_vector) print >>fileobj, "\t\"%s\" -> \"%s\";" % (component_name, node_name) | if node.components is not None: for component in node.components: print >>fileobj, "\t\"%s\" -> \"%s\";" % (vectorstring(component.offset_vector), node_name) | def write(self, fileobj): """ Write a DOT graph representation of the time slide graph to fileobj. """ vectorstring = lambda offset_vector: ",".join("%s=%g" % (instrument, offset) for instrument, offset in sorted(offset_vector.items())) | aadef6db4e4e8b776ea67897539cb82959742b0f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/aadef6db4e4e8b776ea67897539cb82959742b0f/snglcoinc.py |
component_name = vectorstring(component.offset_vector) print >>fileobj, "\t\"%s\" -> \"%s\";" % (component_name, node_name) | print >>fileobj, "\t\"%s\" -> \"%s\";" % (vectorstring(component.offset_vector), node_name) | def write(self, fileobj): """ Write a DOT graph representation of the time slide graph to fileobj. """ vectorstring = lambda offset_vector: ",".join("%s=%g" % (instrument, offset) for instrument, offset in sorted(offset_vector.items())) | aadef6db4e4e8b776ea67897539cb82959742b0f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/aadef6db4e4e8b776ea67897539cb82959742b0f/snglcoinc.py |
html_filename = prefix + opts.suffix +".html" | html_filename = prefix + opts.suffix +".html" | def write_html_output(opts, args, fnameList, tagLists, \ doThumb=True, mapList = [],\ comment=None, CoincSummTable=None,\ html_tag = '', add_box_flag=False): """ @param opts: The options from the calling code @param args: The args from the calling code @param fnameList: A list of the filenames @param tagLists: A list f... | 08b9a01049dc1b59f27566bdf04494b45b5e456a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/08b9a01049dc1b59f27566bdf04494b45b5e456a/InspiralUtils.py |
ds = pi*sqrt(2.0)*resolution/180.0 | ds = pi*resolution/180.0 | def gridsky(resolution): """ grid the sky up into roughly square regions resolution is the length of a side the points get placed at the center of the squares and to first order each square has an area of resolution^2 """ latitude = 0.0 longitude = pi ds = pi*sqrt(2.0)*resolution/180.0 points = [(latitude-0.5*pi, longi... | a628ae45f8774998ed97e41a7a92f65e96084710 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/a628ae45f8774998ed97e41a7a92f65e96084710/skylocutils.py |
fgtemp = finegrid | fgtemp = finegrid[:] | def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid coarsedict = {} ds = coarseres*pi/180 for cpt in coarsegrid: f... | a628ae45f8774998ed97e41a7a92f65e96084710 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/a628ae45f8774998ed97e41a7a92f65e96084710/skylocutils.py |
ds = coarseres*pi/180 | ds = coarseres*pi/180.0 | def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid coarsedict = {} ds = coarseres*pi/180 for cpt in coarsegrid: f... | a628ae45f8774998ed97e41a7a92f65e96084710 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/a628ae45f8774998ed97e41a7a92f65e96084710/skylocutils.py |
if (cpt[0]-fpt[0])*(cpt[0]-fpt[0]) <= ds*ds/4 and \ (cpt[1]-fpt[1])*(cpt[1]-fpt[1])*abs(sin(fpt[1]))*abs(sin(fpt[1])) \ <= ds*ds/4: | if (cpt[0]-fpt[0])*(cpt[0]-fpt[0]) <= ds*ds/4.0 and \ (cpt[1]-fpt[1])*(cpt[1]-fpt[1])*sin(cpt[0])*sin(cpt[0]) \ <= ds*ds/4.0: | def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid coarsedict = {} ds = coarseres*pi/180 for cpt in coarsegrid: f... | a628ae45f8774998ed97e41a7a92f65e96084710 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/a628ae45f8774998ed97e41a7a92f65e96084710/skylocutils.py |
html_tcmp_write+='<td width="30%"><img width="100%" src="2Dkde/'+twoDKdePath+'"/></td>' | html_tcmp_write+='<td width="30%"><img width="100%" src="2Dkde/'+figname+'"/></td>' | def cbcBayesPostProc(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None,bayesfactornoise=None,bayesfactorcoherent=None): """ This is a demonstration script for using the functionality/data structures contained in pylal.bayespputils . It will produce a webpag... | bb781a6a74081f7f519b038ddb9a918ada76ebf3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/bb781a6a74081f7f519b038ddb9a918ada76ebf3/cbcBayesPostProc.py |
a minimum threshold of min_threshold in the most sensitive one, for a source | a minimum threshold of min_threshold in the least sensitive one, for a source | def detector_thresholds(min_threshold, ifos, RA, dec, gps_time, sensitivities=None): """ Return a dictionary of sensitivity thresholds for each detector, based on a minimum threshold of min_threshold in the most sensitive one, for a source at position (RA,dec) specified in radians at time gps_time. Specifying a diction... | 1dd322a12b3f3453f045cd18ced566de10cc30ed /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1dd322a12b3f3453f045cd18ced566de10cc30ed/grbsummary.py |
for name, value in params_func(events, offsetdict, *params_func_extra_args).items(): | for name, value in sorted(params_func(events, offsetdict, *params_func_extra_args).items()): | def P(self, params_func, events, offsetdict, *params_func_extra_args): P_bak = 1.0 P_inj = 1.0 for name, value in params_func(events, offsetdict, *params_func_extra_args).items(): P_bak *= self.background_rates[name](*value)[0] P_inj *= self.injection_rates[name](*value)[0] return P_bak, P_inj | 0728285b67582061ae9df4772f8661c0504c1e4c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/0728285b67582061ae9df4772f8661c0504c1e4c/ligolw_burca2.py |
scatter(plx,ply,s=5,c=numpyasarray(toppoints)[:,2],faceted=False,cmap=matplotlib.cm.jet) | scatter(plx,ply,s=5,c=numpy.asarray(toppoints)[:,2],faceted=False,cmap=matplotlib.cm.jet) | def getinjpar(inj,parnum): if parnum==0: return inj.mchirp if parnum==1: return inj.eta if parnum==2: return inj.get_end() if parnum==3: return inj.phi0 if parnum==4: return inj.distance if parnum==5: return inj.longitude if parnum==6: return inj.latitude if parnum==7: return inj.polarization if parnum==8: return inj.i... | 32555dad1d9ba92ea918455a834f633f6eb55f1b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/32555dad1d9ba92ea918455a834f633f6eb55f1b/OddsPostProc.py |
self.scan_type = type | self.scan_type = type.replace("seismic","seis").upper() | def __init__(self, dag, job, cp, opts, ifo, p_nodes=[], type=""): | 17a68b2f4ab34a52fd284af25bc8fc18d36c784c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/17a68b2f4ab34a52fd284af25bc8fc18d36c784c/stfu_pipe.py |
seps=map(lambda s: ang_dist(sample[6],sample[7],s[1],s[0]),skypoints) | seps=map(lambda s: ang_dist(sample[5],sample[6],s[1],s[0]),skypoints) | def sky_hist(skypoints,samples): N=len(skypoints) print 'operating on %d sky points' % (N) bins=zeros(N) j=0 for sample in samples: seps=map(lambda s: ang_dist(sample[6],sample[7],s[1],s[0]),skypoints) minsep=math.pi for i in range(0,N): if seps[i]<minsep: minsep=seps[i] mindx=i bins[mindx]=bins[mindx]+1 j=j+1 print 'D... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
sampcart=pol2cart(sample[6],sample[7]) | sampcart=pol2cart(sample[5],sample[6]) | def skyhist_cart(skycarts,samples): N=len(skypoints) print 'operating on %d sky points'%(N) bins=zeros(N) j=0 for sample in samples: sampcart=pol2cart(sample[6],sample[7]) dots=map(lambda s: numpy.dot(sampcart,s),skycarts) maxdot=0 for i in range(0,N): if dots[i]>maxdot: maxdot=dots[i] mindx=i bins[mindx]=bins[mindx]+1... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
plot(getinjpar(injection,0),getinjpar(injection,1),'go') | plot(getinjpar(injection,0),getinjpar(injection,1),'go',scalex=False,scaley=False) | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
plot(getinjpar(injection,5),getinjpar(injection,6),'go') | plot(getinjpar(injection,5),getinjpar(injection,6),'go',scalex=False,scaley=False) | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
if injection and getinjpar(injection,7)<max(pos[:,7]) and getinjpar(injection,7)>min(pos[:,7]) and getinjpar(injection,8)<max(pos[:,8]) and getinjpar(injection,8)>min(pos[:,8]): plot(getinjpar(injection,7),getinjpar(injection,8),'go') | if injection and getinjpar(injection,7)<max(pos[:,7]) and getinjpar(injection,7)>min(pos[:,7]) and getinjpar(injection,8)<max(pos[:,8]) and getinjpar(injection,8)>min(pos[:,8]): plot(getinjpar(injection,7),getinjpar(injection,8),'go',scalex=False,scaley=False) | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
plot(injection.mass1,injection.mass2,'go') | plot(injection.mass1,injection.mass2,'go',scalex=False,scaley=False) | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.