rem stringlengths 1 322k | add stringlengths 0 2.05M | context stringlengths 4 228k | meta stringlengths 156 215 |
|---|---|---|---|
plot(injection.mass1,injection.distance,'go') | plot(injection.mass1,injection.distance,'go',scalex=False,scaley=False) | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
plot(getinjpar(injection,4),getinjpar(injection,8),'go') | plot(getinjpar(injection,4),getinjpar(injection,8),'go',scalex=False,scaley=False) | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
plot(getinjpar(injection,i),getinjpar(injection,j),'go') | plot(getinjpar(injection,i),getinjpar(injection,j),'go',scalex=False,scaley=False) | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
plot([getinjpar(injection,i),getinjpar(injection,i)],[0,max(kdepdf)],'r-.') | plot([getinjpar(injection,i),getinjpar(injection,i)],[0,max(kdepdf)],'r-.',scalex=False,scaley=False) | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
htmlfile.write('<img src="'+paramnames[i]+'.png">') | myfig=figure(figsize=(4,3.5),dpi=80) plot(pos[:,i],'.') if injection and min(pos[:,i])<getinjpar(injection,i) and max(pos[:,i])>getinjpar(injection,i): plot([0,len(pos)],[getinjpar(injection,i),getinjpar(injection,i)],'r-.') myfig.savefig(outdir+'/'+paramnames[i]+'_samps.png') htmlfile.write('<img src="'+paramnames[i]+... | def plot2Dkernel(xdat,ydat,Nx,Ny): xax=linspace(min(xdat),max(xdat),Nx) yax=linspace(min(ydat),max(ydat),Ny) x,y=numpy.meshgrid(xax,yax) samp=array([xdat,ydat]) kde=stats.kde.gaussian_kde(samp) grid_coords = numpy.append(x.reshape(-1,1),y.reshape(-1,1),axis=1) z = kde(grid_coords.T) z = z.reshape(Nx,Ny) asp=xax.ptp()/y... | 1c7e777c94d676d98831f95002598d58404c4fe8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/1c7e777c94d676d98831f95002598d58404c4fe8/OddsPostProc.py |
["src/xlal/date.c"], | ["src/xlal/date.c", "src/xlal/misc.c"], | def run(self): # remove the automatically generated user env scripts for script in ["pylal-user-env.sh", "pylal-user-env.csh"]: log.info("removing " + script ) try: os.unlink(os.path.join("etc", script)) except: pass | be80e194ad3e1beb2ce4d17921d438d75f3e5bfd /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/be80e194ad3e1beb2ce4d17921d438d75f3e5bfd/setup.py |
injected_cols.extend(['injected_end_time', 'injected_end_time_ns', 'injected_end_time_utc__Px_click_for_daily_ihope_xP_']) | injected_cols.extend(['injected_decisive_distance','injected_end_time', 'injected_end_time_ns', 'injected_end_time_utc__Px_click_for_daily_ihope_xP_']) | def convert_duration( duration ): return sqlutils.convert_duration( duration, convert_durations ) | 66c62ec79b0aba296ba768048392060c9834f293 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/66c62ec79b0aba296ba768048392060c9834f293/printutils.py |
from glue.lal import LIGOTimeGPS | def get_pyvalue(self): return generic_get_pyvalue(self) | 66c62ec79b0aba296ba768048392060c9834f293 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/66c62ec79b0aba296ba768048392060c9834f293/printutils.py | |
AND rank(""", decisive_distance, """) <= """, str(limit), """ | %s""" % (limit is not None and ''.join(['AND rank(', decisive_distance, ') <= ', str(limit)]) or ''), """ | def get_decisive_distance( *args ): return sorted(args)[1] | 66c62ec79b0aba296ba768048392060c9834f293 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/66c62ec79b0aba296ba768048392060c9834f293/printutils.py |
y.segment_def_cdb = x.creator_db) AND \ NOT (segment.start_time > %s OR %s > segment.end_time) \ | y.segment_def_cdb = x.creator_db) \ | def getSciSegs(ifo=None, gpsStart=None, gpsStop=None, cut=bool(False), serverURL=None, segName="DMT-SCIENCE", seglenmin=None, segpading=0 | 66b3b6bdd1f5fd4c09d95bfa909886e3a8002dc7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/66b3b6bdd1f5fd4c09d95bfa909886e3a8002dc7/fu_utils.py |
sqlQuery=query01%(segName,ifo,gpsStop,gpsStart,gpsStop,gpsStart) | sqlQuery=query01%(segName,ifo,gpsStop,gpsStart) | def getSciSegs(ifo=None, gpsStart=None, gpsStop=None, cut=bool(False), serverURL=None, segName="DMT-SCIENCE", seglenmin=None, segpading=0 | 66b3b6bdd1f5fd4c09d95bfa909886e3a8002dc7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/66b3b6bdd1f5fd4c09d95bfa909886e3a8002dc7/fu_utils.py |
AND y.segment_def_cdb = x.creator_db \ | AND y.segment_def_cdb = x.creator_db \ | def __init__(self,LDBDServerURL=None,quiet=bool(False),pickle=None,blinded=False): """ This class setups of for connecting to a LDBD server specified at command line to do segment queries as part of the follow up pipeline. If the user does not specify the LDBD server to use the method will use the environment variable... | 66b3b6bdd1f5fd4c09d95bfa909886e3a8002dc7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/66b3b6bdd1f5fd4c09d95bfa909886e3a8002dc7/fu_utils.py |
shiftString=shift | shiftString=shiftLabel[shift] | def getFOMLinks(gpsTime=int(0),ifo=("default")): """ Simple method returns a list of links to FOMs ordered by FOM # The list is 2D ie: [['ifo,shift',LINKtoImage,LinktoThumb],['ifo,shift',LinktoImage,LinkToThumb]...] images marked [Eve,Owl,Day] via [p3,p2,p1] in filenames this methd only for S6 and later IFO naming star... | 6cd44b04b081138f0ba71fef8fcfaf1fc7a38e27 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/6cd44b04b081138f0ba71fef8fcfaf1fc7a38e27/stfu_pipe.py |
zero_lag_time_slides = {} background_time_slides = {} for id, instrument, offset, is_background in connection.cursor().execute(""" | offset_vectors = {} for id, instrument, offset in connection.cursor().execute(""" | def get_time_slides(connection): """ Query the database for the IDs and offsets of all time slides, and return two dictionaries one containing the all-zero time slides and the other containing the not-all-zero time slides. """ zero_lag_time_slides = {} background_time_slides = {} for id, instrument, offset, is_backgrou... | 4b4b5780a156863532be0b9f80623ca9e7f9af33 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/4b4b5780a156863532be0b9f80623ca9e7f9af33/SnglBurstUtils.py |
offset, EXISTS ( SELECT * FROM time_slide AS a WHERE a.time_slide_id == time_slide.time_slide_id AND a.offset != 0 ) | offset | def get_time_slides(connection): """ Query the database for the IDs and offsets of all time slides, and return two dictionaries one containing the all-zero time slides and the other containing the not-all-zero time slides. """ zero_lag_time_slides = {} background_time_slides = {} for id, instrument, offset, is_backgrou... | 4b4b5780a156863532be0b9f80623ca9e7f9af33 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/4b4b5780a156863532be0b9f80623ca9e7f9af33/SnglBurstUtils.py |
if is_background: if id not in background_time_slides: background_time_slides[id] = {} background_time_slides[id][instrument] = offset else: if id not in zero_lag_time_slides: zero_lag_time_slides[id] = {} zero_lag_time_slides[id][instrument] = offset | if id not in offset_vectors: offset_vectors[id] = {} offset_vectors[id][instrument] = offset zero_lag_time_slides = dict((id, offset_vector) for id, offset_vector in offset_vectors.items() if not any(offset_vector.values())) background_time_slides = dict((id, offset_vector) for id, offset_vector in offset_vectors.items... | def get_time_slides(connection): """ Query the database for the IDs and offsets of all time slides, and return two dictionaries one containing the all-zero time slides and the other containing the not-all-zero time slides. """ zero_lag_time_slides = {} background_time_slides = {} for id, instrument, offset, is_backgrou... | 4b4b5780a156863532be0b9f80623ca9e7f9af33 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/4b4b5780a156863532be0b9f80623ca9e7f9af33/SnglBurstUtils.py |
return r"%s \\times 10^{%d}" % (m, int(e)) | return r"%s \times 10^{%d}" % (m, int(e)) | def latexnumber(s): """ Convert a string of the form "d.dddde-dd" to "d.dddd \times 10^{-dd}" """ m, e = floatpattern.match(s).groups() return r"%s \\times 10^{%d}" % (m, int(e)) | 4b4b5780a156863532be0b9f80623ca9e7f9af33 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/4b4b5780a156863532be0b9f80623ca9e7f9af33/SnglBurstUtils.py |
rate.to_moving_mean_density(binnedarray, filters.get(name, default_filter)) | def finish(self, filters = {}, verbose = False): default_filter = rate.gaussian_window(21) # normalizing each array so that its sum is 1 has the # effect of making the integral of P(x) dx equal 1 after # the array is transformed to an array of densities (which # is done by dividing each bin by dx). N = len(self.zero_la... | b2c43f402b25fddd2b8a6094c36b902be73280ae /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b2c43f402b25fddd2b8a6094c36b902be73280ae/ligolw_burca_tailor.py | |
twoDKdePath=os.path.join(margdir,par1_name+'-'+par2_name+'_2Dkernel.png') | figname=par1_name+'-'+par2_name+'_2Dkernel.png' twoDKdePath=os.path.join(margdir,figname) | def cbcBayesPostProc(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None,bayesfactornoise=None,bayesfactorcoherent=None): """ This is a demonstration script for using the functionality/data structures contained in pylal.bayespputils . It will produce a webpag... | 25081af9755dc3b90ae7c9418d3526e51860c233 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25081af9755dc3b90ae7c9418d3526e51860c233/cbcBayesPostProc.py |
html_tcmp_write+='<td width="30%"><img width="100%" src="'+twoDKdePath+'"/></td>' | html_tcmp_write+='<td width="30%"><img width="100%" src="2Dkde/'+twoDKdePath+'"/></td>' | def cbcBayesPostProc(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None,bayesfactornoise=None,bayesfactorcoherent=None): """ This is a demonstration script for using the functionality/data structures contained in pylal.bayespputils . It will produce a webpag... | 25081af9755dc3b90ae7c9418d3526e51860c233 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25081af9755dc3b90ae7c9418d3526e51860c233/cbcBayesPostProc.py |
html_tcmp.a("2D/",'All 2D Marginal PDFs') | html_tcmp.a("2Dkde/",'All 2D marginal PDFs (kde)') | def cbcBayesPostProc(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None,bayesfactornoise=None,bayesfactorcoherent=None): """ This is a demonstration script for using the functionality/data structures contained in pylal.bayespputils . It will produce a webpag... | 25081af9755dc3b90ae7c9418d3526e51860c233 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/25081af9755dc3b90ae7c9418d3526e51860c233/cbcBayesPostProc.py |
inspiral.InspiralNode.__init__(self,job) | pipeline.CondorDAGNode.__init__(self,job) pipeline.AnalysisNode.__init__(self) | def __init__(self, dag, job, cp, opts, sngl, frame_cache, chia, tag, p_nodes=[]): | 8397d083fc2a94fb7188f06b24add6fa5dcb7930 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/8397d083fc2a94fb7188f06b24add6fa5dcb7930/stfu_pipe.py |
self.add_var_opt(param,value) | def __init__(self, dag, job, cp, opts, sngl, frame_cache, chia, tag, p_nodes=[]): | 8397d083fc2a94fb7188f06b24add6fa5dcb7930 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/8397d083fc2a94fb7188f06b24add6fa5dcb7930/stfu_pipe.py | |
inspiral.ChiaNode.__init__(self,job) | pipeline.CondorDAGNode.__init__(self,job) pipeline.AnalysisNode.__init__(self) | def __init__(self, dag, job, cp, opts, coinc, inspiral_node_dict, chia_node =None, p_nodes = []): | 8397d083fc2a94fb7188f06b24add6fa5dcb7930 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/8397d083fc2a94fb7188f06b24add6fa5dcb7930/stfu_pipe.py |
f = open(opts.galaxy_priors_dir+'/'+str(int(mineffD))+'Mpc.pkl','r') | f = open(opts.galaxy_priors_dir+'/galaxy_prior_'+str(int(mineffD))+'Mpc.pkl','r') | def get_unique_filename(name): """ use this to avoid name collisions """ counter = 1 base_name, ext = os.path.splitext(name) while os.path.isfile(name): name = base_name + '_' + str(counter) + ext counter += 1 return name | 5d57777704fdde7db5dca49e7d89c4111f8d4dc7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5d57777704fdde7db5dca49e7d89c4111f8d4dc7/run_skypoints.py |
coarsedict = {} | coarsedict = {} | def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarseg... | 18bcbad7a0f00300278669b07bffc33c996bd96d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/18bcbad7a0f00300278669b07bffc33c996bd96d/skylocutils.py |
if (cpt[0]-fpt[0])*(cpt[0]-fpt[0]) <= ds*ds/4.0 and \ (cpt[1]-fpt[1])*(cpt[1]-fpt[1])*sin(cpt[0])*sin(cpt[0]) <= ds*ds/4.0: | if (cpt[0]-fpt[0])*(cpt[0]-fpt[0]) - ds*ds/4.0 <= epsilon and \ (cpt[1]-fpt[1])*(cpt[1]-fpt[1])*sin(cpt[0])*sin(cpt[0]) \ - ds*ds/4.0 <= epsilon: | def map_grids(coarsegrid,finegrid,coarseres=4.0): """ takes the two grids (lists of lat/lon tuples) and returns a dictionary where the points in the coarse grid are the keys and lists of tuples of points in the fine grid are the values """ fgtemp = finegrid[:] coarsedict = {} ds = coarseres*pi/180.0 for cpt in coarseg... | 18bcbad7a0f00300278669b07bffc33c996bd96d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/18bcbad7a0f00300278669b07bffc33c996bd96d/skylocutils.py |
mv $1 $2/. tar -xzvf $2/$1 | currentPath=`pwd` ; mv $1 $2/. ; cd $2 ; tar -xzvf $1 ; cd $currentPath ; | def setup_distrib_script(self,tag_base): distrib_script = open('distribRemoteScan_'+tag_base+'.sh','w') distrib_script.write("""#!/bin/bash | 154fa74bf8e4d8677cc922914fbb92d97e85216c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/154fa74bf8e4d8677cc922914fbb92d97e85216c/stfu_pipe.py |
rm $2/$1 | rm $2/$1 ; | def setup_distrib_script(self,tag_base): distrib_script = open('distribRemoteScan_'+tag_base+'.sh','w') distrib_script.write("""#!/bin/bash | 154fa74bf8e4d8677cc922914fbb92d97e85216c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/154fa74bf8e4d8677cc922914fbb92d97e85216c/stfu_pipe.py |
raise NotImplemented | raise NotImplementedError | def add_content(self, data, label="_nolabel_"): """ Stub. Replace with a method that appends values or lists of values to self.data_sets and appends labels to self.data_labels. Feel free to accept complicated inputs, but try to store only the raw numbers that will enter the plot. """ raise NotImplemented | 26a99ad4b2deb087e7d6b8eba6399c04c01d95b4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/26a99ad4b2deb087e7d6b8eba6399c04c01d95b4/plotutils.py |
raise NotImplemented | raise NotImplementedError | def finalize(self): """ Stub. Replace with a function that creates and makes your plot pretty. Do not do I/O here. """ raise NotImplemented | 26a99ad4b2deb087e7d6b8eba6399c04c01d95b4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/26a99ad4b2deb087e7d6b8eba6399c04c01d95b4/plotutils.py |
for sngl in coincEvent.sngl_inspiral.itervalues(): myArgString=myArgString+"%s,"%sngl.ifo | if hasattr(coincEvent, "sngl_inspiral"): for sngl in coincEvent.sngl_inspiral.itervalues(): myArgString=myArgString+"%s,"%sngl.ifo elif hasattr(coincEvent, "ifos_list"): for ifo in coincEvent.ifos_list: myArgString=myArgString+"%s,"%ifo | def __init__(self, dag, job, cp, opts, coincEvent=None): """ """ self.__conditionalLoadDefaults__(findFlagsNode.defaults,cp) pipeline.CondorDAGNode.__init__(self,job) self.add_var_opt("trigger-time",coincEvent.time) #Output filename oFilename="%s-findFlags_%s_%s.wiki"%(coincEvent.instruments, coincEvent.ifos, coincEven... | 56395bb647ab92635e1398364b06f48bf65527b3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/56395bb647ab92635e1398364b06f48bf65527b3/stfu_pipe.py |
for sngl in coincEvent.sngl_inspiral.itervalues(): myArgString=myArgString+"%s,"%sngl.ifo | if hasattr(coincEvent, "sngl_inspiral"): for sngl in coincEvent.sngl_inspiral.itervalues(): myArgString=myArgString+"%s,"%sngl.ifo elif hasattr(coincEvent, "ifos_list"): for ifo in coincEvent.ifos_list: myArgString=myArgString+"%s,"%ifo | def __init__(self, dag, job, cp, opts, coincEvent=None): """ """ self.__conditionalLoadDefaults__(findVetosNode.defaults,cp) pipeline.CondorDAGNode.__init__(self,job) self.add_var_opt("trigger-time",coincEvent.time) #Output filename oFilename="%s-findVetos_%s_%s.wiki"%(coincEvent.instruments, coincEvent.ifos, coincEven... | 56395bb647ab92635e1398364b06f48bf65527b3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/56395bb647ab92635e1398364b06f48bf65527b3/stfu_pipe.py |
maxbin=0 | def getinjpar(inj,parnum): if parnum==0: return inj.mchirp if parnum==1: return inj.eta if parnum==2: return inj.get_end() if parnum==3: return inj.phi0 if parnum==4: return inj.distance if parnum==5: return inj.longitude if parnum==6: return inj.latitude if parnum==7: return inj.polarization if parnum==8: return inj.i... | be81ea5f646bb2637267e96ded5c2c39e7847000 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/be81ea5f646bb2637267e96ded5c2c39e7847000/OddsPostProc.py | |
hist[maxbin]=0 frac=frac+(maxbin/len(pos)) | hist[maxpos]=0 frac=frac+(float(maxbin)/float(len(pos))) | def getinjpar(inj,parnum): if parnum==0: return inj.mchirp if parnum==1: return inj.eta if parnum==2: return inj.get_end() if parnum==3: return inj.phi0 if parnum==4: return inj.distance if parnum==5: return inj.longitude if parnum==6: return inj.latitude if parnum==7: return inj.polarization if parnum==8: return inj.i... | be81ea5f646bb2637267e96ded5c2c39e7847000 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/be81ea5f646bb2637267e96ded5c2c39e7847000/OddsPostProc.py |
htmlfile.write('<table border=1><tr>') | htmlfile.write('<table border=1 width=100%><tr>') | def cbcBayesSkyRes(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None): if eventnum is not None and injfile is None: print "You specified an event number but no injection file. Ignoring!" if data is None: print 'You must specify an input data file' exit(1)... | 53137fadc21ddd0e3e68ed6dda56793f5e2409e7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53137fadc21ddd0e3e68ed6dda56793f5e2409e7/cbcBayesSkyRes.py |
oneDplotPath=os.path.join(outdir,param+'.png') | figname=param+'.png' oneDplotPath=os.path.join(outdir,figname) | def cbcBayesSkyRes(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None): if eventnum is not None and injfile is None: print "You specified an event number but no injection file. Ignoring!" if data is None: print 'You must specify an input data file' exit(1)... | 53137fadc21ddd0e3e68ed6dda56793f5e2409e7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53137fadc21ddd0e3e68ed6dda56793f5e2409e7/cbcBayesSkyRes.py |
myfig.savefig(os.path.join(outdir,param+'_samps.png')) | myfig.savefig(os.path.join(outdir,figname.replace('.png','_samps.png'))) | def cbcBayesSkyRes(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None): if eventnum is not None and injfile is None: print "You specified an event number but no injection file. Ignoring!" if data is None: print 'You must specify an input data file' exit(1)... | 53137fadc21ddd0e3e68ed6dda56793f5e2409e7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53137fadc21ddd0e3e68ed6dda56793f5e2409e7/cbcBayesSkyRes.py |
oneDplotPaths.append(oneDplotPath) | oneDplotPaths.append(figname) | def cbcBayesSkyRes(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None): if eventnum is not None and injfile is None: print "You specified an event number but no injection file. Ignoring!" if data is None: print 'You must specify an input data file' exit(1)... | 53137fadc21ddd0e3e68ed6dda56793f5e2409e7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53137fadc21ddd0e3e68ed6dda56793f5e2409e7/cbcBayesSkyRes.py |
oneDMenu=['mtotal','m1','m2','mchirp','mc','distance','distMPC','dist','iota','eta','RA','dec','a1','a2','phi1','theta1','phi2','theta2'] | oneDMenu=['mtotal','m1','m2','mchirp','mc','distance','distMPC','dist','iota','psi','eta','RA','dec','a1','a2','phi1','theta1','phi2','theta2'] | def cbcBayesSkyRes(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None): if eventnum is not None and injfile is None: print "You specified an event number but no injection file. Ignoring!" if data is None: print 'You must specify an input data file' exit(1)... | 53137fadc21ddd0e3e68ed6dda56793f5e2409e7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53137fadc21ddd0e3e68ed6dda56793f5e2409e7/cbcBayesSkyRes.py |
twoDplots=[['mc','eta'],['mchirp','eta'],['m1','m2'],['mtotal','eta'],['distance','iota'],['dist','iota'],['RA','dec'],['m1','dist'],['m2','dist']] | twoDplots=[['mc','eta'],['mchirp','eta'],['m1','m2'],['mtotal','eta'],['distance','iota'],['dist','iota'],['RA','dec'],['m1','dist'],['m2','dist'],['psi','iota'],['psi','distance'],['psi','dist'],['psi','phi0']] | def cbcBayesSkyRes(outdir,data,oneDMenu,twoDGreedyMenu,GreedyRes,confidence_levels,twoDplots,injfile=None,eventnum=None,skyres=None): if eventnum is not None and injfile is None: print "You specified an event number but no injection file. Ignoring!" if data is None: print 'You must specify an input data file' exit(1)... | 53137fadc21ddd0e3e68ed6dda56793f5e2409e7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/53137fadc21ddd0e3e68ed6dda56793f5e2409e7/cbcBayesSkyRes.py |
os.path.join("bin", "coh_PTF_html_summary"), os.path.join("bin", "coh_PTF_injfinder"), os.path.join("bin", "coh_PTF_sbv_plotter"), os.path.join("bin", "coh_PTF_trig_cluster"), os.path.join("bin", "coh_PTF_trig_combiner"), os.path.join("bin", "ring_post"), | os.path.join("bin", "coh_PTF_html_summary"), os.path.join("bin", "coh_PTF_injfinder"), os.path.join("bin", "coh_PTF_sbv_plotter"), os.path.join("bin", "coh_PTF_trig_cluster"), os.path.join("bin", "coh_PTF_trig_combiner"), os.path.join("bin", "ring_post"), | def run(self): # remove the automatically generated user env scripts for script in ["pylal-user-env.sh", "pylal-user-env.csh"]: log.info("removing " + script ) try: os.unlink(os.path.join("etc", script)) except: pass | de51ad9d89a7b3093c8fb0b3a547b5f14578ca8e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/de51ad9d89a7b3093c8fb0b3a547b5f14578ca8e/setup.py |
if pt[3] <= dtrss_inj and pt[3] != 0.0: | if pt[3] <= dtrss_inj: | def get_unique_filename(name): """ use this to avoid name collisions """ counter = 1 base_name, ext = os.path.splitext(name) while os.path.isfile(base_name): base_name = base_name + '_' + str(counter) + ext counter += 1 return base_name + ext | 3699225b3ce593e2fd33be880be8bf442ddf1f84 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/3699225b3ce593e2fd33be880be8bf442ddf1f84/run_skypoints.py |
self._InspiralAnalysisNode__pad_data = 0 | self._AnalysisNode__pad_data = 0 | def __init__(self, options, cp, dir='', tag_base=''): """ """ self.__conditionalLoadDefaults__(followUpChiaJob.defaults,cp) #self.__prog__ = 'followUpChiaJob' self.__executable = string.strip(cp.get('condor','chia')) self.__universe = "standard" pipeline.CondorDAGJob.__init__(self,self.__universe,self.__executable) sel... | 24648911372ea1df253fc45ee5f0a9b8ead25a34 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/24648911372ea1df253fc45ee5f0a9b8ead25a34/stfu_pipe.py |
self._InspiralAnalysisNode__pad_data = int(value) | self._AnalysisNode__pad_data = int(value) | def __init__(self, dag, job, cp, opts, sngl, frame_cache, chia, tag, p_nodes=[]): | 24648911372ea1df253fc45ee5f0a9b8ead25a34 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/24648911372ea1df253fc45ee5f0a9b8ead25a34/stfu_pipe.py |
self._InspiralAnalysisNode__pad_data = 0 | self._AnalysisNode__pad_data = 0 | def __init__(self, dag, job, cp, opts, coinc, inspiral_node_dict, chia_node =None, p_nodes = []): | 24648911372ea1df253fc45ee5f0a9b8ead25a34 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/24648911372ea1df253fc45ee5f0a9b8ead25a34/stfu_pipe.py |
flines[:,i]=asin(flines[:,i]) | flines[:,i]=arcsin(flines[:,i]) | def loadDataFile(filename): print filename infile=open(filename,'r') formatstr=infile.readline().lstrip() header=formatstr.split() llines=[] import re dec=re.compile(r'[^\d.-]+') for line in infile: sline=line.split() proceed=True for s in sline: if dec.search(s) is not None: print 'Warning! Ignoring non-numeric data a... | 11f936baaaabc17f21e8899cc738b602b5998e8c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/11f936baaaabc17f21e8899cc738b602b5998e8c/cbcBayesSkyRes.py |
flines[:,i]=acos(flines[:,i]) | flines[:,i]=arccos(flines[:,i]) | def loadDataFile(filename): print filename infile=open(filename,'r') formatstr=infile.readline().lstrip() header=formatstr.split() llines=[] import re dec=re.compile(r'[^\d.-]+') for line in infile: sline=line.split() proceed=True for s in sline: if dec.search(s) is not None: print 'Warning! Ignoring non-numeric data a... | 11f936baaaabc17f21e8899cc738b602b5998e8c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/11f936baaaabc17f21e8899cc738b602b5998e8c/cbcBayesSkyRes.py |
def add(self, *args): packing.Bin.add(self, *args) | def add(self, cache_entry): packing.Bin.add(self, cache_entry, cache_entry.to_segmentlistdict()) | def add(self, *args): packing.Bin.add(self, *args) self.extent = self.size.extent_all() return self | 5d6922774e4d74becff03265f02ee7265be8e613 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5d6922774e4d74becff03265f02ee7265be8e613/ligolw_cafe.py |
new.add(cache_entry, cache_entry.to_segmentlistdict()) | new.add(cache_entry) | def pack(self, cache_entry): """ Find all bins in which this glue.lal.CacheEntry instance belongs, merge them, and add this cache entry to the result. Create a new bin for this cache entry if it does not belong in any of the existing bins. | 5d6922774e4d74becff03265f02ee7265be8e613 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5d6922774e4d74becff03265f02ee7265be8e613/ligolw_cafe.py |
splits = [-segments.infinity()] + [lsctables.LIGOTimeGPS(origbin.extent[0] + i * float(origbin.extent[1] - origbin.extent[0]) / n) for i in range(1, n)] + [+segments.infinity()] | extents = [-segments.infinity()] + [lsctables.LIGOTimeGPS(origbin.extent[0] + i * float(abs(origbin.extent)) / n) for i in range(1, n)] + [+segments.infinity()] | def split_bins(cafepacker, extentlimit, verbose = False): """ Split bins in CafePacker so that each bin has an extent no longer than extentlimit. """ # # loop over all bins in cafepacker.bins # idx = 0 while idx < len(cafepacker.bins): # # retrieve bin # origbin = cafepacker.bins[idx] # # how many pieces? if bin d... | 5d6922774e4d74becff03265f02ee7265be8e613 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5d6922774e4d74becff03265f02ee7265be8e613/ligolw_cafe.py |
print >>sys.stderr, "\tsplitting cache spanning %s at %s" % (str(origbin.extent), ", ".join(str(split) for split in splits[1:-1])) splits = [segments.segmentlist([segments.segment(*bounds)]) for bounds in zip(splits[:-1], splits[1:])] splits = [segments.segmentlistdict.fromkeys(origbin.size, seglist) for seglist in spl... | print >>sys.stderr, "\tsplitting cache spanning %s at %s" % (str(origbin.extent), ", ".join(str(extent) for extent in extents[1:-1])) extents = [segments.segment(*bounds) & origbin.extent for bounds in zip(extents[:-1], extents[1:])] | def split_bins(cafepacker, extentlimit, verbose = False): """ Split bins in CafePacker so that each bin has an extent no longer than extentlimit. """ # # loop over all bins in cafepacker.bins # idx = 0 while idx < len(cafepacker.bins): # # retrieve bin # origbin = cafepacker.bins[idx] # # how many pieces? if bin d... | 5d6922774e4d74becff03265f02ee7265be8e613 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5d6922774e4d74becff03265f02ee7265be8e613/ligolw_cafe.py |
for split in splits: | for extent in extents: | def split_bins(cafepacker, extentlimit, verbose = False): """ Split bins in CafePacker so that each bin has an extent no longer than extentlimit. """ # # loop over all bins in cafepacker.bins # idx = 0 while idx < len(cafepacker.bins): # # retrieve bin # origbin = cafepacker.bins[idx] # # how many pieces? if bin d... | 5d6922774e4d74becff03265f02ee7265be8e613 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5d6922774e4d74becff03265f02ee7265be8e613/ligolw_cafe.py |
newbins[-1].size = origbin.size & split for key in tuple(newbins[-1].size): if not newbins[-1].size[key]: del newbins[-1].size[key] newbins[-1].extent = newbins[-1].size.extent_all() for bin in newbins: bin_extent_plus_max_gap = bin.extent.protract(cafepacker.max_gap) | extent_plus_max_gap = extent.protract(cafepacker.max_gap) | def split_bins(cafepacker, extentlimit, verbose = False): """ Split bins in CafePacker so that each bin has an extent no longer than extentlimit. """ # # loop over all bins in cafepacker.bins # idx = 0 while idx < len(cafepacker.bins): # # retrieve bin # origbin = cafepacker.bins[idx] # # how many pieces? if bin d... | 5d6922774e4d74becff03265f02ee7265be8e613 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5d6922774e4d74becff03265f02ee7265be8e613/ligolw_cafe.py |
if cache_entry.segment.disjoint(bin_extent_plus_max_gap): | if cache_entry.segment.disjoint(extent_plus_max_gap): | def split_bins(cafepacker, extentlimit, verbose = False): """ Split bins in CafePacker so that each bin has an extent no longer than extentlimit. """ # # loop over all bins in cafepacker.bins # idx = 0 while idx < len(cafepacker.bins): # # retrieve bin # origbin = cafepacker.bins[idx] # # how many pieces? if bin d... | 5d6922774e4d74becff03265f02ee7265be8e613 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5d6922774e4d74becff03265f02ee7265be8e613/ligolw_cafe.py |
if cache_entry_segs.intersects_segment(bin.extent): | if cache_entry_segs.intersects_segment(extent): | def split_bins(cafepacker, extentlimit, verbose = False): """ Split bins in CafePacker so that each bin has an extent no longer than extentlimit. """ # # loop over all bins in cafepacker.bins # idx = 0 while idx < len(cafepacker.bins): # # retrieve bin # origbin = cafepacker.bins[idx] # # how many pieces? if bin d... | 5d6922774e4d74becff03265f02ee7265be8e613 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5d6922774e4d74becff03265f02ee7265be8e613/ligolw_cafe.py |
bin.objects.append(cache_entry) | newbins[-1].add(cache_entry) | def split_bins(cafepacker, extentlimit, verbose = False): """ Split bins in CafePacker so that each bin has an extent no longer than extentlimit. """ # # loop over all bins in cafepacker.bins # idx = 0 while idx < len(cafepacker.bins): # # retrieve bin # origbin = cafepacker.bins[idx] # # how many pieces? if bin d... | 5d6922774e4d74becff03265f02ee7265be8e613 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5d6922774e4d74becff03265f02ee7265be8e613/ligolw_cafe.py |
injection_area=bin_size*i | injection_area=bin_size*(i+1) | def _greedy_bin(greedyHist,greedyPoints,injection_bin_index,bin_size,Nsamples,confidence_levels): """ An interal function representing the common, dimensionally-independent part of the greedy binning algorithms. """ #Now call confidence level C extension function to determine top-ranked pixels (injectionconfidence,top... | 927ddfda6dee0b24ed378ab3fbfbeed5d5522965 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/927ddfda6dee0b24ed378ab3fbfbeed5d5522965/bayespputils.py |
def fetchInformation(self,triggerTime=None,window=300): """ Wrapper for fetchInformationDualWindow that mimics original behavior """ return self.fetchInformationDualWindow(triggerTime,window,window,ifoList='DEFAULT') | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py | ||
later. """ | later. """ | def createDQbackground(self,ifoEpochList=list(),pickleLocale=None): """ Two inputs a list of tuples (ifo,epochname) for each instrument. Also a place to save the potential pickle to for quick access later. """ if type(ifoEpochList) != type(list()): raise Exception, \ "Invalid input argument ifoEpochList,%s type(%s)"\ %... | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py |
for (ifo,epoch) in ifoEpochList: | except: backgroundPickle=False sys.stderr.write("Error importing the pickle file! %s\n"\ %(pickleLocale)) return for (ifo,epoch) in ifoEpochList: | def createDQbackground(self,ifoEpochList=list(),pickleLocale=None): """ Two inputs a list of tuples (ifo,epochname) for each instrument. Also a place to save the potential pickle to for quick access later. """ if type(ifoEpochList) != type(list()): raise Exception, \ "Invalid input argument ifoEpochList,%s type(%s)"\ %... | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py |
generated background %s"%self.__backgroundDict__["ifoepoch"] except: backgroundPickle=False sys.stderr.write("Error importing the pickle file!\n") if os.access(os.path.split(pickleLocal)[0],os.W_OK): os.path.rename(pickleLocale,pickleLocale+".corrupt") | generated background expected %s got %s"%(\ self.__backgroundDict__["ifoepoch"], ifoEpochList) | def createDQbackground(self,ifoEpochList=list(),pickleLocale=None): """ Two inputs a list of tuples (ifo,epochname) for each instrument. Also a place to save the potential pickle to for quick access later. """ if type(ifoEpochList) != type(list()): raise Exception, \ "Invalid input argument ifoEpochList,%s type(%s)"\ %... | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py |
ifoEpochList=[(x,getRunEpoch(self.triggerTime,x)) for x in uniqIfos] | ifoEpochList=[(x,getRunEpoch(self.triggerTime,x)) for x in self.ifos] | def estimateDQbackground(self): """ This method looks at the self.resultlist inside the instance. Using this and 1000 generated time stamp it tabulates a ranking of flag prevelance, binomial probability 'p' """ if len(self.resultList) < 1: self.__backgroundResults__=list() self.__backgroundTimesDict__=dict() self.__bac... | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py |
sys.stderr.write("Aborting tabulate of binomial P\n") | sys.stderr.write("Aborting tabulation of binomial P\n") | def estimateDQbackground(self): """ This method looks at the self.resultlist inside the instance. Using this and 1000 generated time stamp it tabulates a ranking of flag prevelance, binomial probability 'p' """ if len(self.resultList) < 1: self.__backgroundResults__=list() self.__backgroundTimesDict__=dict() self.__bac... | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py |
for myIfo,flagList in seenFlags.iteritems(): | for myIfo in seenFlags.keys(): | def estimateDQbackground(self): """ This method looks at the self.resultlist inside the instance. Using this and 1000 generated time stamp it tabulates a ranking of flag prevelance, binomial probability 'p' """ if len(self.resultList) < 1: self.__backgroundResults__=list() self.__backgroundTimesDict__=dict() self.__bac... | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py |
for backgroundTime,backgroundFlags in \ self.__backgroundDict__[myIfo.strip()].iteritems(): tmpFlags.extend([name for ifo,name,ver,com,start,stop in backgroundFlags]) if myIfo not in self.__backgroundResults__.keys(): self.__backgroundResults__[myIfo]=dict() for myFlag in seenFlags[myIfo]: self.__backgroundResults__[my... | if myIfo.strip() not in self.__backgroundDict__.keys(): if myIfo not in self.__backgroundResults__.keys(): self.__backgroundResults__[myIfo]=dict() for outsideFlag in seenFlags[myIfo]: self.__backgroundResults__[myIfo][outsideFlag]=float(-0.0) else: for backgroundTime,backgroundFlags in \ self.__backgroundDict__[myIf... | def estimateDQbackground(self): """ This method looks at the self.resultlist inside the instance. Using this and 1000 generated time stamp it tabulates a ranking of flag prevelance, binomial probability 'p' """ if len(self.resultList) < 1: self.__backgroundResults__=list() self.__backgroundTimesDict__=dict() self.__bac... | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py |
tableString+=rowString%(myColor,ifo,name,version,start,offset1,stop,offset2,size) | tableString+=rowString%(myColor,ifo,name,version,\ start,offset1,stop,offset2,\ size,myBackgroundRank,myCategory) | def generateHTMLTable(self,tableType="BOTH"): """ Return a HTML table already formatted using the module MARKUP to keep the HTML tags complient. This method does nothing but return the result of the last call to self.fetchInformation() The flag names associated with LIGO will have links to the channel wiki in them als... | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py |
tableString+=rowString%(myColor,ifo,name,version,start,offset1,stop,offset2,size) | tableString+=rowString%(myColor,ifo,name,version,\ start,offset1,stop,offset2,\ size,myBackgroundRank,myCategory) | def generateHTMLTable(self,tableType="BOTH"): """ Return a HTML table already formatted using the module MARKUP to keep the HTML tags complient. This method does nothing but return the result of the last call to self.fetchInformation() The flag names associated with LIGO will have links to the channel wiki in them als... | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py |
tableString+=emptyRowString%myColor | tableString+=emptyRowString | def generateMOINMOINTable(self,tableType="BOTH"): """ Return a MOINMOIN table. """ ligo=["L1","H1","H2","V1"] channelWiki="https://ldas-jobs.ligo.caltech.edu/cgi-bin/chanwiki?%s" if self.triggerTime==int(-1): return "" myColor="grey" tableString="" titleString="" emptyRowString="" rowString="" for i,col in enumerate(se... | 5800af6be75ac2540e1e98d21cc7afadf15bf2aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5800af6be75ac2540e1e98d21cc7afadf15bf2aa/fu_utils.py |
myAngle=arcsin(dY/dX) | myAngle=arctan(dY/dX) | def scatterPointer(A=None,B=None): """ Expects two ordered pairs as tuples (x1,y1) and (x2,y2)... Then this defines an angle. From this angle we orient a triangle point and return this as a tuple of requested size. This can be plotted by the scatter plot to point in a particular orientation. The direction right(0,0)->... | b9aca6a019f1e02fdda02f66dcfacc9eb65f7287 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b9aca6a019f1e02fdda02f66dcfacc9eb65f7287/followupPDSurface.py |
myAngle=pi-arcsin(dY/dX) | myAngle=pi-arctan(dY/dX) | def scatterPointer(A=None,B=None): """ Expects two ordered pairs as tuples (x1,y1) and (x2,y2)... Then this defines an angle. From this angle we orient a triangle point and return this as a tuple of requested size. This can be plotted by the scatter plot to point in a particular orientation. The direction right(0,0)->... | b9aca6a019f1e02fdda02f66dcfacc9eb65f7287 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b9aca6a019f1e02fdda02f66dcfacc9eb65f7287/followupPDSurface.py |
myAngle=arcsin(dY/dX)+pi | myAngle=arctan(dY/dX)+pi | def scatterPointer(A=None,B=None): """ Expects two ordered pairs as tuples (x1,y1) and (x2,y2)... Then this defines an angle. From this angle we orient a triangle point and return this as a tuple of requested size. This can be plotted by the scatter plot to point in a particular orientation. The direction right(0,0)->... | b9aca6a019f1e02fdda02f66dcfacc9eb65f7287 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b9aca6a019f1e02fdda02f66dcfacc9eb65f7287/followupPDSurface.py |
myAngle=(2.0*pi)-arcsin(dY/dX) | myAngle=(2.0*pi)-arctan(dY/dX) | def scatterPointer(A=None,B=None): """ Expects two ordered pairs as tuples (x1,y1) and (x2,y2)... Then this defines an angle. From this angle we orient a triangle point and return this as a tuple of requested size. This can be plotted by the scatter plot to point in a particular orientation. The direction right(0,0)->... | b9aca6a019f1e02fdda02f66dcfacc9eb65f7287 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b9aca6a019f1e02fdda02f66dcfacc9eb65f7287/followupPDSurface.py |
type="string",default=None,\ | type="string",default="dummy",\ | def convert2DHistTo3ColVectors(myMatrix,xBins=None,yBins=None): """ Create three vectors X,Y,Z for ease of use in functs like contour, and scatter instead of imshow! If [x,y]Bins==None use bin index for X, Y Input matrix assumed square!! """ if xBins==None: xBins=range(len(myMatrix)) if yBins==None: yBins=range(len(myM... | b9aca6a019f1e02fdda02f66dcfacc9eb65f7287 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b9aca6a019f1e02fdda02f66dcfacc9eb65f7287/followupPDSurface.py |
origData=dict() origData["beam"]=beamSpigot.getDataStream(beamName,gpsStart,gpsEnd) | origData=dict() | def convert2DHistTo3ColVectors(myMatrix,xBins=None,yBins=None): """ Create three vectors X,Y,Z for ease of use in functs like contour, and scatter instead of imshow! If [x,y]Bins==None use bin index for X, Y Input matrix assumed square!! """ if xBins==None: xBins=range(len(myMatrix)) if yBins==None: yBins=range(len(myM... | b9aca6a019f1e02fdda02f66dcfacc9eb65f7287 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b9aca6a019f1e02fdda02f66dcfacc9eb65f7287/followupPDSurface.py |
tmpData=beamSpigot.getDataStream(myLabel[myKey],gpsA,gpsB) mySnipData[myKey]=interp(mySnipData["time"], getTimeStamps(tmpData), tmpData) | if beamName == "dummy" and myKey == "beam": mySnipData[myKey]=ones(size(mySnipData["time"])) else: tmpData=beamSpigot.getDataStream(myLabel[myKey],gpsA,gpsB) mySnipData[myKey]=interp(mySnipData["time"], getTimeStamps(tmpData), tmpData) | def convert2DHistTo3ColVectors(myMatrix,xBins=None,yBins=None): """ Create three vectors X,Y,Z for ease of use in functs like contour, and scatter instead of imshow! If [x,y]Bins==None use bin index for X, Y Input matrix assumed square!! """ if xBins==None: xBins=range(len(myMatrix)) if yBins==None: yBins=range(len(myM... | b9aca6a019f1e02fdda02f66dcfacc9eb65f7287 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b9aca6a019f1e02fdda02f66dcfacc9eb65f7287/followupPDSurface.py |
size=50,color='white') | size=starSize,color='white') | def convert2DHistTo3ColVectors(myMatrix,xBins=None,yBins=None): """ Create three vectors X,Y,Z for ease of use in functs like contour, and scatter instead of imshow! If [x,y]Bins==None use bin index for X, Y Input matrix assumed square!! """ if xBins==None: xBins=range(len(myMatrix)) if yBins==None: yBins=range(len(myM... | b9aca6a019f1e02fdda02f66dcfacc9eb65f7287 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b9aca6a019f1e02fdda02f66dcfacc9eb65f7287/followupPDSurface.py |
size=50,color='white') | size=starSize,color='white') | def convert2DHistTo3ColVectors(myMatrix,xBins=None,yBins=None): """ Create three vectors X,Y,Z for ease of use in functs like contour, and scatter instead of imshow! If [x,y]Bins==None use bin index for X, Y Input matrix assumed square!! """ if xBins==None: xBins=range(len(myMatrix)) if yBins==None: yBins=range(len(myM... | b9aca6a019f1e02fdda02f66dcfacc9eb65f7287 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b9aca6a019f1e02fdda02f66dcfacc9eb65f7287/followupPDSurface.py |
facecolor=None,\ | def convert2DHistTo3ColVectors(myMatrix,xBins=None,yBins=None): """ Create three vectors X,Y,Z for ease of use in functs like contour, and scatter instead of imshow! If [x,y]Bins==None use bin index for X, Y Input matrix assumed square!! """ if xBins==None: xBins=range(len(myMatrix)) if yBins==None: yBins=range(len(myM... | b9aca6a019f1e02fdda02f66dcfacc9eb65f7287 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/b9aca6a019f1e02fdda02f66dcfacc9eb65f7287/followupPDSurface.py | |
for offset_vector in self.offset_vectors: | for offset_vector in cafepacker.offset_vectors: | def split_bins(cafepacker, extentlimit): """ Split bins of stored in CafePacker until each bin has an extent no longer than extentlimit. """ # # loop overall the bins in cafepacker.bins. we pop items out of # cafepacker.bins and append new ones to the end so need a while loop # checking the extent of each bin in cafep... | 4a884fa265544fa4242befa60915a19c7bcfebb7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/4a884fa265544fa4242befa60915a19c7bcfebb7/ligolw_cafe.py |
num_days = int(round((duration)/86400)) | num_days = int(round((int(duration))/86400)) | def daily_ihope_cache(start,end,ifo,cluster=None): """ Generates cache list of daily ihope INSPIRAL xml files for give ifo and clustering (None,'30ms','100ms', or '16s') between start and end time """ #== daily path ihope_daily_path = '/archive/home/cbc/ihope_daily' #== set clustering tag if cluster==None or cluster.... | de85c159a571593db3ec7047a997bccf2f95a751 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/de85c159a571593db3ec7047a997bccf2f95a751/dqDataUtils.py |
cellString=cellString+" %s "%self.linkedRemoteImage(thumbs[ifo][myOmegaIndex], | cellString=cellString+" %s "%self.linkedRemoteImage(thumbs[ifo][myOmegaIndexT], | def insertAnalyzeQscanTable(self, images=None, thumbs=None, indexes=None, imagesAQ=None, thumbsAQ=None, indexesAQ=None, channelRanks=None): """ Insert a multiple IFO table with 5 cols with the AQ underneath this depends on the numer of IFO keys in indexes dictionary. The option channelRanks is not required to change th... | 5e1f81b21a4e936f81d47d9e5881e1ba01bee541 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5e1f81b21a4e936f81d47d9e5881e1ba01bee541/makeCheckListWiki.py |
cellString=cellString+" %s "%self.linkedRemoteImage(thumbsAQ[ifo][myAQIndex], | cellString=cellString+" %s "%self.linkedRemoteImage(thumbsAQ[ifo][myAQIndexT], | def insertAnalyzeQscanTable(self, images=None, thumbs=None, indexes=None, imagesAQ=None, thumbsAQ=None, indexesAQ=None, channelRanks=None): """ Insert a multiple IFO table with 5 cols with the AQ underneath this depends on the numer of IFO keys in indexes dictionary. The option channelRanks is not required to change th... | 5e1f81b21a4e936f81d47d9e5881e1ba01bee541 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/5e1f81b21a4e936f81d47d9e5881e1ba01bee541/makeCheckListWiki.py |
self.FAR = 99 | self.FAR = -1 | def __init__(self): """ here are all the things we need """ #start with data needed for every coinc self.ifo_list = [] self.ifo_coincs = [] self.snr = {} self.gps = {} self.eff_distances = {} self.mass1 = {} self.mass2 = {} self.time = None self.FAR = 99 #this stuff is only needed for injections self.is_injection = ... | c61b0d7232ed4fa06a8cfbef8a6aafce8f37875c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/c61b0d7232ed4fa06a8cfbef8a6aafce8f37875c/skylocutils.py |
def set_FAR(self,FAR_per_day): self.FAR=FAR_per_day | def set_inj_params(self,lat,lon,m1,m2,dist,effDs): """ set all of the injection parameters at once """ self.latitude_inj = lat self.longitude_inj = lon self.mass1_inj = m1 self.mass2_inj = m2 self.distance_inj = dist self.eff_distances_inj = effDs | c61b0d7232ed4fa06a8cfbef8a6aafce8f37875c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/c61b0d7232ed4fa06a8cfbef8a6aafce8f37875c/skylocutils.py | |
if ctab[0].false_alarm_rate is not None: coinc.set_FAR(ctab[0].false_alarm_rate) | def get_coincs_from_coinctable(self,files): """ read data from coinc tables (xml format) FIXME: currently assumes one coinc per file!!! """ for file in files: coinc = CoincData() xmldoc = utils.load_filename(file) sngltab = tab.get_table(xmldoc,lsctables.SnglInspiralTable.tableName) coinc.set_snr(dict((row.ifo, row.sn... | c61b0d7232ed4fa06a8cfbef8a6aafce8f37875c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/c61b0d7232ed4fa06a8cfbef8a6aafce8f37875c/skylocutils.py | |
while os.path.isfile(base_name): base_name = base_name + '_' + str(counter) + ext | while os.path.isfile(name): name = base_name + '_' + str(counter) + ext | def get_unique_filename(name): """ use this to avoid name collisions """ counter = 1 base_name, ext = os.path.splitext(name) while os.path.isfile(base_name): base_name = base_name + '_' + str(counter) + ext counter += 1 return base_name + ext | ed570091132aaa77d9da802494c40b42b3357586 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/ed570091132aaa77d9da802494c40b42b3357586/run_skypoints.py |
return base_name + ext | return name | def get_unique_filename(name): """ use this to avoid name collisions """ counter = 1 base_name, ext = os.path.splitext(name) while os.path.isfile(base_name): base_name = base_name + '_' + str(counter) + ext counter += 1 return base_name + ext | ed570091132aaa77d9da802494c40b42b3357586 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/ed570091132aaa77d9da802494c40b42b3357586/run_skypoints.py |
def append_process(xmldoc, comment = None, force = None, ds_sq_threshold = None, save_small_coincs = None, vetoes_name = None, verbose = None): | def append_process(xmldoc, comment = None, force = None, ds_sq_threshold = None, save_small_coincs = None, vetoes_name = None, coinc_end_time_segment = None, verbose = None): | def append_process(xmldoc, comment = None, force = None, ds_sq_threshold = None, save_small_coincs = None, vetoes_name = None, verbose = None): process = llwapp.append_process(xmldoc, program = process_program_name, version = __version__, cvs_repository = u"lscsoft", cvs_entry_time = __date__, comment = comment) param... | 33c375988e5fbb955cb57c83b181ddc3e9420c4e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/33c375988e5fbb955cb57c83b181ddc3e9420c4e/ligolw_rinca.py |
def append_coinc(self, process_id, time_slide_id, coinc_def_id, events): | def append_coinc(self, process_id, node, coinc_def_id, events): | def append_coinc(self, process_id, time_slide_id, coinc_def_id, events): # # populate the coinc_event and coinc_event_map tables # | 33c375988e5fbb955cb57c83b181ddc3e9420c4e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/33c375988e5fbb955cb57c83b181ddc3e9420c4e/ligolw_rinca.py |
time_slide_id = node.time_slide_id | def append_coinc(self, process_id, time_slide_id, coinc_def_id, events): # # populate the coinc_event and coinc_event_map tables # | 33c375988e5fbb955cb57c83b181ddc3e9420c4e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/33c375988e5fbb955cb57c83b181ddc3e9420c4e/ligolw_rinca.py | |
tstart = events[0].get_start() + self.time_slide_index[time_slide_id][events[0].ifo] coinc_ringdown.set_start(tstart + sum(event.snr * float(event.get_start() + self.time_slide_index[time_slide_id][event.ifo] - tstart) for event in events) / sum(event.snr for event in events)) | tstart = coinc_ringdown_start(events, node.offset_vector) coinc_ringdown.set_start(tstart) | def append_coinc(self, process_id, time_slide_id, coinc_def_id, events): # # populate the coinc_event and coinc_event_map tables # | 33c375988e5fbb955cb57c83b181ddc3e9420c4e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/33c375988e5fbb955cb57c83b181ddc3e9420c4e/ligolw_rinca.py |
tstart = coinc_ringdown.get_start() | def append_coinc(self, process_id, time_slide_id, coinc_def_id, events): # # populate the coinc_event and coinc_event_map tables # | 33c375988e5fbb955cb57c83b181ddc3e9420c4e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/33c375988e5fbb955cb57c83b181ddc3e9420c4e/ligolw_rinca.py | |
coinc_tables.append_coinc(process_id, node.time_slide_id, coinc_def_id, ntuple) | coinc_tables.append_coinc(process_id, node, coinc_def_id, ntuple) | def ligolw_rinca( xmldoc, process_id, EventListType, CoincTables, coinc_definer_row, event_comparefunc, thresholds, ntuple_comparefunc = lambda events, offset_vector: False, small_coincs = False, veto_segments = None, verbose = False | 33c375988e5fbb955cb57c83b181ddc3e9420c4e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/33c375988e5fbb955cb57c83b181ddc3e9420c4e/ligolw_rinca.py |
coinc_tables.append_coinc(process_id, node.time_slide_id, coinc_def_id, ntuple) | coinc_tables.append_coinc(process_id, node, coinc_def_id, ntuple) | def ligolw_rinca( xmldoc, process_id, EventListType, CoincTables, coinc_definer_row, event_comparefunc, thresholds, ntuple_comparefunc = lambda events, offset_vector: False, small_coincs = False, veto_segments = None, verbose = False | 33c375988e5fbb955cb57c83b181ddc3e9420c4e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/33c375988e5fbb955cb57c83b181ddc3e9420c4e/ligolw_rinca.py |
def __init__(self, config=None): | def __init__(self, configfile=None): | def __init__(self, config=None): cp = ConfigParser.ConfigParser() self.cp = cp self.time_now = "_".join([str(i) for i in time_method.gmtime()[0:6]]) self.ini_file=self.time_now + ".ini" home_base = home_dirs() # CONDOR SECTION NEEDED BY THINGS IN INSPIRAL.PY cp.add_section("condor") cp.set("condor","datafind",self.whi... | 7b60b47b4fbaad3d8fa52b8e122b7a950318aa85 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/7b60b47b4fbaad3d8fa52b8e122b7a950318aa85/stfu_pipe.py |
cp.set("condor","chia", self.which("lalapps_coherent_inspiral")) | cp.set("condor","chia", self.which("lalapps_coherent_inspiral")) | def __init__(self, config=None): cp = ConfigParser.ConfigParser() self.cp = cp self.time_now = "_".join([str(i) for i in time_method.gmtime()[0:6]]) self.ini_file=self.time_now + ".ini" home_base = home_dirs() # CONDOR SECTION NEEDED BY THINGS IN INSPIRAL.PY cp.add_section("condor") cp.set("condor","datafind",self.whi... | 7b60b47b4fbaad3d8fa52b8e122b7a950318aa85 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/7b60b47b4fbaad3d8fa52b8e122b7a950318aa85/stfu_pipe.py |
if config: | if configfile: | def __init__(self, config=None): cp = ConfigParser.ConfigParser() self.cp = cp self.time_now = "_".join([str(i) for i in time_method.gmtime()[0:6]]) self.ini_file=self.time_now + ".ini" home_base = home_dirs() # CONDOR SECTION NEEDED BY THINGS IN INSPIRAL.PY cp.add_section("condor") cp.set("condor","datafind",self.whi... | 7b60b47b4fbaad3d8fa52b8e122b7a950318aa85 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/7b60b47b4fbaad3d8fa52b8e122b7a950318aa85/stfu_pipe.py |
user_cp.read(config) | user_cp.read(configfile) | def __init__(self, config=None): cp = ConfigParser.ConfigParser() self.cp = cp self.time_now = "_".join([str(i) for i in time_method.gmtime()[0:6]]) self.ini_file=self.time_now + ".ini" home_base = home_dirs() # CONDOR SECTION NEEDED BY THINGS IN INSPIRAL.PY cp.add_section("condor") cp.set("condor","datafind",self.whi... | 7b60b47b4fbaad3d8fa52b8e122b7a950318aa85 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/7b60b47b4fbaad3d8fa52b8e122b7a950318aa85/stfu_pipe.py |
if user_cp: self.overwrite_config(user_cp) | if user_cp: self.overwrite_config(user_cp,cp) | def __init__(self, config=None): cp = ConfigParser.ConfigParser() self.cp = cp self.time_now = "_".join([str(i) for i in time_method.gmtime()[0:6]]) self.ini_file=self.time_now + ".ini" home_base = home_dirs() # CONDOR SECTION NEEDED BY THINGS IN INSPIRAL.PY cp.add_section("condor") cp.set("condor","datafind",self.whi... | 7b60b47b4fbaad3d8fa52b8e122b7a950318aa85 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/7b60b47b4fbaad3d8fa52b8e122b7a950318aa85/stfu_pipe.py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.