rem
stringlengths
1
226k
add
stringlengths
0
227k
context
stringlengths
6
326k
meta
stringlengths
143
403
input_ids
listlengths
256
256
attention_mask
listlengths
256
256
labels
listlengths
128
128
headers, 'file:'+pathname2url(file))
headers, 'file:'+file)
def open_local_file(self, url): import mimetypes, mimetools, StringIO mtype = mimetypes.guess_type(url)[0] headers = mimetools.Message(StringIO.StringIO( 'Content-Type: %s\n' % (mtype or 'text/plain'))) host, file = splithost(url) if not host: return addinfourl(open(url2pathname(file), 'rb'), headers, 'file:'+pathname2url(file)) host, port = splitport(host) if not port \ and socket.gethostbyname(host) in (localhost(), thishost()): return addinfourl(open(url2pathname(file), 'rb'), headers, 'file:'+pathname2url(file)) raise IOError, ('local file error', 'not on local host')
3764595c980ccba821a29947dbef14947b71825e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/3764595c980ccba821a29947dbef14947b71825e/urllib.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1696, 67, 3729, 67, 768, 12, 2890, 16, 880, 4672, 1930, 20369, 15180, 16, 20369, 278, 8192, 16, 15777, 22189, 273, 20369, 15180, 18, 20885, 67, 723, 12, 718, 25146, 20, 65, 1607, 273, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1696, 67, 3729, 67, 768, 12, 2890, 16, 880, 4672, 1930, 20369, 15180, 16, 20369, 278, 8192, 16, 15777, 22189, 273, 20369, 15180, 18, 20885, 67, 723, 12, 718, 25146, 20, 65, 1607, 273, ...
hour * self.hourHeight,
hour * self.hourHeight,
def DrawBackground(self, dc): styles = self.parent self._doDrawingCalculations()
590a7421d4a56497721ad83a36a6f3eb4226407e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/590a7421d4a56497721ad83a36a6f3eb4226407e/CalendarCanvas.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 10184, 8199, 12, 2890, 16, 6744, 4672, 5687, 273, 365, 18, 2938, 365, 6315, 2896, 26885, 4844, 1012, 1435, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 10184, 8199, 12, 2890, 16, 6744, 4672, 5687, 273, 365, 18, 2938, 365, 6315, 2896, 26885, 4844, 1012, 1435, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
self.__updateJobStatus('Failed','%s Undefined' %(param))
self.__updateJobStatus(job,'Failed','%s Undefined' %(param))
def __preparePilot(self,job,classadJob,attributes): """The pilot agent job can now be formulated for a job that is deemed eligible. """ jdlCPU = 0 if classadJob.lookupAttribute("MaxCPUTime"): jdlCPU = int(string.replace(classadJob.get_expression("MaxCPUTime"),'"','') )
e4e152162784e68280a5aedde1e5bad201fa5992 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12864/e4e152162784e68280a5aedde1e5bad201fa5992/AgentDirector.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 9366, 52, 22797, 12, 2890, 16, 4688, 16, 1106, 361, 2278, 16, 4350, 4672, 3536, 1986, 293, 22797, 4040, 1719, 848, 2037, 506, 646, 11799, 364, 279, 1719, 716, 353, 443, 351, 329, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 9366, 52, 22797, 12, 2890, 16, 4688, 16, 1106, 361, 2278, 16, 4350, 4672, 3536, 1986, 293, 22797, 4040, 1719, 848, 2037, 506, 646, 11799, 364, 279, 1719, 716, 353, 443, 351, 329, ...
if line.split()[0].lower() == 'rem' \ and line[0] in "rR":
if line.split(None, 1)[0].lower() == 'rem' and line[0] in "rR":
def __read(self, fp, fpname): """Parse a sectioned setup file.
05e74c488d7085fb533b40a9d6860ee732e784e4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/05e74c488d7085fb533b40a9d6860ee732e784e4/ConfigParser.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 896, 12, 2890, 16, 4253, 16, 4253, 529, 4672, 3536, 3201, 279, 2442, 329, 3875, 585, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 896, 12, 2890, 16, 4253, 16, 4253, 529, 4672, 3536, 3201, 279, 2442, 329, 3875, 585, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100...
head = _colorize(doc, heading_token, errors, warnings, 'heading')
head = _colorize(doc, heading_token, errors, 'heading')
def _add_section(doc, heading_token, stack, indent_stack, errors, warnings): """Add a new section to the DOM tree, with the given heading.""" if indent_stack[-1] == None: indent_stack[-1] = heading_token.indent elif indent_stack[-1] != heading_token.indent: estr = "Improper heading indentation." errors.append(StructuringError(estr, heading_token)) # Check for errors. for tok in stack[2:]: if tok.tagName != "section": estr = "Headings must occur at the top level." errors.append(StructuringError(estr, heading_token)) break if (heading_token.level+2) > len(stack): estr = "Wrong underline character for heading." errors.append(StructuringError(estr, heading_token)) # Pop the appropriate number of headings so we're at the # correct level. stack[heading_token.level+2:] = [] indent_stack[heading_token.level+2:] = [] # Colorize the heading head = _colorize(doc, heading_token, errors, warnings, 'heading') # Add the section's and heading's DOM elements. sec = doc.createElement("section") stack[-1].appendChild(sec) stack.append(sec) sec.appendChild(head) indent_stack.append(None)
36ace726c8fc4e9c1a450d16c75087d594ead7cf /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11420/36ace726c8fc4e9c1a450d16c75087d594ead7cf/epytext.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 1289, 67, 3464, 12, 2434, 16, 11053, 67, 2316, 16, 2110, 16, 3504, 67, 3772, 16, 1334, 16, 5599, 4672, 3536, 986, 279, 394, 2442, 358, 326, 4703, 2151, 16, 598, 326, 864, 11053, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 1289, 67, 3464, 12, 2434, 16, 11053, 67, 2316, 16, 2110, 16, 3504, 67, 3772, 16, 1334, 16, 5599, 4672, 3536, 986, 279, 394, 2442, 358, 326, 4703, 2151, 16, 598, 326, 864, 11053, ...
self.logger.info("Forcing state to true for pkg %s" % (pkg.get('name')))
self.logger.info("Forcing state to true for pkg %s" % \ (pkg.get('name')))
def Install(self, packages, states): """ Run a one-pass install, followed by single pkg installs in case of failure. """ self.logger.info("Trying single pass package install for pkgtype %s" % \ self.pkgtype)
decb4dd187c808e2f1c0ff7040e914abc7adeae2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11867/decb4dd187c808e2f1c0ff7040e914abc7adeae2/__init__.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 10284, 12, 2890, 16, 5907, 16, 5493, 4672, 3536, 1939, 279, 1245, 17, 5466, 3799, 16, 10860, 635, 2202, 3475, 31011, 316, 648, 434, 5166, 18, 3536, 365, 18, 4901, 18, 1376, 2932, 18038, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 10284, 12, 2890, 16, 5907, 16, 5493, 4672, 3536, 1939, 279, 1245, 17, 5466, 3799, 16, 10860, 635, 2202, 3475, 31011, 316, 648, 434, 5166, 18, 3536, 365, 18, 4901, 18, 1376, 2932, 18038, ...
if datasettype in [ 'TAG', 'TAG_REC', 'TNT_DOWNLOAD', 'TNT_LOCAL']: if os.access('./tag.tar.gz',os.R_OK): cmd = 'tar xvzf tag.tar.gz' rc, out = getstatusoutput(cmd) if (rc!=0): print "ERROR: error during extraction of tar.tar.gz" print out sys.exit(EC_UNSPEC) dir = "." filepat = "\.root" pat = re.compile(filepat) filelist = os.listdir(dir) joName = 'input.py' outFile = open(joName,'w') if datasettype in [ 'TAG_REC', 'TNT_DOWNLOAD', 'TNT_LOCAL' ]: if atlas_release_major >= 13: outFile.write('PoolTAGInput = [') else: outFile.write('CollInput = [') else: if atlas_release_major >= 13: versionString='ServiceMgr.' else: versionString = '' outFile.write('%sEventSelector.CollectionType="ExplicitROOT"\n'%versionString) outFile.write('%sEventSelector.RefName = "StreamAOD"\n'%versionString) outFile.write('%sEventSelector.InputCollections = ['%versionString) for file in filelist: found = re.findall(pat, file) if found: filename = re.sub('\.root\.\d+$','',file) if atlas_release_major <= 12: filename = re.sub('\.root$','',file) outFile.write('"%s",' % filename) outFile.write(']\n') outFile.close() else: try: tagdatasetnames = os.environ['TAGDATASETNAME'].split(":") except: raise NameError, "ERROR: TAGDATASETNAME not defined" sys.exit(EC_Configuration) cmd = 'which python32' pythoncmd = '' rc, out = commands.getstatusoutput(cmd) if (rc!=0): print 'No python32 found' pythoncmd = '' else: pythoncmd = out.strip() if os.environ.has_key('pybin'): pythoncmd = os.environ['pybin'] dq2setuppath = '$VO_ATLAS_SW_DIR/ddm/latest/setup.sh' inputtxt = 'dq2localid.txt' try: temp_dq2localsiteid = [ line.strip() for line in file(inputtxt) ] dq2localsiteid = temp_dq2localsiteid[0] except: dq2localsiteid = os.environ[ 'DQ2_LOCAL_SITE_ID' ] pass for tagdatasetname in tagdatasetnames: cmd = 'source %s; dq2-get --client-id=ganga --automatic --local-site=%s --no-directories --timeout %s -p lcg %s' % (dq2setuppath, dq2localsiteid ,timeout, tagdatasetname) cmdretry = 'source %s; dq2-get --client-id=ganga --automatic --local-site=CERN-PROD_DATADISK --no-directories --timeout %s -p lcg %s' % (dq2setuppath, dq2localsiteid ,timeout, tagdatasetname) rc, out = getstatusoutput(cmd) print out if (rc!=0): print "ERROR: error during dq2-get occured" rc, out = getstatusoutput(cmdretry) print out if (rc!=0): print "ERROR: error during retry of dq2-get occured" sys.exit(EC_DQ2GET) taglfns = [ line.strip() for line in file('input_tag_files') ] tagguids = [ line.strip() for line in file('input_tag_guids') ] tagddmFileMap = {} for i in xrange(0,len(taglfns)): tagddmFileMap[taglfns[i]] = tagguids[i] files = {} pfnsnew = [] for lfn, guid in tagddmFileMap.iteritems(): name = os.path.basename(lfn) pfn = os.path.join(directory,name) try: open(pfn) fsize = os.stat(pfn).st_size except IOError: print "ERROR %s not found" % name continue if (fsize>0): item = {'pfn':pfn,'guid':guid} files[lfn] = item _makeJobO(files, tag=True, type=datasettype, version=atlas_release_major, dtype=datatype) dir = "." filepat = "pool\.root\.\d+$" pat = re.compile(filepat) filelist = os.listdir(dir) for file in filelist: found = re.findall(pat, file) if found: filenew = re.sub('\.\d+$','',file) try: os.symlink(file,filenew) except OSError: pass if datasettype == 'TAG_REC': if os.environ.has_key('ATHENA_OPTIONS'): joboptions = os.environ['ATHENA_OPTIONS'].split(' ') if atlas_release_major >= 13: linepat = "^PoolInputQuery=" else: linepat = "^CollInputQuery=" pat = re.compile(linepat) for jfile in joboptions: try: jolines = [ rline.strip() for rline in open(jfile,'r') ] except IOError: jolines = [] newlines = [] for l in jolines: found = re.findall(pat, l) if found: newlines.append("include ( \"input.py\" )") newlines.append(l) outFile = open(jfile,'w') for l in newlines: outFile.write(l+'\n') outFile.close()
def findsetype(sitesrm): setype= 'NULL' if sitesrm.find('castor')>=0: setype = 'CASTOR' elif sitesrm.find('dpm')>=0: setype = 'DPM' elif sitesrm.find('pnfs')>=0: setype = 'DCACHE' elif sitesrm.find('/nfs/')>=0: setype = 'NFS' return setype
8126e8e033d86f07c6cd5ab3c324635fbccddab7 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1488/8126e8e033d86f07c6cd5ab3c324635fbccddab7/ganga-stage-in-out-dq2.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1104, 542, 388, 12, 12180, 8864, 4672, 444, 388, 33, 296, 8560, 11, 225, 309, 9180, 8864, 18, 4720, 2668, 4155, 280, 6134, 34, 33, 20, 30, 444, 388, 273, 296, 21871, 916, 11, 1327, 9...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1104, 542, 388, 12, 12180, 8864, 4672, 444, 388, 33, 296, 8560, 11, 225, 309, 9180, 8864, 18, 4720, 2668, 4155, 280, 6134, 34, 33, 20, 30, 444, 388, 273, 296, 21871, 916, 11, 1327, 9...
while self.current_control_tick < control_tick: self.current_control_tick += 1 win32api.PostMessage(self.hprogress, commctrl.PBM_STEPIT, 0, 0)
if verbose: print "Tick", self.current_stage_tick, "is", this_prop, "through the stage,", total_prop, "through the total - ctrl tick is", control_tick win32api.PostMessage(self.hprogress, commctrl.PBM_SETPOS, control_tick)
def tick(self): self.current_stage_tick += 1 # Calc how far through this stage. this_prop = float(self.current_stage_tick) / self.current_stage_max # How far through the total. stage_name, start, end = self._get_current_stage() # Calc the perc of the total control. stage_name, start, prop = self._get_current_stage() total_prop = start + this_prop * prop # How may ticks is this on the control (but always have 1, so the # user knows the process has actually started.) control_tick = max(1,int(total_prop * self.total_control_ticks)) #print "Tick", self.current_stage_tick, "is", this_prop, "through the stage,", total_prop, "through the total - ctrl tick is", control_tick while self.current_control_tick < control_tick: self.current_control_tick += 1 #print "ticking control", self.current_control_tick win32api.PostMessage(self.hprogress, commctrl.PBM_STEPIT, 0, 0)
4719b68626c47e5b4e78e019c8a48ff3506576c8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6126/4719b68626c47e5b4e78e019c8a48ff3506576c8/async_processor.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4024, 12, 2890, 4672, 365, 18, 2972, 67, 12869, 67, 6470, 1011, 404, 468, 29128, 3661, 10247, 3059, 333, 6009, 18, 333, 67, 5986, 273, 1431, 12, 2890, 18, 2972, 67, 12869, 67, 6470, 13...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4024, 12, 2890, 4672, 365, 18, 2972, 67, 12869, 67, 6470, 1011, 404, 468, 29128, 3661, 10247, 3059, 333, 6009, 18, 333, 67, 5986, 273, 1431, 12, 2890, 18, 2972, 67, 12869, 67, 6470, 13...
arcIdMatch = re.search("(\w+://([a-zA-Z0-9.-]+)\S*/\d*)", output)
arcIdMatch = re.search("(\w+://([a-zA-Z0-9.-]+)\S*/\d*)", jobstring)
def query(self, obj, service='', objType='node'): """ Query status and eventually other scheduler related information, and store it in the job.runningJob data structure.
41f02dda6faeaa4fd642e5315ef1d6eb56a334a3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8886/41f02dda6faeaa4fd642e5315ef1d6eb56a334a3/SchedulerARC.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 843, 12, 2890, 16, 1081, 16, 1156, 2218, 2187, 30078, 2218, 2159, 11, 4672, 3536, 2770, 1267, 471, 18011, 1308, 8129, 3746, 1779, 16, 471, 1707, 518, 316, 326, 1719, 18, 8704, 2278, 501,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 843, 12, 2890, 16, 1081, 16, 1156, 2218, 2187, 30078, 2218, 2159, 11, 4672, 3536, 2770, 1267, 471, 18011, 1308, 8129, 3746, 1779, 16, 471, 1707, 518, 316, 326, 1719, 18, 8704, 2278, 501,...
Popen([os.path.join(am_dir, "../unit_tree/TeamColorizer"), src, path + "/" + imgurl])
def w(x): f.write(x + "\n")
40959658132c7772a3b77bd1fd4f42f57abbe0c4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9355/40959658132c7772a3b77bd1fd4f42f57abbe0c4/html.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 341, 12, 92, 4672, 284, 18, 2626, 12, 92, 397, 1548, 82, 7923, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 341, 12, 92, 4672, 284, 18, 2626, 12, 92, 397, 1548, 82, 7923, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
trace(2, depth, "early abort of _updown.upg at %s", (upgpkg))
self.trace(2, "early abort of _updown.upg at %s", (upgpkg))
def _updown(self, pkg, force): trans = self._trans changeset = self._changeset locked = self._locked depth = self._depth pruneweight = self._pruneweight trace(1, depth, "_updown(%s, pw=%f, yw=%f, f=%d)", (pkg, pruneweight, self._yieldweight, force))
b26ec53d5174d18eebf24ded272ad218fb19d0b4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8317/b26ec53d5174d18eebf24ded272ad218fb19d0b4/transaction.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 416, 2378, 12, 2890, 16, 3475, 16, 2944, 4672, 906, 273, 365, 6315, 2338, 22463, 273, 365, 6315, 6329, 278, 8586, 273, 365, 6315, 15091, 3598, 273, 365, 6315, 5979, 846, 318, 359, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 416, 2378, 12, 2890, 16, 3475, 16, 2944, 4672, 906, 273, 365, 6315, 2338, 22463, 273, 365, 6315, 6329, 278, 8586, 273, 365, 6315, 15091, 3598, 273, 365, 6315, 5979, 846, 318, 359, ...
os.system('/usr/sbin/undomatroxcolor')
def run(self): while 1: if self.mode == 'idle': self.mode_flag.wait() self.mode_flag.clear()
4d4cbecd1b64b161427892bd2f2bed9b00002280 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11399/4d4cbecd1b64b161427892bd2f2bed9b00002280/mplayer.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 4672, 1323, 404, 30, 309, 365, 18, 3188, 422, 296, 20390, 4278, 365, 18, 3188, 67, 6420, 18, 7048, 1435, 365, 18, 3188, 67, 6420, 18, 8507, 1435, 2, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 4672, 1323, 404, 30, 309, 365, 18, 3188, 422, 296, 20390, 4278, 365, 18, 3188, 67, 6420, 18, 7048, 1435, 365, 18, 3188, 67, 6420, 18, 8507, 1435, 2, -100, -100, -100, ...
buffer.append(data, t) buffer.append(None)
queue.put(data, int(id*tpf)) t1 = time.millitimer() gl.wintitle('(busy) ' + filename) print lastid, 'fields in', t1-t0, 'msec', print '--', 0.1 * int(lastid * 10000.0 / (t1-t0)), 'fields/sec' if audiofilename: audiostop.append(None) audiodone.acquire_lock() v.EndContinuousCapture() queue.put(None) done.acquire_lock()
def record(v, filename, audiofilename): import thread x, y = gl.getsize() vout = VFile.VoutFile().init(filename) vout.format = 'rgb8' vout.width = x vout.height = y vout.writeheader() buffer = [] thread.start_new_thread(saveframes, (vout, buffer)) if audiofilename: initaudio(audiofilename, buffer) gl.wintitle('(rec) ' + filename) v.StartCapture() t0 = time.millitimer() while not gl.qtest(): if v.GetCaptured() > 2: t = time.millitimer() - t0 cd, st = v.GetCaptureData() data = cd.interleave(x, y) cd.UnlockCaptureData() buffer.append(data, t) else: time.millisleep(10) v.StopCapture() while v.GetCaptured() > 0: t = time.millitimer() - t0 cd, st = v.GetCaptureData() data = cd.interleave(x, y) cd.UnlockCaptureData() buffer.append(data, t) buffer.append(None) # Sentinel gl.wintitle('(done) ' + filename)
62f6bc8e55ef2d0e226b1c3ad7b9aca58407ab7a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/62f6bc8e55ef2d0e226b1c3ad7b9aca58407ab7a/Vrec.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1409, 12, 90, 16, 1544, 16, 20232, 77, 792, 2550, 4672, 1930, 2650, 619, 16, 677, 273, 5118, 18, 588, 1467, 1435, 331, 659, 273, 776, 812, 18, 58, 659, 812, 7675, 2738, 12, 3459, 13,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1409, 12, 90, 16, 1544, 16, 20232, 77, 792, 2550, 4672, 1930, 2650, 619, 16, 677, 273, 5118, 18, 588, 1467, 1435, 331, 659, 273, 776, 812, 18, 58, 659, 812, 7675, 2738, 12, 3459, 13,...
log_and_exec(CHECK + " user -v --name utilisager.normal --auto-yes")
log_and_exec(CHK + " user -v --name utilisager.normal --auto-yes")
def test_users(): """Test ADD/MOD/DEL on user accounts in various ways.""" test_message('''starting users related tests.''') log_and_exec(ADD + " group --name test_users_A --description 'groupe créé pour la suite de tests sur les utilisateurs, vous pouvez effacer'") log_and_exec(ADD + " profile --name Utilisagers --group utilisagers --comment 'profil normal créé pour la suite de tests utilisateurs'") log_and_exec(ADD + " profile --name Responsibilisateurs --group responsibilisateurs --groups cdrom,lpadmin,plugdev,audio,video,scanner,fuse --comment 'profil power user créé pour la suite de tests utilisateurs.'") log_and_exec(ADD + " group --name test_users_B --description 'groupe créé pour la suite de tests sur les utilisateurs, vous pouvez effacer'") os.system(GETENT + " groups") os.system(GETENT + " profiles") log_and_exec(ADD + " user --firstname Utiliçateur --lastname Accentué") log_and_exec(ADD + " user --gecos 'Utilisateur Accentué n°2'", True, 12, comment = "can't build a login from only a GECOS field.") log_and_exec(ADD + " user --login utilisager.normal --profile utilisagers") log_and_exec(MODIFY + " user --login=utilisager.normal -v --add-groups test_users_A") log_and_exec(MODIFY + " user --login=utilisager.normal -v --add-groups test_users_B") # should produce nothing, because nothing is wrong. log_and_exec(CHECK + " group -v --name test_users_B") os.system("rm ~utilisager.normal/test_users_A") # all must be OK, extended checks are not enabled, the program will not "see" the missing link. log_and_exec(CHECK + " group -v --name test_users_A") # the link to group_A isn't here ! log_and_exec(CHECK + " group -vv --name test_users_A --extended --auto-no", True, 7, comment = "a user lacks a symlink.") log_and_exec(CHECK + " group -vv --name test_users_A --extended --auto-yes") # the same check, but checking from users.py #os.system("rm ~utilisager.normal/test_users_A") #log_and_exec(CHECK + " user --name utilisager.normal") # not yet implemented #log_and_exec(CHECK + " user --name utilisager.normal --extended --auto-no", True, 7, comment="user lacks symlink") #log_and_exec(CHECK + " user --name utilisager.normal --extended --auto-yes") # checking for Maildir repair capacity... if configuration.users.mailbox_type == configuration.MAIL_TYPE_HOME_MAILDIR: os.system("rm -rf ~utilisager.normal/" + configuration.users.mailbox) log_and_exec(CHECK + " user -v --name utilisager.normal --auto-no", True, 7, comment="user lacks ~/" + configuration.users.mailbox) log_and_exec(CHECK + " user -v --name utilisager.normal --auto-yes") os.system("touch ~utilisager.normal/.dmrc ; chmod 666 ~utilisager.normal/.dmrc") log_and_exec(CHECK + " user -v --name utilisager.normal --auto-yes") os.system("mv -f ~utilisager.normal/test_users_B ~utilisager.normal/mon_groupe_B_préféré") # all must be ok, the link is just renamed... log_and_exec(CHECK + " group -vv --name test_users_B --extended") # FIXME: verify the user can create things in shared group dirs. log_and_exec(MODIFY + " user --login=utilisager.normal --del-groups test_users_A") # should fail log_and_exec(MODIFY + " user --login=utilisager.normal --del-groups test_users_A", comment = "already not a member.") log_and_exec(ADD + " user --login test.responsibilly --profile responsibilisateurs") log_and_exec(MODIFY + " profile --group utilisagers --add-groups cdrom") log_and_exec(MODIFY + " profile --group utilisagers --add-groups cdrom,test_users_B") log_and_exec(MODIFY + " profile --group utilisagers --apply-groups") log_and_exec(MODIFY + " profile --group responsibilisateurs --add-groups plugdev,audio,test_users_A") log_and_exec(MODIFY + " profile --group responsibilisateurs --del-groups audio") log_and_exec(MODIFY + " profile --group responsibilisateurs --apply-groups") # clean the system log_and_exec(DELETE + " user --login utilicateur.accentue") log_and_exec(DELETE + " user --login utilisateur.accentuen2", True, 5, comment = "this user has *NOT* been created previously.") log_and_exec(DELETE + " profile -vvv --group utilisagers --del-users --no-archive") #os.system(GETENT + " users") log_and_exec(DELETE + " profile --group responsibilisateurs", True, 12, comment = "there are still some users in the pri group of this profile.") log_and_exec(DELETE + " group --name=test_users_A --del-users --no-archive") log_and_exec(DELETE + " user --login test.responsibilly") # this should work now that the last user has been deleted log_and_exec(DELETE + " profile --group responsibilisateurs") log_and_exec(DELETE + " group --name=test_users_B -vv") # already deleted before #log_and_exec(DELETE + " user --login utilisager.normal") #log_and_exec(DELETE + " user --login test.responsibilly") test_message('''users related tests finished.''')
07cb13c975b820cb15074a8378ec89357e5846f4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7650/07cb13c975b820cb15074a8378ec89357e5846f4/core.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 5577, 13332, 3536, 4709, 11689, 19, 6720, 19, 24733, 603, 729, 9484, 316, 11191, 16226, 12123, 225, 1842, 67, 2150, 2668, 6309, 18526, 3677, 3746, 7434, 1093, 11, 6134, 225, 613,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 5577, 13332, 3536, 4709, 11689, 19, 6720, 19, 24733, 603, 729, 9484, 316, 11191, 16226, 12123, 225, 1842, 67, 2150, 2668, 6309, 18526, 3677, 3746, 7434, 1093, 11, 6134, 225, 613,...
wr.position.x = center_col_vec[0,0] wr.position.y = center_col_vec[1,0] wr.position.z = center_col_vec[2,0]
lock_pos = tr.translation_matrix(np.matrix([0,-.268,-.150]))*tfu.transform(self.omni_name+'_sensable', self.omni_name, self.tflistener)*np.row_stack((center_col_vec, np.matrix([1.]))) wr.position.x = (lock_pos[0,0])*1000.0 wr.position.y = (lock_pos[1,0])*1000.0 wr.position.z = (lock_pos[2,0])*1000.0
def omni_pose_cb(self, msg): if self.enabled: #Get the omni's tip pose in the PR2's torso frame tip_omni, msg_frame = tfu.posestamped_as_matrix(msg) self.torso_T_omni(tip_omni, msg_frame) tip_tt = self.tip_tt tip_tq = self.tip_tq #Publish new arm pose ps = PoseStamped() ps.header.frame_id = '/torso_lift_link' ps.header.stamp = rospy.get_rostime() ps.pose.position.x = tip_tt[0] ps.pose.position.y = tip_tt[1] ps.pose.position.z = tip_tt[2] ps.pose.orientation.x = tip_tq[0] ps.pose.orientation.y = tip_tq[1] ps.pose.orientation.z = tip_tq[2] ps.pose.orientation.w = tip_tq[3] self.pr2_pub.publish(ps) if self.zero_out_forces: wr = OmniFeedback() wr.force.x = 0 wr.force.y = 0 wr.force.z = 0 self.omni_fb.publish(wr) self.zero_out_forces = False else: #this is a zero order hold publishing the last received values until the control loop is active again tip_tt = self.tip_tt tip_tq = self.tip_tq ps = PoseStamped() ps.header.frame_id = '/torso_lift_link' ps.header.stamp = rospy.get_rostime() ps.pose.position.x = tip_tt[0] ps.pose.position.y = tip_tt[1] ps.pose.position.z = tip_tt[2] ps.pose.orientation.x = tip_tq[0] ps.pose.orientation.y = tip_tq[1] ps.pose.orientation.z = tip_tq[2] ps.pose.orientation.w = tip_tq[3] self.pr2_pub.publish(ps)
7b7ca3a268c5daaec0461231ecdc51422ad69239 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8117/7b7ca3a268c5daaec0461231ecdc51422ad69239/pr2_omni_teleop.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8068, 15834, 67, 4150, 67, 7358, 12, 2890, 16, 1234, 4672, 309, 365, 18, 5745, 30, 468, 967, 326, 8068, 15834, 1807, 9529, 26948, 316, 326, 11770, 22, 1807, 268, 1383, 83, 2623, 9529, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8068, 15834, 67, 4150, 67, 7358, 12, 2890, 16, 1234, 4672, 309, 365, 18, 5745, 30, 468, 967, 326, 8068, 15834, 1807, 9529, 26948, 316, 326, 11770, 22, 1807, 268, 1383, 83, 2623, 9529, ...
self.logger.debug('Calling ' + functionPath + str(args))
if not s.loggedin: self.logger.debug("RPC method call from unauthenticated user: %s" % functionPath + str(args)) else: self.logger.debug("RPC method call from user %s: %s" % (s.userid, functionPath + str(args)))
def render(self, request): """ override method of xmlrpc python twisted framework
e02ca120fa18409a237d312ac7d02ea34220bf72 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5988/e02ca120fa18409a237d312ac7d02ea34220bf72/agent.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1743, 12, 2890, 16, 590, 4672, 3536, 3849, 707, 434, 31811, 5790, 2339, 25444, 8257, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1743, 12, 2890, 16, 590, 4672, 3536, 3849, 707, 434, 31811, 5790, 2339, 25444, 8257, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100...
@type plainSyllable: str @param plainSyllable: syllable without tonal information
@type plainEntity: str @param plainEntity: syllable without tonal information
def getExplicitTone(self, plainSyllable, baseTone): """ Gets the explicit tone for the given plain syllable and base tone.
3393604da067ca36b1da468c96a925bae792aff6 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11128/3393604da067ca36b1da468c96a925bae792aff6/operator.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 17186, 56, 476, 12, 2890, 16, 7351, 10876, 2906, 429, 16, 1026, 56, 476, 4672, 3536, 11881, 326, 5515, 30600, 364, 326, 864, 7351, 29136, 471, 1026, 30600, 18, 2, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 17186, 56, 476, 12, 2890, 16, 7351, 10876, 2906, 429, 16, 1026, 56, 476, 4672, 3536, 11881, 326, 5515, 30600, 364, 326, 864, 7351, 29136, 471, 1026, 30600, 18, 2, -100, -100, -100, ...
""" Searches in the fields of the given related model and returns the result as a simple string to be used by the jQuery Autocomplete plugin """ query = request.GET.get('q', None) app_label = request.GET.get('app_label', None) model_name = request.GET.get('model_name', None) search_fields = request.GET.get('search_fields', None) if search_fields and app_label and model_name and query: def construct_search(field_name): if field_name.startswith('^'): return "%s__istartswith" % field_name[1:] elif field_name.startswith('='): return "%s__iexact" % field_name[1:] elif field_name.startswith('@'): return "%s__search" % field_name[1:] else: return "%s__icontains" % field_name model = models.get_model(app_label, model_name) qs = model._default_manager.all() for bit in query.split(): or_queries = [models.Q(**{construct_search( smart_str(field_name)): smart_str(bit)}) for field_name in search_fields.split(',')] other_qs = QuerySet(model) other_qs.dup_select_related(qs) other_qs = other_qs.filter(reduce(operator.or_, or_queries)) qs = qs & other_qs data = ''.join([u'%s|%s\n' % (f.__unicode__(), f.pk) for f in qs]) return HttpResponse(data) return HttpResponseNotFound()
""" Searches in the fields of the given related model and returns the result as a simple string to be used by the jQuery Autocomplete plugin """ query = request.GET.get('q', None) app_label = request.GET.get('app_label', None) model_name = request.GET.get('model_name', None) search_fields = request.GET.get('search_fields', None) if search_fields and app_label and model_name and query: def construct_search(field_name): if field_name.startswith('^'): return "%s__istartswith" % field_name[1:] elif field_name.startswith('='): return "%s__iexact" % field_name[1:] elif field_name.startswith('@'): return "%s__search" % field_name[1:] else: return "%s__icontains" % field_name model = models.get_model(app_label, model_name) qs = model._default_manager.all() for bit in query.split(): or_queries = [models.Q(**{construct_search( smart_str(field_name)): smart_str(bit)}) for field_name in search_fields.split(',')] other_qs = QuerySet(model) other_qs.dup_select_related(qs) other_qs = other_qs.filter(reduce(operator.or_, or_queries)) qs = qs & other_qs data = ''.join([u'%s|%s\n' % (f.__unicode__(), f.pk) for f in qs]) return HttpResponse(data) return HttpResponseNotFound()
def search(request): """ Searches in the fields of the given related model and returns the result as a simple string to be used by the jQuery Autocomplete plugin """ query = request.GET.get('q', None) app_label = request.GET.get('app_label', None) model_name = request.GET.get('model_name', None) search_fields = request.GET.get('search_fields', None) if search_fields and app_label and model_name and query: def construct_search(field_name): # use different lookup methods depending on the notation if field_name.startswith('^'): return "%s__istartswith" % field_name[1:] elif field_name.startswith('='): return "%s__iexact" % field_name[1:] elif field_name.startswith('@'): return "%s__search" % field_name[1:] else: return "%s__icontains" % field_name model = models.get_model(app_label, model_name) qs = model._default_manager.all() for bit in query.split(): or_queries = [models.Q(**{construct_search( smart_str(field_name)): smart_str(bit)}) for field_name in search_fields.split(',')] other_qs = QuerySet(model) other_qs.dup_select_related(qs) other_qs = other_qs.filter(reduce(operator.or_, or_queries)) qs = qs & other_qs data = ''.join([u'%s|%s\n' % (f.__unicode__(), f.pk) for f in qs]) return HttpResponse(data) return HttpResponseNotFound()
7759d6eb08a7e6568691e648ed314450c8afa48e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11995/7759d6eb08a7e6568691e648ed314450c8afa48e/views.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1623, 12, 2293, 4672, 3536, 5167, 281, 316, 326, 1466, 434, 326, 864, 3746, 938, 471, 1135, 326, 563, 487, 279, 4143, 533, 358, 506, 1399, 635, 326, 11215, 11809, 12760, 1909, 3536, 843,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1623, 12, 2293, 4672, 3536, 5167, 281, 316, 326, 1466, 434, 326, 864, 3746, 938, 471, 1135, 326, 563, 487, 279, 4143, 533, 358, 506, 1399, 635, 326, 11215, 11809, 12760, 1909, 3536, 843,...
if wxPlatform == '__WXGTK__': self.navigationBar = self.GetToolBar() else: self.navigationBar = self.FindWindowByName("NavigationBar")
self.navigationBar = self.FindWindowByName("NavigationBar")
def OnInit(self, model): """ There's a tricky problem here. We need to postpone wiring up OnMove, OnSize, etc. after __init__, otherwise OnMove, etc. will get called before we've had a chance to set the windows size using the value in our model. """ self.model = model
5845268ad32cf037101ba85b25022c508e0700e1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/5845268ad32cf037101ba85b25022c508e0700e1/ChandlerWindow.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2755, 2570, 12, 2890, 16, 938, 4672, 3536, 6149, 1807, 279, 433, 13055, 6199, 2674, 18, 1660, 1608, 358, 1603, 84, 476, 341, 11256, 731, 2755, 7607, 16, 2755, 1225, 16, 5527, 18, 1839, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2755, 2570, 12, 2890, 16, 938, 4672, 3536, 6149, 1807, 279, 433, 13055, 6199, 2674, 18, 1660, 1608, 358, 1603, 84, 476, 341, 11256, 731, 2755, 7607, 16, 2755, 1225, 16, 5527, 18, 1839, ...
connection = self._getConnection( connection )
def addReplica( self, lfns, connection = False ): connection = self._getConnection( connection ) """ Add replica to the catalog """ successful = {} failed = {} for lfn, info in lfns.items(): res = self._checkInfo( info, ['PFN', 'SE'] ) if not res['OK']: failed[lfn] = res['Message'] lfns.pop( lfn ) res = self._addReplicas( lfns, connection = connection ) if not res['OK']: for lfn in lfns.keys(): failed[lfn] = res['Message'] else: failed.update( res['Value']['Failed'] ) successful.update( res['Value']['Successful'] ) return S_OK( {'Successful':successful, 'Failed':failed} )
730c8c8e8acbd225fa02fcc8d670bcbd786db665 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/730c8c8e8acbd225fa02fcc8d670bcbd786db665/FileManagerBase.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 14222, 12, 365, 16, 18594, 2387, 16, 1459, 273, 1083, 262, 30, 3536, 1436, 12335, 358, 326, 6222, 3536, 6873, 273, 2618, 2535, 273, 2618, 364, 328, 4293, 16, 1123, 316, 18594, 2387,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 14222, 12, 365, 16, 18594, 2387, 16, 1459, 273, 1083, 262, 30, 3536, 1436, 12335, 358, 326, 6222, 3536, 6873, 273, 2618, 2535, 273, 2618, 364, 328, 4293, 16, 1123, 316, 18594, 2387,...
yearmin = data[0][0] yearmax = data[-1][0] full_left_y = int(round(a + yearmin * b)) full_right_y = int(round(a + yearmax * b)) valid_count = 0 last_valid = -9999 for i,(x,y) in reversed(list(enumerate(data))): if y is None: continue last_valid = max(last_valid, i) valid_count += 1 if valid_count >= 30: break (a_30,b_30,r2_30) = trend(data[i:]) left_y = int(round(a_30 + (yearmin+i) * b_30)) left_x = -100 + 200*float(i)/(yearmax-yearmin) right_y = int(round(a_30 + (yearmin+last_valid) * b_30)) right_x = -100 + 200*float(last_valid)/(yearmax-yearmin) result = Struct() result.urlfrag = ("-100,100|%d,%d|%.0f,%.0f|%d,%d" % (full_left_y, full_right_y, left_x, right_x, left_y, right_y)) result.b_full = b result.b_short = b_30 result.r2_full = r2 result.r2_short = r2_30
if a is not None: yearmin = data[0][0] yearmax = data[-1][0] full_left_y = int(round(a + yearmin * b)) full_right_y = int(round(a + yearmax * b)) valid_count = 0 last_valid = -9999 for i,(x,y) in reversed(list(enumerate(data))): if y is None: continue last_valid = max(last_valid, i) valid_count += 1 if valid_count >= 30: break (a_30,b_30,r2_30) = trend(data[i:]) left_y = int(round(a_30 + (yearmin+i) * b_30)) left_x = -100 + 200*float(i)/(yearmax-yearmin) right_y = int(round(a_30 + (yearmin+last_valid) * b_30)) right_x = -100 + 200*float(last_valid)/(yearmax-yearmin) result.urlfrag = ("-100,100|%d,%d|%.0f,%.0f|%d,%d" % (full_left_y, full_right_y, left_x, right_x, left_y, right_y)) result.b_full = b result.b_short = b_30 result.r2_full = r2 result.r2_short = r2_30 else: result.b_full = None result.b_short = None result.r2_full = None result.r2_short = None result.urlfrag = "-999|-999|-999|-999"
def trendlines(data): """Return a a triple of (url,slopelong,slopeshort) for the full and 30-year trend lines (url is a fragment).""" # full trend (a,b,r2) = trend(data) yearmin = data[0][0] yearmax = data[-1][0] full_left_y = int(round(a + yearmin * b)) full_right_y = int(round(a + yearmax * b)) # thirty-year trend # Find most recent 30 years of _valid_ data valid_count = 0 last_valid = -9999 # largest index with valid data for i,(x,y) in reversed(list(enumerate(data))): if y is None: continue last_valid = max(last_valid, i) valid_count += 1 if valid_count >= 30: break (a_30,b_30,r2_30) = trend(data[i:]) left_y = int(round(a_30 + (yearmin+i) * b_30)) left_x = -100 + 200*float(i)/(yearmax-yearmin) right_y = int(round(a_30 + (yearmin+last_valid) * b_30)) right_x = -100 + 200*float(last_valid)/(yearmax-yearmin) result = Struct() result.urlfrag = ("-100,100|%d,%d|%.0f,%.0f|%d,%d" % (full_left_y, full_right_y, left_x, right_x, left_y, right_y)) result.b_full = b result.b_short = b_30 result.r2_full = r2 result.r2_short = r2_30 return result
f48f5e90a10c46dc1af612da90b8566080e2297e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6890/f48f5e90a10c46dc1af612da90b8566080e2297e/vischeck.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 268, 20262, 3548, 12, 892, 4672, 3536, 990, 279, 279, 14543, 434, 261, 718, 16, 87, 16884, 292, 932, 16, 87, 16884, 4607, 499, 13, 364, 326, 1983, 471, 5196, 17, 6874, 268, 20262, 2362...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 268, 20262, 3548, 12, 892, 4672, 3536, 990, 279, 279, 14543, 434, 261, 718, 16, 87, 16884, 292, 932, 16, 87, 16884, 4607, 499, 13, 364, 326, 1983, 471, 5196, 17, 6874, 268, 20262, 2362...
return self.entry["thumbnail"]["url"].decode('ascii', 'replace')
thumb = element["thumbnail"] except KeyError: return None if isinstance(thumb, str): return thumb elif isinstance(thumb, unicode): return thumb.decode('ascii', 'replace') try: return thumb["url"].decode('ascii', 'replace')
def getThumbnailURL(self): """Returns a link to the thumbnail of the video. """ self.confirmDBThread() # Try to get the thumbnail specific to the video enclosure videoEnclosure = self.getFirstVideoEnclosure() if videoEnclosure is not None: try: return videoEnclosure["thumbnail"]["url"].decode("ascii", "replace") except KeyError: pass # Try to get any enclosure thumbnail for enclosure in self.entry.enclosures: try: return enclosure["thumbnail"]["url"].decode('ascii', 'replace') except KeyError: pass
ad3bc528acb93eb1885468e21f921d1448fed66c /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12354/ad3bc528acb93eb1885468e21f921d1448fed66c/item.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 16270, 1785, 12, 2890, 4672, 3536, 1356, 279, 1692, 358, 326, 9134, 434, 326, 6191, 18, 3536, 365, 18, 10927, 2290, 3830, 1435, 468, 6161, 358, 336, 326, 9134, 2923, 358, 326, 6191,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 16270, 1785, 12, 2890, 4672, 3536, 1356, 279, 1692, 358, 326, 9134, 434, 326, 6191, 18, 3536, 365, 18, 10927, 2290, 3830, 1435, 468, 6161, 358, 336, 326, 9134, 2923, 358, 326, 6191,...
test_capi2()
if not sys.platform.startswith('java'): test_capi1() test_capi2()
def test_capi2(): try: _testcapi.raise_exception(BadException, 0) except RuntimeError, err: exc, err, tb = sys.exc_info() co = tb.tb_frame.f_code assert co.co_name == "__init__" assert co.co_filename.endswith('test_exceptions.py') co2 = tb.tb_frame.f_back.f_code assert co2.co_name == "test_capi2" else: print "Expected exception"
4b7557c37c2b47b46719f0e60f85e7d9f35da077 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/4b7557c37c2b47b46719f0e60f85e7d9f35da077/test_exceptions.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 71, 2425, 22, 13332, 775, 30, 389, 3813, 71, 2425, 18, 11628, 67, 4064, 12, 6434, 503, 16, 374, 13, 1335, 7265, 16, 393, 30, 3533, 16, 393, 16, 8739, 273, 2589, 18, 10075, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 71, 2425, 22, 13332, 775, 30, 389, 3813, 71, 2425, 18, 11628, 67, 4064, 12, 6434, 503, 16, 374, 13, 1335, 7265, 16, 393, 30, 3533, 16, 393, 16, 8739, 273, 2589, 18, 10075, ...
warn('imap.search() return: %s' % flist[0])
Log().warn('imap.search() return: %s' % flist[0])
def get_message(self, folder, uid): """Return the status and the message body for the given uid in the given folder.
37e720541da9eb3d7c758e4ba3bba9e24328637d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2245/37e720541da9eb3d7c758e4ba3bba9e24328637d/imapbackup.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 2150, 12, 2890, 16, 3009, 16, 4555, 4672, 3536, 990, 326, 1267, 471, 326, 883, 1417, 364, 326, 864, 4555, 316, 326, 864, 3009, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 2150, 12, 2890, 16, 3009, 16, 4555, 4672, 3536, 990, 326, 1267, 471, 326, 883, 1417, 364, 326, 864, 4555, 316, 326, 864, 3009, 18, 2, -100, -100, -100, -100, -100, -100, -100,...
I0 = self.index_set() J0 = other.index_set()
I0 = self.index_object() J0 = other.index_object()
def convolution(self, other): """ Convolves two sequences of the same length (automatically expands the shortest one by extending it by 0 if they have different lengths). If {a_n} and {b_n} are sequences of length N (n=0,1,...,N-1), extended by zero for all n in ZZ, then the convolution is
5d0c7d543a1b24ac0a1d29ac68884870541a3a98 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9890/5d0c7d543a1b24ac0a1d29ac68884870541a3a98/dft.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 26828, 12, 2890, 16, 1308, 4672, 3536, 735, 15416, 3324, 2795, 8463, 434, 326, 1967, 769, 261, 5854, 20271, 4542, 87, 326, 20579, 1245, 635, 27147, 518, 635, 374, 309, 2898, 1240, 3775, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 26828, 12, 2890, 16, 1308, 4672, 3536, 735, 15416, 3324, 2795, 8463, 434, 326, 1967, 769, 261, 5854, 20271, 4542, 87, 326, 20579, 1245, 635, 27147, 518, 635, 374, 309, 2898, 1240, 3775, ...
assert len(self.pendingEntId2ParentId) == 0
assert len(self.parent2ChildIds) == 0
def handleZoneEnter(collisionEntry, self=self, zoneNum=zoneNum): # eat the collisionEntry self.enterZone(zoneNum)
3c00fa597cd01b9c43d742c85f1d921aa688ebfa /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8543/3c00fa597cd01b9c43d742c85f1d921aa688ebfa/DistributedLevel.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1640, 4226, 10237, 12, 12910, 1951, 1622, 16, 365, 33, 2890, 16, 4157, 2578, 33, 3486, 2578, 4672, 468, 20729, 326, 17740, 1622, 365, 18, 2328, 4226, 12, 3486, 2578, 13, 2, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1640, 4226, 10237, 12, 12910, 1951, 1622, 16, 365, 33, 2890, 16, 4157, 2578, 33, 3486, 2578, 4672, 468, 20729, 326, 17740, 1622, 365, 18, 2328, 4226, 12, 3486, 2578, 13, 2, -100, -100, ...
sage: magma.function_call('PolynomialRing', [QQ,2])
sage: magma.function_call('PolynomialRing', [QQ,2])
def function_call(self, function, args=[], params={}, nvals=1): """ Return result of evaluating a Magma function with given input, parameters, and asking for nvals as output.
04917fdfb19c54d3ca3aa421b8fa4cc30257acef /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/04917fdfb19c54d3ca3aa421b8fa4cc30257acef/magma.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 445, 67, 1991, 12, 2890, 16, 445, 16, 833, 22850, 6487, 859, 28793, 290, 4524, 33, 21, 4672, 3536, 2000, 563, 434, 26704, 279, 490, 9454, 445, 598, 864, 810, 16, 1472, 16, 471, 29288, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 445, 67, 1991, 12, 2890, 16, 445, 16, 833, 22850, 6487, 859, 28793, 290, 4524, 33, 21, 4672, 3536, 2000, 563, 434, 26704, 279, 490, 9454, 445, 598, 864, 810, 16, 1472, 16, 471, 29288, ...
def initialize(self):
def setUp(self, request, node, data):
def initialize(self): self['type'] = 'password'
0815ea022f851ae2e2bcb7754a6b6f197e8cd751 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12595/0815ea022f851ae2e2bcb7754a6b6f197e8cd751/widgets.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24292, 12, 2890, 16, 590, 16, 756, 16, 501, 4672, 365, 3292, 723, 3546, 273, 296, 3664, 11, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24292, 12, 2890, 16, 590, 16, 756, 16, 501, 4672, 365, 3292, 723, 3546, 273, 296, 3664, 11, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
seh = process.read_uint( self.get_linear_address('SegFs', 0) )
process = self.get_process() seh = self.get_seh_chain_pointer()
def get_seh_chain(self): """ @rtype: list of tuple( int, int ) @return: List of structured exception handlers. Each SEH is represented as a tuple of two addresses: - Address of the SEH block - Address of the SEH callback function
015d670d265cb2758bfa6d6c506d4c7735582b07 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7261/015d670d265cb2758bfa6d6c506d4c7735582b07/system.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 307, 76, 67, 5639, 12, 2890, 4672, 3536, 632, 86, 723, 30, 225, 666, 434, 3193, 12, 509, 16, 509, 262, 632, 2463, 30, 987, 434, 19788, 1520, 4919, 18, 8315, 3174, 44, 353, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 307, 76, 67, 5639, 12, 2890, 4672, 3536, 632, 86, 723, 30, 225, 666, 434, 3193, 12, 509, 16, 509, 262, 632, 2463, 30, 987, 434, 19788, 1520, 4919, 18, 8315, 3174, 44, 353, ...
x = complex(x, 0)
x = _to_complex(x)
def sin(x): """sin(x) Return the sine of x.""" x = complex(x, 0) real = math.sin(x.real) * math.cosh(x.imag) imag = math.cos(x.real) * math.sinh(x.imag) return complex(real, imag)
d265b40c2f1e228c0cbd532ea68ae4f9079e1876 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6934/d265b40c2f1e228c0cbd532ea68ae4f9079e1876/cmath.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5367, 12, 92, 4672, 3536, 21861, 12, 92, 13, 225, 2000, 326, 272, 558, 434, 619, 12123, 225, 619, 273, 389, 869, 67, 14259, 12, 92, 13, 2863, 273, 4233, 18, 21861, 12, 92, 18, 7688, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5367, 12, 92, 4672, 3536, 21861, 12, 92, 13, 225, 2000, 326, 272, 558, 434, 619, 12123, 225, 619, 273, 389, 869, 67, 14259, 12, 92, 13, 2863, 273, 4233, 18, 21861, 12, 92, 18, 7688, ...
fd, name = tempfile.mkstemp() fileobj = os.fdopen(fd, "w+b") try:
with TemporaryFile("w+") as fileobj:
def test_read_long(self): fd, name = tempfile.mkstemp() fileobj = os.fdopen(fd, "w+b") try: fileobj.write("1,2,abc,4,5,6\r\n") fileobj.seek(0) reader = csv.DictReader(fileobj, fieldnames=["f1", "f2"]) self.assertEqual(next(reader), {"f1": '1', "f2": '2', None: ["abc", "4", "5", "6"]}) finally: fileobj.close() os.unlink(name)
02b1d9a951b66534b2fd77809400d61c745873e5 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/3187/02b1d9a951b66534b2fd77809400d61c745873e5/test_csv.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 896, 67, 5748, 12, 2890, 4672, 598, 22791, 812, 2932, 91, 15, 7923, 487, 17041, 30, 17041, 18, 2626, 2932, 21, 16, 22, 16, 18947, 16, 24, 16, 25, 16, 26, 64, 86, 64, 82, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 896, 67, 5748, 12, 2890, 4672, 598, 22791, 812, 2932, 91, 15, 7923, 487, 17041, 30, 17041, 18, 2626, 2932, 21, 16, 22, 16, 18947, 16, 24, 16, 25, 16, 26, 64, 86, 64, 82, ...
elif u.dis_mode == 32 and not(inst.opr_mode == 64 or REX_W(inst.pfx.rex) == 0):
elif u.dis_mode == 32 or not(inst.opr_mode == 32 and REX_W(inst.pfx.rex) == 0):
def resolve_gpr64(u, inst, gpr_op): """64bit General Purpose Register-Selection.""" if gpr_op in range(OP_rAXr8, OP_rDIr15) : index = (gpr_op - OP_rAXr8) |(REX_B(inst.pfx.rex) << 3) else: index = gpr_op - OP_rAX if inst.opr_mode == 16: return GPR[16][index] elif u.dis_mode == 32 and not(inst.opr_mode == 64 or REX_W(inst.pfx.rex) == 0): return GPR[32][index] return GPR[64][index]
dac0d7b28a954921eec1862b83256e49619ccafb /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11108/dac0d7b28a954921eec1862b83256e49619ccafb/decode.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2245, 67, 75, 683, 1105, 12, 89, 16, 1804, 16, 314, 683, 67, 556, 4672, 3536, 1105, 3682, 9544, 14466, 4150, 5433, 17, 6233, 12123, 309, 314, 683, 67, 556, 316, 225, 1048, 12, 3665, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2245, 67, 75, 683, 1105, 12, 89, 16, 1804, 16, 314, 683, 67, 556, 4672, 3536, 1105, 3682, 9544, 14466, 4150, 5433, 17, 6233, 12123, 309, 314, 683, 67, 556, 316, 225, 1048, 12, 3665, ...
self.parent = weakParent() self._gal = weakGalObj()
self._weakParent = weakParent self._weakGal = weakGalObj
def __init__(self , respObj , weakGalObj , weakParent=None): self._setAttrItems(respObj.items()) if 'entity' in respObj: self._setAttrItems(respObj['entity'].items()) if weakParent is not None: self.parent = weakParent() self._gal = weakGalObj() self.fh = None self._postInit()
3ec4e710c574957aed44062e61443be9b89841b8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/13485/3ec4e710c574957aed44062e61443be9b89841b8/G3Items.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 269, 1718, 2675, 269, 16046, 43, 287, 2675, 269, 16046, 3054, 33, 7036, 4672, 365, 6315, 542, 3843, 3126, 12, 12243, 2675, 18, 3319, 10756, 309, 296, 1096, 11,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 269, 1718, 2675, 269, 16046, 43, 287, 2675, 269, 16046, 3054, 33, 7036, 4672, 365, 6315, 542, 3843, 3126, 12, 12243, 2675, 18, 3319, 10756, 309, 296, 1096, 11,...
Return the q-expansion of self, to as high precision as it is known.
Return the `q`-expansion of self, to as high precision as it is known.
def q_expansion(self, prec=None): r""" Return the q-expansion of self, to as high precision as it is known.
d7c9d61652cabcfb658c91a720989214f7a469bd /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/d7c9d61652cabcfb658c91a720989214f7a469bd/genus0.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1043, 67, 2749, 12162, 12, 2890, 16, 13382, 33, 7036, 4672, 436, 8395, 2000, 326, 1043, 17, 2749, 12162, 434, 365, 16, 358, 487, 3551, 6039, 487, 518, 353, 4846, 18, 2, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1043, 67, 2749, 12162, 12, 2890, 16, 13382, 33, 7036, 4672, 436, 8395, 2000, 326, 1043, 17, 2749, 12162, 434, 365, 16, 358, 487, 3551, 6039, 487, 518, 353, 4846, 18, 2, -100, -100, -10...
def main ():
def on_idle(self, event): self.set_status_msg(self.status_msg) def main():
def on_tick(self, evt): #if option.autotune #tune offset if self.print_status: self.print_stats()
b5a766ad3cf7ace8184a93912fcda3cbfb5c1279 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/10611/b5a766ad3cf7ace8184a93912fcda3cbfb5c1279/gsm_scan.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 603, 67, 6470, 12, 2890, 16, 6324, 4672, 468, 430, 1456, 18, 5854, 352, 7556, 468, 88, 7556, 1384, 225, 309, 365, 18, 1188, 67, 2327, 30, 365, 18, 1188, 67, 5296, 1435, 225, 2, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 603, 67, 6470, 12, 2890, 16, 6324, 4672, 468, 430, 1456, 18, 5854, 352, 7556, 468, 88, 7556, 1384, 225, 309, 365, 18, 1188, 67, 2327, 30, 365, 18, 1188, 67, 5296, 1435, 225, 2, -100,...
repme += "\n**[[:Image:%s]]" % duplicate
repme += "\n**[[:Image:%s]]" % self.convert_to_url(duplicate)
def checkImageDuplicated(self, image): """ Function to check the duplicated images. """ self.image = image duplicateRegex = r'\n\*(?:\[\[:Image:%s\]\] has the following duplicates:|\*\[\[:Image:%s\]\])$' % (self.image, self.image) imagePage = wikipedia.ImagePage(self.site, 'Image:%s' % self.image) wikipedia.output(u'Checking if %s has duplicates...' % image) get_hash = self.site.getUrl('/w/api.php?action=query&format=xml&titles=Image:%s&prop=imageinfo&iiprop=sha1' % self.convert_to_url(self.image)) hash_found_list = re.findall(r'<ii sha1="(.*?)" />', get_hash) if hash_found_list != []: hash_found = hash_found_list[0] else: if imagePage.exists(): raise NoHash('No Hash found in the APIs! Maybe the regex to catch it is wrong or someone has changed the APIs structure.') else: wikipedia.output(u'Image deleted before getting the Hash. Skipping...') return False # Error, we need to skip the page. get_duplicates = self.site.getUrl('/w/api.php?action=query&format=xml&list=allimages&aisha1=%s' % hash_found) duplicates = re.findall(r'<img name="(.*?)".*?/>', get_duplicates) if len(duplicates) > 1: if len(duplicates) == 2: wikipedia.output(u'%s has a duplicate! Reporting it...' % self.image) else: wikipedia.output(u'%s has %s duplicates! Reporting them...' % (self.image, len(duplicates) - 1)) repme = "\n*[[:Image:%s]] has the following duplicates:" % self.image for duplicate in duplicates: if duplicate == self.image: continue # the image itself, not report also this as duplicate repme += "\n**[[:Image:%s]]" % duplicate self.report_image(self.image, self.rep_page, self.com, repme, addings = False, regex = duplicateRegex) return True # Ok - No problem. Let's continue the checking phase
6918b3b10235a13ab06a42c1d70403c4f6526397 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/4404/6918b3b10235a13ab06a42c1d70403c4f6526397/checkimages.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 866, 2040, 19682, 690, 12, 2890, 16, 1316, 4672, 3536, 4284, 358, 866, 326, 16975, 4602, 18, 3536, 365, 18, 2730, 273, 1316, 6751, 6628, 273, 436, 8314, 82, 64, 14, 15599, 7438, 10531, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 866, 2040, 19682, 690, 12, 2890, 16, 1316, 4672, 3536, 4284, 358, 866, 326, 16975, 4602, 18, 3536, 365, 18, 2730, 273, 1316, 6751, 6628, 273, 436, 8314, 82, 64, 14, 15599, 7438, 10531, ...
for i in range(len(document.body)): document.body[i] = document.body[i].replace('\\InsetSpace \\hfill{}', '\\hfill') document.body[i] = document.body[i].replace('\\InsetSpace \\dotfill{}', \ '\\begin_inset ERT\nstatus collapsed\n\n' \ '\\begin_layout Standard\n\n\n\\backslash\n' \ 'dotfill{}\n\\end_layout\n\n\\end_inset\n\n') document.body[i] = document.body[i].replace('\\InsetSpace \\hrulefill{}', \ '\\begin_inset ERT\nstatus collapsed\n\n' \ '\\begin_layout Standard\n\n\n\\backslash\n' \ 'hrulefill{}\n\\end_layout\n\n\\end_inset\n\n')
hfill = re.compile(r'\\hfill') dotfill = re.compile(r'\\dotfill') hrulefill = re.compile(r'\\hrulefill') i = 0 while True: i = find_token(document.body, "\\InsetSpace", i) if i == -1: return if hfill.search(document.body[i]): document.body[i] = \ document.body[i].replace('\\InsetSpace \\hfill{}', '\\hfill') i += 1 continue if dotfill.search(document.body[i]): subst = document.body[i].replace('\\InsetSpace \\dotfill{}', \ '\\begin_inset ERT\nstatus collapsed\n\n' \ '\\begin_layout Standard\n\n\n\\backslash\n' \ 'dotfill{}\n\\end_layout\n\n\\end_inset\n\n') subst = subst.split('\n') document.body[i : i+1] = subst i += len(subst) continue if hrulefill.search(document.body[i]): subst = document.body[i].replace('\\InsetSpace \\hrulefill{}', \ '\\begin_inset ERT\nstatus collapsed\n\n' \ '\\begin_layout Standard\n\n\n\\backslash\n' \ 'hrulefill{}\n\\end_layout\n\n\\end_inset\n\n') subst = subst.split('\n') document.body[i : i+1] = subst i += len(subst) continue i += 1
def revert_hfills(document): ' Revert \\hfill commands ' for i in range(len(document.body)): document.body[i] = document.body[i].replace('\\InsetSpace \\hfill{}', '\\hfill') document.body[i] = document.body[i].replace('\\InsetSpace \\dotfill{}', \ '\\begin_inset ERT\nstatus collapsed\n\n' \ '\\begin_layout Standard\n\n\n\\backslash\n' \ 'dotfill{}\n\\end_layout\n\n\\end_inset\n\n') document.body[i] = document.body[i].replace('\\InsetSpace \\hrulefill{}', \ '\\begin_inset ERT\nstatus collapsed\n\n' \ '\\begin_layout Standard\n\n\n\\backslash\n' \ 'hrulefill{}\n\\end_layout\n\n\\end_inset\n\n')
a97fbaa6858229f4bef56786ffb07688ef5a4d82 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7514/a97fbaa6858229f4bef56786ffb07688ef5a4d82/lyx_1_6.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 15226, 67, 76, 5935, 87, 12, 5457, 4672, 296, 868, 1097, 1736, 76, 5935, 4364, 296, 366, 5935, 273, 283, 18, 11100, 12, 86, 11, 1695, 76, 5935, 6134, 3928, 5935, 273, 283, 18, 11100, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 15226, 67, 76, 5935, 87, 12, 5457, 4672, 296, 868, 1097, 1736, 76, 5935, 4364, 296, 366, 5935, 273, 283, 18, 11100, 12, 86, 11, 1695, 76, 5935, 6134, 3928, 5935, 273, 283, 18, 11100, ...
if bootCode: loaderFile = os.path.join(filetool.root(), os.pardir, "data", "generator", "loader-build.tmpl.js") else: loaderFile = os.path.join(filetool.root(), os.pardir, "data", "generator", "loader-source.tmpl.js") result = filetool.read(loaderFile) result = fillTemplate(vals, result)
template = loadTemplate(bootCode) result = fillTemplate(vals, template)
def packageUrisToJS(): # Translate URI data to JavaScript allUris = [] allUrisSmall = [] for packageId, package in enumerate(packages): packageUris = [] packageUrisSmall = []
3978f530a7d64dbbf06451cc3ce692ba259795e9 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5718/3978f530a7d64dbbf06451cc3ce692ba259795e9/CodeGenerator.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2181, 23900, 774, 6479, 13332, 468, 16820, 3699, 501, 358, 11905, 225, 777, 23900, 273, 5378, 777, 23900, 19187, 273, 5378, 364, 2181, 548, 16, 2181, 316, 4241, 12, 10308, 4672, 2181, 2390...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2181, 23900, 774, 6479, 13332, 468, 16820, 3699, 501, 358, 11905, 225, 777, 23900, 273, 5378, 777, 23900, 19187, 273, 5378, 364, 2181, 548, 16, 2181, 316, 4241, 12, 10308, 4672, 2181, 2390...
print("WARNING: Optional "+file+" loaded when it shouldn't've"+" ("+str(round(dt,2))+"s)")
print("WARNING: Optional file "+file+" loaded when it shouldn't've!"+" ("+str(round(dt,2))+"s)")
def run_tests(argv): clean = 0 printerr = 1 #scan for --help and --clean if len(argv) > 1: for arg in argv: if "--help" in arg: do_help(); exit(0) if "--clean" in arg: clean = 1 if "--error" in arg: if printerr == 0: printerr = 1 if printerr == 1: printerr = 0 print("Starting autotest script") there_dir = os.getcwd() err_ct = 0 ex_ct = 0 first_time = time.time() end_time = 0 #if clean == 1: # print("Go clean?") # determine where the script starts here_dir = os.getcwd() print("Starting from \'"+here_dir+"\'") #determine where we _want_ the script to start if len(argv) > 1: argn = 1 while argn < len(argv): if argv[argn][0] == '-': argn+=1 # ignore anything that looks like a flag else: there_dir = argv[argn] # take the first path and be done with it # poor form but we're not doing much complicated break #locate autotest dirs #autotestdirs = find_autotest_dirs(there_dir) autotestdirs = [] for path, dirs, files in os.walk(there_dir): if "autotest" in dirs: autotestdirs.append(os.path.abspath(os.path.join(path,"autotest"))) #for path in autotestdirs: # print("Found dir: \'"+path+"\'") #locate autotest test files #autotestfiles = find_autotest_files(autotestdirs) autotestfiles = [] for path in autotestdirs: for file in os.listdir(path): if file.startswith("test_") and file.endswith(".glm") and file[0] != '.': autotestfiles.append((path, file)) #for path, file in autotestfiles: # print("Found file: \'"+file+"\'") #build test dirs #for file in autotestfiles: errlist=[] cleanlist=[] for path, file in autotestfiles: err = False slice = file[:-4] currpath = os.getcwd() xpath = os.path.join(path,slice) # path/slice xfile = os.path.join(xpath,file) # path/slice/file.glm if not os.path.exists(xpath): os.mkdir(xpath) if os.path.exists(os.path.join(xpath,"gridlabd.xml")): os.remove(os.path.join(xpath,"gridlabd.xml")) shutil.copy2(os.path.join(path,file), xfile) # path/file to path/slice/file os.chdir(xpath) #print("cwd: "+xpath) # build conf files # moo? #run files with popen #run file with: outfile = open(os.path.join(xpath,"outfile.txt"), "w") errfile = open(os.path.join(xpath,"errfile.txt"), "w") #print("NOTICE: Running \'"+xfile+"\'") start_time = time.time(); rv = subprocess.call(["gridlabd",xfile],stdout=outfile,stderr=errfile) end_time = time.time(); dt = end_time - start_time outfile.close() errfile.close() if os.path.exists(os.path.join(xpath,"gridlabd.xml")): statinfo = os.stat(os.path.join(xpath, "gridlabd.xml")) if(statinfo.st_mtime > start_time): if rv == 0: # didn't succeed if gridlabd.xml exists & updated since runtime rv = 1 # handle results if "err_" in file or "_err" in file: if rv == 0: if "opt_" in file or "_opt" in file: print("WARNING: Optional "+file+" converged when it shouldn't've"+" ("+str(round(dt,2))+"s)") cleanlist.append((path, file)) err = False else: print("ERROR: "+file+" converged when it shouldn't've"+" ("+str(round(dt,2))+"s)") print_error(path, printerr) err_ct += 1 err = True elif rv == 2: print("SUCCESS: "+file+" failed to converge, as planned."+" ("+str(round(dt,2))+"s)") cleanlist.append((path, file)) elif rv == 1: print("EXCEPTION: "+file+" failed to load"+" ("+str(dt)+"s)") print_error(path, printerr) ex_ct += 1 err = True else: print("EXCEPTION: "+file+" unrecognized return value ("+str(rv)+")"+" ("+str(round(dt,2))+"s)") print_error(path, printerr) ex_ct += 1 err = True elif "exc_" in file or "_exc" in file: if rv == 0: if "opt_" in file or "_opt" in file: print("WARNING: Optional "+file+" loaded when it shouldn't've"+" ("+str(round(dt,2))+"s)") cleanlist.append((path, file)) err = False else: print("ERROR: "+file+" loaded when it shouldn't've"+" ("+str(round(dt,2))+"s)") err_ct += 1 err = True elif rv == 1: print("SUCCESS: "+file+" failed to load, as planned"+" ("+str(round(dt,2))+"s)") cleanlist.append((path, file)) else: print("EXCEPTION: "+file+" unrecognized return value ("+str(rv)+")"+" ("+str(round(dt,2))+"s)") print_error(path, printerr) ex_ct += 1 err = True else: if rv == 2: if "opt_" in file or "_opt" in file: print("WARNING: Optional "+file+" failed to converge"+" ("+str(round(dt,2))+"s)") cleanlist.append((path, file)) err = False else: err_ct += 1 print("ERROR: "+file+" failed to converge"+" ("+str(round(dt,2))+"s)") print_error(path, printerr) err = True elif rv == 1: print("EXCEPTION: "+file+" failed to load"+" ("+str(round(dt,2))+"s)") print_error(path, printerr) ex_ct += 1 err = True elif rv == 0: print("SUCCESS: "+file+" converged"+" ("+str(round(dt,2))+"s)") cleanlist.append((path, file)) else: print("EXCEPTION: "+file+": unrecognized return value ("+str(rv)+")"+" ("+str(round(dt,2))+"s)") print_error(path, printerr) ex_ct += 1 err = True if err: # zip target directory errlist.append((path,file)) os.chdir(currpath) #print("cwd: "+currpath) # end autotestfiles loop #cleanup as appropriate #for cleanfile, cleanpath in cleanlist: # for path, dirs, file in os.walk(cleanpath): # print("foo") #end #print("bar") last_time = time.time() dt = last_time - first_time #return success/failure print("Validation detected "+str(err_ct)+" models with errors and "+str(ex_ct)+" models with exceptions in "+str(round(dt,2))+" seconds.") for errpath, errfile in errlist: print(" * "+os.path.join(errpath, errfile)) exit(err_ct+ex_ct)
df94e5b8a9553981a05c88f42b7f6175b9153ec1 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/43/df94e5b8a9553981a05c88f42b7f6175b9153ec1/validate.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 67, 16341, 12, 19485, 4672, 2721, 273, 374, 1172, 370, 273, 404, 468, 9871, 364, 1493, 5201, 471, 1493, 6200, 309, 562, 12, 19485, 13, 405, 404, 30, 364, 1501, 316, 5261, 30, 309...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 67, 16341, 12, 19485, 4672, 2721, 273, 374, 1172, 370, 273, 404, 468, 9871, 364, 1493, 5201, 471, 1493, 6200, 309, 562, 12, 19485, 13, 405, 404, 30, 364, 1501, 316, 5261, 30, 309...
c = Chain2(verbose=True)
c = Chain2(verbose=0)
def install_order(self, req): """ Return the list of distributions which need to be installed. The returned list is given in dependency order, i.e. the distributions can be installed in this order without any package being installed before its dependencies got installed. """ # all requirements necessary for install reqs = self.get_reqs(req)
6a8b09561f5be59b89be57f7cf16f575a7469a14 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7707/6a8b09561f5be59b89be57f7cf16f575a7469a14/t.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3799, 67, 1019, 12, 2890, 16, 1111, 4672, 3536, 2000, 326, 666, 434, 23296, 1492, 1608, 358, 506, 5876, 18, 1021, 2106, 666, 353, 864, 316, 4904, 1353, 16, 277, 18, 73, 18, 326, 23296,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3799, 67, 1019, 12, 2890, 16, 1111, 4672, 3536, 2000, 326, 666, 434, 23296, 1492, 1608, 358, 506, 5876, 18, 1021, 2106, 666, 353, 864, 316, 4904, 1353, 16, 277, 18, 73, 18, 326, 23296,...
icons = [os.path.join("Icons",name) for name in Icons]
icons = [os.path.join(package_dir, "Icons",name) for name in Icons] txts = [os.path.join(package_dir, name) for name in txt_files]
def get_source_files(self): # returns the .py files, the .txt files, and the icons icons = [os.path.join("Icons",name) for name in Icons] return build_py.get_source_files(self)+txt_files+icons
b34f34c1e4459597214b7dc957cfd2ff910722e7 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/b34f34c1e4459597214b7dc957cfd2ff910722e7/setup.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 3168, 67, 2354, 12, 2890, 4672, 468, 1135, 326, 263, 2074, 1390, 16, 326, 263, 5830, 1390, 16, 471, 326, 17455, 17455, 273, 306, 538, 18, 803, 18, 5701, 12, 5610, 67, 1214, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 3168, 67, 2354, 12, 2890, 4672, 468, 1135, 326, 263, 2074, 1390, 16, 326, 263, 5830, 1390, 16, 471, 326, 17455, 17455, 273, 306, 538, 18, 803, 18, 5701, 12, 5610, 67, 1214, ...
src.writeh('// %s<PREFIX>%s = %s,' % ('ep', name[0].upper() + name[1:], _formatcode(code)))
src.writeh('// %s<PREFIX>%s = %s,' % ('ep', _capname(name), _formatcode(code)))
def render_aemheader(self, src): src.newglue('AEMConstants', False) src.comment('Types, enumerators, properties') src.writeh('enum {') for name, (kind, code) in self.typebyname: self._renderaemconstant('k', name, code, src) src.writeh('};\n') src.writeh('enum {') for name, (kind, data) in self.referencebyname: if kind != kCommand: self._renderaemconstant(kind, name, data, src) src.writeh('};\n') src.comment('Events') knownparams = set() for name, (kind, data) in self.referencebyname: if kind == kCommand: src.writeh('enum {') code = data[0] self._renderaemconstant('ec', name, code[:4], src) self._renderaemconstant('ei', name, code[4:], src) params = data[1].items() params.sort() for name, code in params: if name in knownparams: src.writeh('// %s<PREFIX>%s = %s,' % ('ep', name[0].upper() + name[1:], _formatcode(code))) else: knownparams.add(name) self._renderaemconstant('ep', name, code, src) src.writeh('};\n')
4f805079f12ced16aef2a53d56a693223625d575 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/10961/4f805079f12ced16aef2a53d56a693223625d575/osaglue.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1743, 67, 69, 351, 3374, 12, 2890, 16, 1705, 4672, 1705, 18, 2704, 7043, 344, 2668, 37, 3375, 2918, 2187, 1083, 13, 1705, 18, 3469, 2668, 2016, 16, 3557, 3062, 16, 1790, 6134, 1705, 18...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1743, 67, 69, 351, 3374, 12, 2890, 16, 1705, 4672, 1705, 18, 2704, 7043, 344, 2668, 37, 3375, 2918, 2187, 1083, 13, 1705, 18, 3469, 2668, 2016, 16, 3557, 3062, 16, 1790, 6134, 1705, 18...
elif os.path.isfile(res):
elif os.path.exists(res):
def get_module_resource(module, *args): """Return the full path of a resource of the given module. @param module: the module @param args: the resource path components @return: absolute path to the resource """ a = get_module_path(module) res = a and opj(a, *args) or False if zipfile.is_zipfile( a +'.zip') : zip = zipfile.ZipFile( a + ".zip") files = ['/'.join(f.split('/')[1:]) for f in zip.namelist()] res = '/'.join(args) if res in files: return opj(a, res) elif os.path.isfile(res): return res return False
312e62c10a681aae2ef7dbf9cc76a831f25be8d0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/312e62c10a681aae2ef7dbf9cc76a831f25be8d0/__init__.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 2978, 67, 3146, 12, 2978, 16, 380, 1968, 4672, 3536, 990, 326, 1983, 589, 434, 279, 1058, 434, 326, 864, 1605, 18, 225, 632, 891, 1605, 30, 326, 1605, 632, 891, 833, 30, 326...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 2978, 67, 3146, 12, 2978, 16, 380, 1968, 4672, 3536, 990, 326, 1983, 589, 434, 279, 1058, 434, 326, 864, 1605, 18, 225, 632, 891, 1605, 30, 326, 1605, 632, 891, 833, 30, 326...
while map:
while client_running and map:
def client_loop(): map = client_map read = asyncore.read write = asyncore.write _exception = asyncore._exception loop_failures = 0 while map: try: # The next two lines intentionally don't use # iterators. Other threads can close dispatchers, causeing # the socket map to shrink. r = e = client_map.keys() w = [fd for (fd, obj) in map.items() if obj.writable()] try: r, w, e = select.select(r, w, e, client_timeout) except select.error, err: if err[0] != errno.EINTR: if err[0] == errno.EBADF: # If a connection is closed while we are # calling select on it, we can get a bad # file-descriptor error. We'll check for this # case by looking for entries in r and w that # are not in the socket map. if [fd for fd in r if fd not in client_map]: continue if [fd for fd in w if fd not in client_map]: continue raise else: continue if not (r or w or e): for obj in client_map.itervalues(): if isinstance(obj, Connection): # Send a heartbeat message as a reply to a # non-existent message id. try: obj.send_reply(-1, None) except DisconnectedError: pass global client_timeout_count client_timeout_count += 1 continue for fd in r: obj = map.get(fd) if obj is None: continue read(obj) for fd in w: obj = map.get(fd) if obj is None: continue write(obj) for fd in e: obj = map.get(fd) if obj is None: continue _exception(obj) except: if map: try: client_logger.critical('The ZEO client loop failed.', exc_info=sys.exc_info()) except: pass for fd, obj in map.items(): if obj is client_trigger: continue try: obj.mgr.client.close() except: map.pop(fd, None) try: client_logger.critical("Couldn't close a dispatcher.", exc_info=sys.exc_info()) except: pass
96933dd70337c5e1b8f2c2666b8e0ed1fd9358a7 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/10048/96933dd70337c5e1b8f2c2666b8e0ed1fd9358a7/connection.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1004, 67, 6498, 13332, 852, 273, 1004, 67, 1458, 225, 855, 273, 4326, 479, 18, 896, 1045, 273, 4326, 479, 18, 2626, 389, 4064, 273, 4326, 479, 6315, 4064, 2798, 67, 26268, 273, 374, 22...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1004, 67, 6498, 13332, 852, 273, 1004, 67, 1458, 225, 855, 273, 4326, 479, 18, 896, 1045, 273, 4326, 479, 18, 2626, 389, 4064, 273, 4326, 479, 6315, 4064, 2798, 67, 26268, 273, 374, 22...
class MSSQLDialect(ansisql.ANSIDialect):
class MSSQLDialect(ansisql.ANSIDialect):
def post_exec(self, engine, proxy, compiled, parameters, **kwargs): """ Turn off the INDENTITY_INSERT mode if it's been activated, and fetch recently inserted IDENTIFY values (works only for one column) """ if getattr(compiled, "isinsert", False): if self.IINSERT: proxy("SET IDENTITY_INSERT %s OFF" % compiled.statement.table.name) self.IINSERT = False elif self.HASIDENT: cursor = proxy("SELECT @@IDENTITY AS lastrowid") row = cursor.fetchone() self._last_inserted_ids = [int(row[0])] # print "LAST ROW ID", self._last_inserted_ids self.HASIDENT = False
4a1a2fec8071124775f05bddcdae855d008db8e8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1074/4a1a2fec8071124775f05bddcdae855d008db8e8/mssql.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1603, 67, 4177, 12, 2890, 16, 4073, 16, 2889, 16, 7743, 16, 1472, 16, 2826, 4333, 4672, 3536, 22425, 3397, 326, 2120, 40, 11101, 67, 11356, 1965, 309, 518, 1807, 2118, 14892, 16, 471, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1603, 67, 4177, 12, 2890, 16, 4073, 16, 2889, 16, 7743, 16, 1472, 16, 2826, 4333, 4672, 3536, 22425, 3397, 326, 2120, 40, 11101, 67, 11356, 1965, 309, 518, 1807, 2118, 14892, 16, 471, ...
def videoEventFilterCb(self, event): self.xlock.acquire() self.overlay.expose() self.xlock.release() return gtk.gdk.FILTER_CONTINUE
def backgroundRealizeCb(self, widget): # Create the video window. self.videoWin = gtk.gdk.Window( self.background.window, self.allocation.width, self.allocation.height, gtk.gdk.WINDOW_CHILD, gtk.gdk.EXPOSURE_MASK, gtk.gdk.INPUT_OUTPUT, "", 0, 0)
df60b4f9844d03743fe59fc739547c98777f8bcf /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5222/df60b4f9844d03743fe59fc739547c98777f8bcf/videowidget.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5412, 6955, 554, 15237, 12, 2890, 16, 3604, 4672, 468, 1788, 326, 6191, 2742, 18, 365, 18, 9115, 18049, 273, 22718, 18, 75, 2883, 18, 3829, 12, 365, 18, 9342, 18, 5668, 16, 365, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5412, 6955, 554, 15237, 12, 2890, 16, 3604, 4672, 468, 1788, 326, 6191, 2742, 18, 365, 18, 9115, 18049, 273, 22718, 18, 75, 2883, 18, 3829, 12, 365, 18, 9342, 18, 5668, 16, 365, 18, ...
method_display = dict(Order.METHOD_CHOICES)[method]
def clean(self): cleaned_data = super(OrderForm, self).clean() method = self.cleaned_data.get('method') if method == Order.METHOD_DELIVERY: address_fields = ['street', 'city', 'state', 'zip'] if not all(cleaned_data.get(field) for field in address_fields): msg = 'You must enter an address for delivery.' raise forms.ValidationError(msg) address = ' '.join(cleaned_data.get(field) for field in address_fields) msg = """We apologize, but it appears you are outside of our delivery area. Please choose one of the other options or call us at %s.""" % self.site.phone try: lon, lat = geocode(address) except: raise forms.ValidationError(msg) point = Point(lon, lat) area = self.site.ordersettings.delivery_area if not area.contains(point): raise forms.ValidationError(msg) lead_time = self.site.ordersettings.delivery_lead_time else: lead_time = self.site.ordersettings.lead_time method_display = dict(Order.METHOD_CHOICES)[method] ready_by = self.cleaned_data.get('ready_by') if ready_by: server_tz = timezone(settings.TIME_ZONE) if ready_by < server_tz.localize(datetime.now() + timedelta(minutes=lead_time)): raise forms.ValidationError('%s orders must be placed %d minutes in advance.' % (method_display, lead_time)) return cleaned_data
327d8a95f1cc048f9c08d4474d37d8219c70079c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9903/327d8a95f1cc048f9c08d4474d37d8219c70079c/forms.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2721, 12, 2890, 4672, 9426, 67, 892, 273, 2240, 12, 2448, 1204, 16, 365, 2934, 6200, 1435, 707, 273, 365, 18, 6200, 329, 67, 892, 18, 588, 2668, 2039, 6134, 309, 707, 422, 4347, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2721, 12, 2890, 4672, 9426, 67, 892, 273, 2240, 12, 2448, 1204, 16, 365, 2934, 6200, 1435, 707, 273, 365, 18, 6200, 329, 67, 892, 18, 588, 2668, 2039, 6134, 309, 707, 422, 4347, 18, ...
assert fc[0]>=0 and fc[1]<v.shape[idim], 'Slice index %i exceeds dimension upper boundary (%i).' % (fc,v.shape[idim]-1)
assert fc[0]>=0, 'Slice index %i lies below dimension lower boundary (0).' % (fc[0],) assert fc[1]<v.shape[idim], 'Slice index %i exceeds dimension upper boundary (%i).' % (fc[1],v.shape[idim]-1)
def getSlice(self,bounds): nc = self.store.getcdf() v = nc.variables[self.varname] dimnames = list(v.dimensions) assert len(bounds)==len(dimnames), 'Number of specified bounds (%i) does not match number of dimensions (%i).' % (len(bounds),len(dimnames)) varslice = self.Slice(dimnames) # Get initial slices, taking into acount specified integer slices and fixed # coordinates, but not float-based slices. boundindices,isfloatslice = [],[] for idim,dimname in enumerate(dimnames): assert bounds[idim].step==None,'Step argument is not yet supported.' fc = self.fixedcoords.get(dimname,None) isfloatslice.append(False) if fc!=None: assert fc[0]>=0 and fc[1]<v.shape[idim], 'Slice index %i exceeds dimension upper boundary (%i).' % (fc,v.shape[idim]-1) boundindices.append(slice(fc[0],fc[1])) elif isinstance(bounds[idim].start,float) or isinstance(bounds[idim].stop,float): boundindices.append(slice(0,v.shape[idim])) isfloatslice[-1] = True else: start,stop,step = bounds[idim].indices(v.shape[idim]) boundindices.append(slice(start,stop)) # Translate slices based on floating point values to slices based on integers. for idim,dimname in enumerate(dimnames): if not isfloatslice[idim]: continue (coords,coords_stag) = self.store.getCoordinates(dimname) if coords==None: return None flstart,flstop = bounds[idim].start,bounds[idim].stop if coords.ndim==1: istart,istop = coords.searchsorted((flstart,flstop)) if istart>0: istart-=1 if istop<len(coords): istop +=1 boundindices[idim] = slice(istart,istop) # Retrieve coordinate values for idim,dimname in enumerate(dimnames): (coords,coords_stag) = self.store.getCoordinates(dimname) if coords==None: return None if coords.ndim==1: start,stop,step = boundindices[idim].indices(v.shape[idim]) varslice.coords [idim] = coords [boundindices[idim]] varslice.coords_stag[idim] = coords_stag[slice(start,stop+1)] else: coorddims = self.store.getCoordinateDimensions(dimname) coordslices = [boundindices[dimnames.index(cd)] for cd in coorddims] coordslices_stag = [slice(s.start,s.stop+1) for s in coordslices] varslice.coords [idim] = coords[tuple(coordslices)] varslice.coords_stag[idim] = coords_stag[tuple(coordslices_stag)]
ac0074fc81e79d18ab4d7c30fab50bef77b2b1a1 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/381/ac0074fc81e79d18ab4d7c30fab50bef77b2b1a1/data.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1322, 2008, 12, 2890, 16, 10576, 4672, 8194, 273, 365, 18, 2233, 18, 588, 24799, 1435, 225, 331, 273, 8194, 18, 7528, 63, 2890, 18, 1401, 529, 65, 2464, 1973, 273, 666, 12, 90, 18, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1322, 2008, 12, 2890, 16, 10576, 4672, 8194, 273, 365, 18, 2233, 18, 588, 24799, 1435, 225, 331, 273, 8194, 18, 7528, 63, 2890, 18, 1401, 529, 65, 2464, 1973, 273, 666, 12, 90, 18, 1...
def zipTest(f, compression, srccontents): zip = zipfile.ZipFile(f, "w", compression) zip.write(srcname, "another"+os.extsep+"name") zip.write(srcname, srcname) zip.close()
TESTFN2 = TESTFN + "2"
def zipTest(f, compression, srccontents): zip = zipfile.ZipFile(f, "w", compression) # Create the ZIP archive zip.write(srcname, "another"+os.extsep+"name") zip.write(srcname, srcname) zip.close() zip = zipfile.ZipFile(f, "r", compression) # Read the ZIP archive readData2 = zip.read(srcname) readData1 = zip.read("another"+os.extsep+"name") zip.close() if readData1 != srccontents or readData2 != srccontents: raise TestFailed, "Written data doesn't equal read data."
f51eab5da1d45826c2db27f13ed719eae8fe5c8a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/f51eab5da1d45826c2db27f13ed719eae8fe5c8a/test_zipfile.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3144, 4709, 12, 74, 16, 9154, 16, 1705, 3980, 4672, 3144, 273, 18823, 18, 29394, 12, 74, 16, 315, 91, 3113, 9154, 13, 282, 468, 1788, 326, 18277, 5052, 3144, 18, 2626, 12, 4816, 529, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3144, 4709, 12, 74, 16, 9154, 16, 1705, 3980, 4672, 3144, 273, 18823, 18, 29394, 12, 74, 16, 315, 91, 3113, 9154, 13, 282, 468, 1788, 326, 18277, 5052, 3144, 18, 2626, 12, 4816, 529, ...
(?<=\n)(?=%| (?= (?=</?%) | (?=\$\{) |
(?<=\n)(?=%|\ (?=\ (?=</?%) | (?=\$\{) |
def __init__(self, **options): super(MyghtyCssLexer, self).__init__(CssLexer, MyghtyLexer, **options)
68250587b0ad3b391049169af2bdf9ae7c23ff82 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6148/68250587b0ad3b391049169af2bdf9ae7c23ff82/templates.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 2826, 2116, 4672, 2240, 12, 49, 4338, 647, 93, 7359, 13356, 16, 365, 2934, 972, 2738, 972, 12, 7359, 13356, 16, 490, 4338, 647, 93, 13356, 16, 2826, 2116...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 2826, 2116, 4672, 2240, 12, 49, 4338, 647, 93, 7359, 13356, 16, 365, 2934, 972, 2738, 972, 12, 7359, 13356, 16, 490, 4338, 647, 93, 13356, 16, 2826, 2116...
dst = os.path.join(os.path.basename(path)) self.LOG.info(' to "%s"' % (dst,))
dst = os.path.join(dst, os.path.basename(path)) self.LOG.info(" to %s" % (pretty_path(dst),))
def guarded(call, *args): "Helper for filesystem calls." self.LOG.debug('%s("%s")' % ( call.__name__, '", "'.join(args), )) if not self.options.dry_run: try: call(*args) except EnvironmentError, exc: self.fatal('%s("%s") failed [%s]' % ( call.__name__, '", "'.join(args), exc, ))
7be5c896b244abd96852611847844ab3d05894ff /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4865/7be5c896b244abd96852611847844ab3d05894ff/rtmv.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3058, 17212, 12, 1991, 16, 380, 1968, 4672, 315, 2276, 364, 6496, 4097, 1199, 365, 18, 4842, 18, 4148, 29909, 87, 27188, 87, 31241, 738, 261, 745, 16186, 529, 972, 16, 19197, 2491, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3058, 17212, 12, 1991, 16, 380, 1968, 4672, 315, 2276, 364, 6496, 4097, 1199, 365, 18, 4842, 18, 4148, 29909, 87, 27188, 87, 31241, 738, 261, 745, 16186, 529, 972, 16, 19197, 2491, 18, ...
-sort: all equal
=sort: all equal
def tabulate(r): """Tabulate sort speed for lists of various sizes. The sizes are 2**i for i in r (the argument, a list). The output displays i, 2**i, and the time to sort arrays of 2**i floating point numbers with the following properties: *sort: random data \sort: descending data /sort: ascending data ~sort: many duplicates -sort: all equal !sort: worst case scenario """ cases = ("*sort", "\\sort", "/sort", "~sort", "-sort", "!sort") fmt = ("%2s %6s" + " %6s"*len(cases)) print fmt % (("i", "2**i") + cases) for i in r: n = 1<<i L = randrange(n) ##assert len(L) == n print "%2d %6d" % (i, n), fl() doit(L) # *sort L.reverse() doit(L) # \sort doit(L) # /sort if n > 4: del L[4:] L = L*(n/4) L = map(lambda x: --x, L) doit(L) # ~sort del L L = map(abs, [-0.5]*n) doit(L) # -sort L = range(n/2-1, -1, -1) L[len(L):] = range(n/2) doit(L) # !sort print
8b6ec79b74284873696b24ab979fb1cb579b86f8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/8b6ec79b74284873696b24ab979fb1cb579b86f8/sortperf.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3246, 6243, 12, 86, 4672, 3536, 5661, 6243, 1524, 8632, 364, 6035, 434, 11191, 8453, 18, 225, 1021, 8453, 854, 576, 636, 77, 364, 277, 316, 436, 261, 5787, 1237, 16, 279, 666, 2934, 22...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3246, 6243, 12, 86, 4672, 3536, 5661, 6243, 1524, 8632, 364, 6035, 434, 11191, 8453, 18, 225, 1021, 8453, 854, 576, 636, 77, 364, 277, 316, 436, 261, 5787, 1237, 16, 279, 666, 2934, 22...
def sync_readline (self, line): assert None == self.select_trigger_log (line)
def sync_stdin (self): self.sync_stderr ('>>> ') line = sys.stdin.readline ()[:-1] if line == '': self.select_trigger ((self.sync_prompt, ())) return
9e54825d994516066d14c27fc5298d0d679a0fa1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2577/9e54825d994516066d14c27fc5298d0d679a0fa1/sync_stdio.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3792, 67, 21772, 261, 2890, 4672, 365, 18, 8389, 67, 11241, 7707, 23012, 8624, 980, 273, 2589, 18, 21772, 18, 896, 1369, 1832, 10531, 17, 21, 65, 309, 980, 422, 875, 30, 365, 18, 4025,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3792, 67, 21772, 261, 2890, 4672, 365, 18, 8389, 67, 11241, 7707, 23012, 8624, 980, 273, 2589, 18, 21772, 18, 896, 1369, 1832, 10531, 17, 21, 65, 309, 980, 422, 875, 30, 365, 18, 4025,...
equivNumericTypes.append(UINT16) equivNumericTypes.append(UINT32)
def eol(n=1): printf("%s" % chr(10) * n)
1b242d94323b90b8ea50866cb613de74bd1ed7ac /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9106/1b242d94323b90b8ea50866cb613de74bd1ed7ac/SD.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 14775, 12, 82, 33, 21, 4672, 12710, 27188, 87, 6, 738, 4513, 12, 2163, 13, 380, 290, 13, 225, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 14775, 12, 82, 33, 21, 4672, 12710, 27188, 87, 6, 738, 4513, 12, 2163, 13, 380, 290, 13, 225, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10...
normalUri = trueUriPrefix + uriSuffix
normalUri = os.path.normpath(os.path.join(trueUriPrefix, uriSuffix))
def normalizeImgUri(uriFromMetafile, trueCombinedUri, combinedUriFromMetafile): # get the "wrong" prefix (in mappedUriPrefix) trueUriPrefix, mappedUriPrefix, sfx = Path.getCommonSuffix(trueCombinedUri, combinedUriFromMetafile) # ...and strip it from contained image uri, to get a correct suffix (in uriSuffix) pre, mappedUriSuffix, uriSuffix = Path.getCommonPrefix(mappedUriPrefix, uriFromMetafile) # ...then compose the correct prefix with the correct suffix normalUri = trueUriPrefix + uriSuffix return normalUri
f091ad523d23e2495476c80e0939b102b6d54921 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5718/f091ad523d23e2495476c80e0939b102b6d54921/Generator.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3883, 12804, 3006, 12, 1650, 1265, 2781, 768, 16, 638, 27994, 3006, 16, 8224, 3006, 1265, 2781, 768, 4672, 468, 336, 326, 315, 21530, 6, 1633, 261, 267, 5525, 3006, 2244, 13, 638, 3006, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3883, 12804, 3006, 12, 1650, 1265, 2781, 768, 16, 638, 27994, 3006, 16, 8224, 3006, 1265, 2781, 768, 4672, 468, 336, 326, 315, 21530, 6, 1633, 261, 267, 5525, 3006, 2244, 13, 638, 3006, ...
class IEEEOperationsTestCase(unittest.TestCase): if float.__getformat__("double").startswith("IEEE"): def test_double_infinity(self): big = 4.8e159 pro = big*big self.assertEquals(repr(pro), 'inf') sqr = big**2 self.assertEquals(repr(sqr), 'inf')
def test_float_specials_do_unpack(self): for fmt, data in [('>f', BE_FLOAT_INF), ('>f', BE_FLOAT_NAN), ('<f', LE_FLOAT_INF), ('<f', LE_FLOAT_NAN)]: struct.unpack(fmt, data)
348dc88097412cc229254f20f2759ce4cd192261 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/348dc88097412cc229254f20f2759ce4cd192261/test_float.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 5659, 67, 9371, 87, 67, 2896, 67, 17309, 12, 2890, 4672, 364, 1325, 16, 501, 316, 306, 2668, 34, 74, 2187, 9722, 67, 15640, 67, 19212, 3631, 7707, 34, 74, 2187, 9722, 67, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 5659, 67, 9371, 87, 67, 2896, 67, 17309, 12, 2890, 4672, 364, 1325, 16, 501, 316, 306, 2668, 34, 74, 2187, 9722, 67, 15640, 67, 19212, 3631, 7707, 34, 74, 2187, 9722, 67, 1...
sage: P1xP1 = ToricVariety(fan) sage: P1xP1.inject_coefficients() The last command does nothing, since ``P1xP1`` is defined over `\QQ`. Let's construct a toric variety over a more complicated field::
def inject_coefficients(self, scope=None, verbose=True): r""" Inject generators of the base field of ``self`` into ``scope``.
bc08271a149642c16f0a0e051f10fcd76466364e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9890/bc08271a149642c16f0a0e051f10fcd76466364e/toric_variety.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4590, 67, 8075, 12344, 12, 2890, 16, 2146, 33, 7036, 16, 3988, 33, 5510, 4672, 436, 8395, 20085, 13327, 434, 326, 1026, 652, 434, 12176, 2890, 10335, 1368, 12176, 4887, 68, 8338, 2, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4590, 67, 8075, 12344, 12, 2890, 16, 2146, 33, 7036, 16, 3988, 33, 5510, 4672, 436, 8395, 20085, 13327, 434, 326, 1026, 652, 434, 12176, 2890, 10335, 1368, 12176, 4887, 68, 8338, 2, -100...
print "ERROR:",result['Message'] print "Setup:", result['Value']['Setup'] print "DIRAC version:",result['Value']['DIRAC'] if result['Value']['Extensions']: for e,v in result['Value']['Extensions'].items(): print "%s version" % e,v else: print "Unknown option:",option def getLog(self,argss):
print "ERROR:", result['Message'] else: print print "Setup:", result['Value']['Setup'] print "DIRAC version:", result['Value']['DIRAC'] if result['Value']['Extensions']: for e, v in result['Value']['Extensions'].items(): print "%s version" % e, v else: print "Unknown option:", option def getLog( self, argss ):
def do_show(self,args): """ Add a record to the File Catalog usage: show software - show components for which software is available show installed - show components installed in the host show status - show status of the installed components show database - show the status of the databases show mysql - show the status of the MySQL server show log <system> <service|agent> show info - show version of software and setup """ argss = args.split() option = argss[0] del argss[0] if option == 'software': client = SystemAdministratorClient(self.host) result = client.getSoftwareComponents() if not result['OK']: print "ERROR:",result['Message'] else: serviceCount = 0 agentCount = 0 for compType in result['Value']: for system in result['Value'][compType]: for component in result['Value'][compType][system]: print compType.ljust(8),system.ljust(28),component.ljust(28) if compType == 'Services': serviceCount += 1 if compType == 'Agents': agentCount += 1 print "Total: %d services, %d agents" % (serviceCount,agentCount) elif option == 'installed': client = SystemAdministratorClient(self.host) result = client.getSetupComponents() if not result['OK']: print "ERROR:",result['Message'] else: serviceCount = 0 agentCount = 0 for compType in result['Value']: for system in result['Value'][compType]: for component in result['Value'][compType][system]: print compType.ljust(8),system.ljust(28),component.ljust(28) if compType == 'Services': serviceCount += 1 if compType == 'Agents': agentCount += 1 print "Total: %d services, %d agents" % (serviceCount,agentCount) elif option == 'status': client = SystemAdministratorClient(self.host) result = client.getOverallStatus() if not result['OK']: print "ERROR:",result['Message'] else: rDict = result['Value'] print " System",' '*20,'Name',' '*5,'Type',' '*23,'Setup Installed Runit Uptime PID' print '-'*116 for compType in rDict: for system in rDict[compType]: for component in rDict[compType][system]: if rDict[compType][system][component]['Installed']: print system.ljust(28),component.ljust(28),compType.lower()[:-1].ljust(7), if rDict[compType][system][component]['Setup']: print 'SetUp'.rjust(12), else: print 'NotSetup'.rjust(12), if rDict[compType][system][component]['Installed']: print 'Installed'.rjust(12), else: print 'NotInstalled'.rjust(12), print str(rDict[compType][system][component]['RunitStatus']).ljust(7), print str(rDict[compType][system][component]['Timeup']).rjust(7), print str(rDict[compType][system][component]['PID']).rjust(8), print elif option == 'database' or option == 'databases': client = SystemAdministratorClient(self.host) result = client.getDatabases(self.rootPwd) if not result['OK']: print "ERROR:",result['Message'] return if result.has_key('MySQLPassword'): self.rootPwd = result['MySQLPassword'] resultSW = client.getSoftwareDatabases() if not resultSW['OK']: print "ERROR:",resultSW['Message'] return sw = resultSW['Value'] installed = result['Value'] print for db in sw: if db in installed: print db.rjust(25),': Installed' else: print db.rjust(25),': Not installed' if not sw: print "No database found" elif option == 'mysql': client = SystemAdministratorClient(self.host) result = client.getMySQLStatus() if not result['OK']: print "ERROR:",result['Message'] elif result['Value']: for par,value in result['Value'].items(): print par.rjust(28),':',value else: print "No MySQL database found" elif option == "log": self.getLog(argss) elif option == "info": client = SystemAdministratorClient(self.host) result = client.getInfo() if not result['OK']: print "ERROR:",result['Message'] print "Setup:", result['Value']['Setup'] print "DIRAC version:",result['Value']['DIRAC'] if result['Value']['Extensions']: for e,v in result['Value']['Extensions'].items(): print "%s version" % e,v else: print "Unknown option:",option
b68208f4ea74e3c61aead9e641ff73054212c46c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/b68208f4ea74e3c61aead9e641ff73054212c46c/SystemAdministratorClientCLI.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 741, 67, 4500, 12, 2890, 16, 1968, 4672, 3536, 1436, 279, 1409, 358, 326, 1387, 14953, 225, 4084, 30, 225, 2405, 17888, 1377, 300, 2405, 4085, 364, 1492, 17888, 353, 2319, 2405, 5876, 37...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 741, 67, 4500, 12, 2890, 16, 1968, 4672, 3536, 1436, 279, 1409, 358, 326, 1387, 14953, 225, 4084, 30, 225, 2405, 17888, 1377, 300, 2405, 4085, 364, 1492, 17888, 353, 2319, 2405, 5876, 37...
while line: if len(line) > 0: self.logger.debug('< %s' % line[:-1]) output.append(line[:-1]) line = runpipe.fromchild.readline() cmdstat = runpipe.poll()
cmdstat = -1 while cmdstat == -1: while line: if len(line) > 0: self.logger.debug('< %s' % line[:-1]) output.append(line[:-1]) line = runpipe.fromchild.readline() cmdstat = runpipe.poll()
def run(self, command): '''Run a command in a pipe dealing with stdout buffer overloads''' self.logger.debug('> %s' % command)
984698527dd76d52bebcf18fb4695f214367928c /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/11867/984698527dd76d52bebcf18fb4695f214367928c/__init__.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 16, 1296, 4672, 9163, 1997, 279, 1296, 316, 279, 6010, 21964, 598, 3909, 1613, 1879, 17135, 26418, 365, 18, 4901, 18, 4148, 2668, 34, 738, 87, 11, 738, 1296, 13, 2, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 16, 1296, 4672, 9163, 1997, 279, 1296, 316, 279, 6010, 21964, 598, 3909, 1613, 1879, 17135, 26418, 365, 18, 4901, 18, 4148, 2668, 34, 738, 87, 11, 738, 1296, 13, 2, -10...
log.info(" compare");
log.debug(" compare");
def compare(self, expect, result): match = {}
035075459b24be1f7ca9d4ad3495e3f6b4ba18d5 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5327/035075459b24be1f7ca9d4ad3495e3f6b4ba18d5/attr.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3400, 12, 2890, 16, 4489, 16, 563, 4672, 845, 273, 2618, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3400, 12, 2890, 16, 4489, 16, 563, 4672, 845, 273, 2618, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
sub_message.SetInParent()
sub_message.SetInParent()
def _MergeField(tokenizer, message): """Merges a single protocol message field into a message. Args: tokenizer: A tokenizer to parse the field name and values. message: A protocol message to record the data. Raises: ParseError: In case of ASCII parsing problems. """ message_descriptor = message.DESCRIPTOR if tokenizer.TryConsume('['): name = [tokenizer.ConsumeIdentifier()] while tokenizer.TryConsume('.'): name.append(tokenizer.ConsumeIdentifier()) name = '.'.join(name) if not message_descriptor.is_extendable: raise tokenizer.ParseErrorPreviousToken( 'Message type "%s" does not have extensions.' % message_descriptor.full_name) field = message.Extensions._FindExtensionByName(name) if not field: raise tokenizer.ParseErrorPreviousToken( 'Extension "%s" not registered.' % name) elif message_descriptor != field.containing_type: raise tokenizer.ParseErrorPreviousToken( 'Extension "%s" does not extend message type "%s".' % ( name, message_descriptor.full_name)) tokenizer.Consume(']') else: name = tokenizer.ConsumeIdentifier() field = message_descriptor.fields_by_name.get(name, None) # Group names are expected to be capitalized as they appear in the # .proto file, which actually matches their type names, not their field # names. if not field: field = message_descriptor.fields_by_name.get(name.lower(), None) if field and field.type != descriptor.FieldDescriptor.TYPE_GROUP: field = None if (field and field.type == descriptor.FieldDescriptor.TYPE_GROUP and field.message_type.name != name): field = None if not field: raise tokenizer.ParseErrorPreviousToken( 'Message type "%s" has no field named "%s".' % ( message_descriptor.full_name, name)) if field.cpp_type == descriptor.FieldDescriptor.CPPTYPE_MESSAGE: tokenizer.TryConsume(':') if tokenizer.TryConsume('<'): end_token = '>' else: tokenizer.Consume('{') end_token = '}' if field.label == descriptor.FieldDescriptor.LABEL_REPEATED: if field.is_extension: sub_message = message.Extensions[field].add() else: sub_message = getattr(message, field.name).add() else: if field.is_extension: sub_message = message.Extensions[field] else: sub_message = getattr(message, field.name) sub_message.SetInParent() while not tokenizer.TryConsume(end_token): if tokenizer.AtEnd(): raise tokenizer.ParseErrorPreviousToken('Expected "%s".' % (end_token)) _MergeField(tokenizer, sub_message) else: _MergeScalarField(tokenizer, message, field)
949e5f8b62bdd7cf8dbfc47dfeb690d77a004e86 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10745/949e5f8b62bdd7cf8dbfc47dfeb690d77a004e86/text_format.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 6786, 974, 12, 2316, 1824, 16, 883, 4672, 3536, 18725, 279, 2202, 1771, 883, 652, 1368, 279, 883, 18, 225, 6634, 30, 10123, 30, 432, 10123, 358, 1109, 326, 652, 508, 471, 924, 18,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 6786, 974, 12, 2316, 1824, 16, 883, 4672, 3536, 18725, 279, 2202, 1771, 883, 652, 1368, 279, 883, 18, 225, 6634, 30, 10123, 30, 432, 10123, 358, 1109, 326, 652, 508, 471, 924, 18,...
__all__ = ['dict', 'inet_pton', 'inet_ntop', 'isinstance']
__all__ = ['dict', 'inet_pton', 'inet_ntop', 'isinstance', 'True', 'False', 'bool']
def isinstance(object, class_or_type_or_tuple): if type(class_or_type_or_tuple) == types.TupleType: for t in class_or_type_or_tuple: if __builtin__.isinstance(object, t): return 1 return 0 else: return __builtin__.isinstance(object, class_or_type_or_tuple)
f64711af541c49c019649b0510030fd740ad27e0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12595/f64711af541c49c019649b0510030fd740ad27e0/compat.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1549, 12, 1612, 16, 667, 67, 280, 67, 723, 67, 280, 67, 8052, 4672, 309, 618, 12, 1106, 67, 280, 67, 723, 67, 280, 67, 8052, 13, 422, 1953, 18, 9038, 559, 30, 364, 268, 316, 667, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1549, 12, 1612, 16, 667, 67, 280, 67, 723, 67, 280, 67, 8052, 4672, 309, 618, 12, 1106, 67, 280, 67, 723, 67, 280, 67, 8052, 13, 422, 1953, 18, 9038, 559, 30, 364, 268, 316, 667, ...
None,
# typedef void (*lc_callback)(uint32_t, uint32_t, uint32_t, void*);
252469d959d952d07c6c2b796be96c97a8628abc /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/3138/252469d959d952d07c6c2b796be96c97a8628abc/libconcord.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 468, 618, 536, 918, 261, 14, 17704, 67, 3394, 21433, 11890, 1578, 67, 88, 16, 2254, 1578, 67, 88, 16, 2254, 1578, 67, 88, 16, 918, 14, 1769, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 468, 618, 536, 918, 261, 14, 17704, 67, 3394, 21433, 11890, 1578, 67, 88, 16, 2254, 1578, 67, 88, 16, 2254, 1578, 67, 88, 16, 918, 14, 1769, 2, -100, -100, -100, -100, -100, -100, -100, -1...
result=os.popen('exiftool.exe -n -GPSLatitude -GPSLatitudeRef "%s" ' % self.picPath).read().split("\n")
result=os.popen('%s -n -GPSLatitude -GPSLatitudeRef "%s" ' % (self.exifcmd, self.picPath)).read().split("\n")
def readLatitude(self): """read the latitute tag is available and return a float""" result=os.popen('exiftool.exe -n -GPSLatitude -GPSLatitudeRef "%s" ' % self.picPath).read().split("\n") print result if len(result)>1: latitude=float(result[0].split(":")[1]) print "latitude= ",latitude return latitude else: return "None"
e7a9683d9553c4ca742eb3b4829d1cdc566d9a89 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/11453/e7a9683d9553c4ca742eb3b4829d1cdc566d9a89/geoexif.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 855, 16890, 12, 2890, 4672, 3536, 896, 326, 2516, 305, 624, 1047, 353, 2319, 471, 327, 279, 1431, 8395, 563, 33, 538, 18, 84, 3190, 29909, 87, 300, 82, 300, 9681, 4559, 270, 3540, 300,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 855, 16890, 12, 2890, 4672, 3536, 896, 326, 2516, 305, 624, 1047, 353, 2319, 471, 327, 279, 1431, 8395, 563, 33, 538, 18, 84, 3190, 29909, 87, 300, 82, 300, 9681, 4559, 270, 3540, 300,...
if data is None: data = self._getDefaultValues(isGroup)
defaults = self._getDefaultValues(isGroup) if not data: data = {} for key, val in defaults.items(): if not data.has_key(key): data[key] = val
def getPropertiesForUser(self, user, request=None): """Get property values for a user or group. Returns a dictionary of values or a PropertySheet.
3120e0574b472e701f43bc9172afffe37d2eb4ab /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12230/3120e0574b472e701f43bc9172afffe37d2eb4ab/property.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 9392, 19894, 12, 2890, 16, 729, 16, 590, 33, 7036, 4672, 3536, 967, 1272, 924, 364, 279, 729, 578, 1041, 18, 2860, 279, 3880, 434, 924, 578, 279, 4276, 8229, 18, 2, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 9392, 19894, 12, 2890, 16, 729, 16, 590, 33, 7036, 4672, 3536, 967, 1272, 924, 364, 279, 729, 578, 1041, 18, 2860, 279, 3880, 434, 924, 578, 279, 4276, 8229, 18, 2, -100, -100, -100, ...
self.fields = dict.fromkeys(FIELDS, "")
self.fields = dict.fromkeys(FIELDS, '')
def __init__(self, entry=None, **kwargs): """If `entry` is given, copy all fields from `entry`. Any keyword arguments are regarded as field values, and are stored if no other value has been given""" self.parent = None self.fields = dict.fromkeys(FIELDS, "") if entry: # copy constructor self.fields.update(entry) for k in kwargs: self.setdefault(k, kwargs[k])
914907c5f1149ba5eb599c36f071c23d079f8e90 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2215/914907c5f1149ba5eb599c36f071c23d079f8e90/phonebook.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 1241, 33, 7036, 16, 2826, 4333, 4672, 3536, 2047, 1375, 4099, 68, 353, 864, 16, 1610, 777, 1466, 628, 1375, 4099, 8338, 5502, 4932, 1775, 854, 960, 17212, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 1241, 33, 7036, 16, 2826, 4333, 4672, 3536, 2047, 1375, 4099, 68, 353, 864, 16, 1610, 777, 1466, 628, 1375, 4099, 8338, 5502, 4932, 1775, 854, 960, 17212, ...
"""
TESTS:: sage: digraphs.DeBruijn(5,0) De Bruijn digraph (n=5, k=0): Looped multi-digraph on 1 vertex sage: digraphs.DeBruijn(0,0) De Bruijn digraph (n=0, k=0): Looped multi-digraph on 0 vertices """
def DeBruijn(self,n,k): r""" Returns the De Bruijn diraph with parameters `n,k`.
61bbe1cebccced5851e1137f8603db211fa85fa2 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/61bbe1cebccced5851e1137f8603db211fa85fa2/graph_generators.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1505, 19114, 4881, 78, 82, 12, 2890, 16, 82, 16, 79, 4672, 436, 8395, 2860, 326, 1505, 605, 86, 4881, 78, 82, 4314, 1483, 598, 1472, 1375, 82, 16, 79, 8338, 2, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1505, 19114, 4881, 78, 82, 12, 2890, 16, 82, 16, 79, 4672, 436, 8395, 2860, 326, 1505, 605, 86, 4881, 78, 82, 4314, 1483, 598, 1472, 1375, 82, 16, 79, 8338, 2, -100, -100, -100, -100...
print `cf.get('Foo Bar', 'foo', raw=1)` print `cf.get('Spacey Bar', 'foo', raw=1)` print `cf.get('Commented Bar', 'foo', raw=1)`
verify(cf.get('Foo Bar', 'foo', raw=1) == 'bar') verify(cf.get('Spacey Bar', 'foo', raw=1) == 'bar') verify(cf.get('Commented Bar', 'foo', raw=1) == 'bar')
def basic(src): print print "Testing basic accessors..." cf = ConfigParser.ConfigParser() sio = StringIO.StringIO(src) cf.readfp(sio) L = cf.sections() L.sort() print L for s in L: print "%s: %s" % (s, cf.options(s)) # The use of spaces in the section names serves as a regression test for # SourceForge bug #115357. # http://sourceforge.net/bugs/?func=detailbug&group_id=5470&bug_id=115357 print `cf.get('Foo Bar', 'foo', raw=1)` print `cf.get('Spacey Bar', 'foo', raw=1)` print `cf.get('Commented Bar', 'foo', raw=1)` if '__name__' in cf.options("Foo Bar"): print '__name__ "option" should not be exposed by the API!' else: print '__name__ "option" properly hidden by the API.' # Make sure the right things happen for remove_option(); # added to include check for SourceForge bug #123324: if not cf.remove_option('Foo Bar', 'foo'): raise TestFailed( "remove_option() failed to report existance of option") if cf.has_option('Foo Bar', 'foo'): raise TestFailed("remove_option() failed to remove option") if cf.remove_option('Foo Bar', 'foo'): raise TestFailed( "remove_option() failed to report non-existance of option" " that was removed") try: cf.remove_option('No Such Section', 'foo') except ConfigParser.NoSectionError: pass else: raise TestFailed( "remove_option() failed to report non-existance of option" " that never existed")
0e2e6d4a7110b251ed7f8e7e8919a5e6d67290cf /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/0e2e6d4a7110b251ed7f8e7e8919a5e6d67290cf/test_cfgparser.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5337, 12, 4816, 4672, 1172, 1172, 315, 22218, 5337, 28088, 7070, 6080, 273, 25076, 18, 809, 2678, 1435, 272, 1594, 273, 15777, 18, 780, 4294, 12, 4816, 13, 6080, 18, 896, 7944, 12, 87, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5337, 12, 4816, 4672, 1172, 1172, 315, 22218, 5337, 28088, 7070, 6080, 273, 25076, 18, 809, 2678, 1435, 272, 1594, 273, 15777, 18, 780, 4294, 12, 4816, 13, 6080, 18, 896, 7944, 12, 87, ...
self._update_window_title()
self._update_window_title()
def on_select_tag(self, widget, row=None, col=None): #When you clic on a tag, you want to unselect the tasks self.task_tv.get_selection().unselect_all() self.ctask_tv.get_selection().unselect_all() task_model = self.task_tv.get_model() self.task_modelfilter.refilter() self._update_window_title()
3de01fe23181b3554804a97d98603435ca30e1cc /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8234/3de01fe23181b3554804a97d98603435ca30e1cc/browser.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 603, 67, 4025, 67, 2692, 12, 2890, 16, 3604, 16, 1027, 33, 7036, 16, 645, 33, 7036, 4672, 468, 9434, 1846, 927, 335, 603, 279, 1047, 16, 1846, 2545, 358, 640, 4025, 326, 4592, 365, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 603, 67, 4025, 67, 2692, 12, 2890, 16, 3604, 16, 1027, 33, 7036, 16, 645, 33, 7036, 4672, 468, 9434, 1846, 927, 335, 603, 279, 1047, 16, 1846, 2545, 358, 640, 4025, 326, 4592, 365, 1...
if self.__version == 7:
if self.__version >= 7:
def get_msvc_paths(self, path, platform='x86'): """Get a list of devstudio directories (include, lib or path).
89ddf73fbbc39e1209c66ea797dbe4e4fb5ee74f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/89ddf73fbbc39e1209c66ea797dbe4e4fb5ee74f/msvccompiler.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 959, 4227, 67, 4481, 12, 2890, 16, 589, 16, 4072, 2218, 92, 5292, 11, 4672, 3536, 967, 279, 666, 434, 4461, 334, 4484, 6402, 261, 6702, 16, 2561, 578, 589, 2934, 2, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 959, 4227, 67, 4481, 12, 2890, 16, 589, 16, 4072, 2218, 92, 5292, 11, 4672, 3536, 967, 279, 666, 434, 4461, 334, 4484, 6402, 261, 6702, 16, 2561, 578, 589, 2934, 2, -100, -1...
for j in xrange(4):
for j in xrange(4) :
def setup(): global CylList, GridList, CapList, CubeList, solidCubeList global sphereList listbase = glGenLists(numSphereSizes + 4) for i in range(numSphereSizes): sphereList += [listbase+i] glNewList(sphereList[i], GL_COMPILE) glBegin(GL_TRIANGLES) ocdec = getSphereTriangles(i) for tri in ocdec: glNormal3fv(tri[0]) glVertex3fv(tri[0]) glNormal3fv(tri[1]) glVertex3fv(tri[1]) glNormal3fv(tri[2]) glVertex3fv(tri[2]) glEnd() glEndList() CylList = listbase+numSphereSizes glNewList(CylList, GL_COMPILE) glBegin(GL_TRIANGLES) for i in range(len(circle)): glNormal3fv(circlen[i][0]) glVertex3fv(circle[i][0]) glNormal3fv(circlen[i][1]) glVertex3fv(circle[i][1]) glNormal3fv(circlen[i][2]) glVertex3fv(circle[i][2]) glEnd() glEndList() CapList = CylList + 1 glNewList(CapList, GL_COMPILE) glNormal3fv(cap0n) glBegin(GL_POLYGON) for p in drum0: glVertex3fv(p) glEnd() glNormal3fv(cap1n) glBegin(GL_POLYGON) for p in drum1: glVertex3fv(p) glEnd() glEndList() GridList = CapList + 1 glNewList(GridList, GL_COMPILE) glBegin(GL_LINES) for p in digrid: glVertex(p[0]) glVertex(p[1]) glEnd() glEndList() CubeList = GridList + 1 glNewList(CubeList, GL_COMPILE) glBegin(GL_QUAD_STRIP) glVertex((-1,-1,-1)) glVertex(( 1,-1,-1)) glVertex((-1, 1,-1)) glVertex(( 1, 1,-1)) glVertex((-1, 1, 1)) glVertex(( 1, 1, 1)) glVertex((-1,-1, 1)) glVertex(( 1,-1, 1)) glVertex((-1,-1,-1)) glVertex(( 1,-1,-1)) glEnd() glEndList() solidCubeList = CubeList + 1 glNewList(solidCubeList, GL_COMPILE) glBegin(GL_QUADS) for i in xrange(len(cubeIndices)): for j in xrange(4): nTuple = tuple(cubeNormals[cubeIndices[i][j]]) vTuple = tuple(cubeVertices[cubeIndices[i][j]]) glNormal3fv(nTuple) glVertex3fv(vTuple) glEnd() glEndList()
8261a6068dcdeb9e7076f9336f1a12251d792b6d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/8261a6068dcdeb9e7076f9336f1a12251d792b6d/drawer.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3875, 13332, 2552, 22337, 80, 682, 16, 7145, 682, 16, 11200, 682, 16, 385, 4895, 682, 16, 18035, 39, 4895, 682, 2552, 20041, 682, 225, 666, 1969, 273, 5118, 7642, 7432, 12, 2107, 22747, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3875, 13332, 2552, 22337, 80, 682, 16, 7145, 682, 16, 11200, 682, 16, 385, 4895, 682, 16, 18035, 39, 4895, 682, 2552, 20041, 682, 225, 666, 1969, 273, 5118, 7642, 7432, 12, 2107, 22747, ...
res = RecapDecompose(m)
res = RecapDecompose(m,onlyUseReactions=[2])
def testEsterRxn(self): m = Chem.MolFromSmiles('C1CC1C(=O)OC1OC1') res = RecapDecompose(m) self.failUnless(res) self.failUnless(len(res.GetLeaves())==2) ks = res.GetLeaves().keys() self.failUnless('[*]C(=O)C1CC1' in ks) self.failUnless('[*]OC1CO1' in ks) m = Chem.MolFromSmiles('C1CC1C(=O)CC1OC1') res = RecapDecompose(m) self.failUnless(res) self.failUnless(len(res.GetLeaves())==0)
e793cf1d5a1abc952de6d7bd7feb70328e8a57c4 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9195/e793cf1d5a1abc952de6d7bd7feb70328e8a57c4/Recap.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 41, 8190, 54, 22695, 12, 2890, 4672, 312, 273, 26542, 18, 49, 355, 1265, 9552, 1449, 2668, 39, 21, 6743, 21, 39, 12, 33, 51, 13, 10809, 21, 10809, 21, 6134, 400, 273, 868, 5909...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 41, 8190, 54, 22695, 12, 2890, 4672, 312, 273, 26542, 18, 49, 355, 1265, 9552, 1449, 2668, 39, 21, 6743, 21, 39, 12, 33, 51, 13, 10809, 21, 10809, 21, 6134, 400, 273, 868, 5909...
edges = Set()
edges = set()
def class_tree_graph(bases, linker, context=None, **options): """ Return a `DotGraph` that graphically displays the package hierarchies for the given packages. """ graph = DotGraph('Class Hierarchy for %s' % name_list(bases), body='ranksep=0.3\n', node_defaults={'shape':'box', 'width': 0, 'height': 0}, edge_defaults={'sametail':True, 'dir':'none'}) # Options if options.get('dir', 'TB') != 'TB': # default: top-down graph.body += 'rankdir=%s\n' % options.get('dir', 'TB') # Find all superclasses & subclasses of the given classes. classes = Set(bases) queue = list(bases) for cls in queue: if cls.subclasses not in (None, UNKNOWN): queue.extend(cls.subclasses) classes.update(cls.subclasses) queue = list(bases) for cls in queue: if cls.bases not in (None, UNKNOWN): queue.extend(cls.bases) classes.update(cls.bases) # Add a node for each cls. classes = [d for d in classes if isinstance(d, ClassDoc) if d.pyval is not object] nodes = add_valdoc_nodes(graph, classes, linker, context) # Add an edge for each package/subclass relationship. edges = Set() for cls in classes: for subcls in cls.subclasses: if cls in nodes and subcls in nodes: edges.add((nodes[cls], nodes[subcls])) graph.edges = [DotGraphEdge(src,dst) for (src,dst) in edges] return graph
8ef87f79a31bd78539934474395f14d31c533ab2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11420/8ef87f79a31bd78539934474395f14d31c533ab2/dotgraph.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 667, 67, 3413, 67, 4660, 12, 18602, 16, 28058, 16, 819, 33, 7036, 16, 2826, 2116, 4672, 3536, 2000, 279, 1375, 10412, 4137, 68, 716, 2667, 6478, 22461, 326, 2181, 366, 5148, 606, 364, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 667, 67, 3413, 67, 4660, 12, 18602, 16, 28058, 16, 819, 33, 7036, 16, 2826, 2116, 4672, 3536, 2000, 279, 1375, 10412, 4137, 68, 716, 2667, 6478, 22461, 326, 2181, 366, 5148, 606, 364, ...
('localhost', 8070), my_crazy_app, server_name='localhost')
('0.0.0.0', 8070), my_crazy_app, server_name='www.cherrypy.example')
def my_crazy_app(environ, start_response): status = '200 OK' response_headers = [('Content-type','text/plain')] start_response(status, response_headers) return ['Hello world!\n']
ce80eb18442543ff1f5cbd1c75c6de16be4a6cb9 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/82/ce80eb18442543ff1f5cbd1c75c6de16be4a6cb9/__init__.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3399, 67, 71, 354, 21832, 67, 2910, 12, 28684, 16, 787, 67, 2740, 4672, 1267, 273, 296, 6976, 7791, 11, 766, 67, 2485, 273, 306, 2668, 1350, 17, 723, 17023, 955, 19, 7446, 6134, 65, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3399, 67, 71, 354, 21832, 67, 2910, 12, 28684, 16, 787, 67, 2740, 4672, 1267, 273, 296, 6976, 7791, 11, 766, 67, 2485, 273, 306, 2668, 1350, 17, 723, 17023, 955, 19, 7446, 6134, 65, ...
assert python_cmd('setup.py install')
assert python_cmd('setup.py', 'install')
def _install(tarball): # extracting the tarball tmpdir = tempfile.mkdtemp() log.warn('Extracting in %s', tmpdir) old_wd = os.getcwd() try: os.chdir(tmpdir) tar = tarfile.open(tarball) extractall(tar) tar.close() # going in the directory subdir = os.path.join(tmpdir, os.listdir(tmpdir)[0]) os.chdir(subdir) log.warn('Now working in %s', subdir) # installing log.warn('Installing Distribute') assert python_cmd('setup.py install') finally: os.chdir(old_wd)
269f12efa9ac5640bb9ad4a79c26e77c66f1a950 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/495/269f12efa9ac5640bb9ad4a79c26e77c66f1a950/distribute_setup.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 5425, 12, 88, 23846, 4672, 468, 27117, 326, 29441, 20213, 273, 13275, 18, 24816, 72, 5814, 1435, 613, 18, 8935, 2668, 4976, 310, 316, 738, 87, 2187, 20213, 13, 1592, 67, 3623, 273, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 5425, 12, 88, 23846, 4672, 468, 27117, 326, 29441, 20213, 273, 13275, 18, 24816, 72, 5814, 1435, 613, 18, 8935, 2668, 4976, 310, 316, 738, 87, 2187, 20213, 13, 1592, 67, 3623, 273, ...
'generic_fblas3.pyf'])
'generic_fblas3.pyf', 'interface_gen.py'])
def generate_pyf(extension, build_dir): name = extension.name.split('.')[-1] target = join(build_dir,target_dir,name+'.pyf') if name[0]=='c' and atlas_version is None and newer(__file__,target): f = open(target,'w') f.write('python module '+name+'\n') f.write('usercode void empty_module(void) {}\n') f.write('interface\n') f.write('subroutine empty_module()\n') f.write('intent(c) empty_module\n') f.write('end subroutine empty_module\n') f.write('end interface\nend python module'+name+'\n') f.close() return target if newer_group(extension.depends,target): generate_interface(name, extension.depends[0], target, skip_names[name]) return target
d22345c4009069cb0d5194f7bcc58445d8409832 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12971/d22345c4009069cb0d5194f7bcc58445d8409832/setup_linalg.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2103, 67, 2074, 74, 12, 6447, 16, 1361, 67, 1214, 4672, 508, 273, 2710, 18, 529, 18, 4939, 2668, 1093, 13, 18919, 21, 65, 1018, 273, 1233, 12, 3510, 67, 1214, 16, 3299, 67, 1214, 16,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2103, 67, 2074, 74, 12, 6447, 16, 1361, 67, 1214, 4672, 508, 273, 2710, 18, 529, 18, 4939, 2668, 1093, 13, 18919, 21, 65, 1018, 273, 1233, 12, 3510, 67, 1214, 16, 3299, 67, 1214, 16,...
iter = self.size_units_combo.get_active_iter() units = self.size_units_combo.get_model().get_value(iter, 0)
def change_size_units(self): iter = self.size_units_combo.get_active_iter() units = self.size_units_combo.get_model().get_value(iter, 0) lower = self.__get_num(self.size_lower) upper = self.__get_num(self.size_upper) size = self.__get_num(self.size_new) size_beg_label = self.glade_xml.get_widget('size_beg') size_beg_label.set_text(str(lower)) size_end_label = self.glade_xml.get_widget('size_end') size_end_label.set_text(str(upper)) if self.size_lower < self.size_upper: self.size_scale.set_range(lower, upper) self.size_scale.set_value(size) self.size_entry.set_text(str(size)) self.update_remaining_space_label()
9853486877ffa9129d0f5ea9d5bb5e86f67600dd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3880/9853486877ffa9129d0f5ea9d5bb5e86f67600dd/InputController.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2549, 67, 1467, 67, 7705, 12, 2890, 4672, 2612, 273, 365, 16186, 588, 67, 2107, 12, 2890, 18, 1467, 67, 8167, 13, 3854, 273, 365, 16186, 588, 67, 2107, 12, 2890, 18, 1467, 67, 5797, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2549, 67, 1467, 67, 7705, 12, 2890, 4672, 2612, 273, 365, 16186, 588, 67, 2107, 12, 2890, 18, 1467, 67, 8167, 13, 3854, 273, 365, 16186, 588, 67, 2107, 12, 2890, 18, 1467, 67, 5797, ...
(select count(*) from event_registration r , crm_case c , event_event e where c.section_id=e.section_id and r.case_id=c.id and c.state='draft' and e.type=t.id ) as draft_state , (select count(*) from event_registration r , crm_case c , event_event e where c.section_id=e.section_id and r.case_id=c.id and c.state='open' and e.type=t.id ) as confirm_state,
(select sum(nb_register) from event_registration r , crm_case c , event_event e where c.section_id=e.section_id and r.case_id=c.id and c.state='draft' and e.type=t.id ) as draft_state , (select sum(nb_register) from event_registration r , crm_case c , event_event e where c.section_id=e.section_id and r.case_id=c.id and c.state='open' and e.type=t.id ) as confirm_state,
def init(self, cr): cr.execute(""" create or replace view report_event_type_registration as ( select count(t.id) as id, t.name as name, (select count(*) from event_registration r , crm_case c , event_event e where c.section_id=e.section_id and r.case_id=c.id and c.state='draft' and e.type=t.id ) as draft_state , (select count(*) from event_registration r , crm_case c , event_event e where c.section_id=e.section_id and r.case_id=c.id and c.state='open' and e.type=t.id ) as confirm_state, count(t.id) as nbevent from event_event e inner join crm_case_section c1 on (e.section_id=c1.id) inner join event_type t on (e.type=t.id) group by t.name,t.id
ad2fc47780e68710dea43a85f1c75fdf2ff685f3 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7339/ad2fc47780e68710dea43a85f1c75fdf2ff685f3/event.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1208, 12, 2890, 16, 4422, 4672, 4422, 18, 8837, 2932, 3660, 752, 578, 1453, 1476, 2605, 67, 2575, 67, 723, 67, 14170, 487, 261, 2027, 1056, 12, 88, 18, 350, 13, 487, 612, 16, 268, 18...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1208, 12, 2890, 16, 4422, 4672, 4422, 18, 8837, 2932, 3660, 752, 578, 1453, 1476, 2605, 67, 2575, 67, 723, 67, 14170, 487, 261, 2027, 1056, 12, 88, 18, 350, 13, 487, 612, 16, 268, 18...
cls._parser = etree.XMLParser(**cls._parser_opts) class XMLApplication(XMLElement): __metaclass__ = XMLApplicationType accept_types = [] build_types = []
cls._xml_parser = etree.XMLParser(remove_blank_text=True) class XMLRootElement(XMLElement): __metaclass__ = XMLRootElementType encoding = 'UTF-8' content_type = None
def __init__(cls, name, bases, dct): if cls._xml_schema is None and cls._xml_schema_file is not None: cls._xml_schema = etree.XMLSchema(etree.parse(open(os.path.join(cls._xml_schema_dir, cls._xml_schema_file), 'r'))) if cls._parser is None: if cls._xml_schema is not None and cls._validate_input: cls._parser = etree.XMLParser(schema=cls._xml_schema, **cls._parser_opts) else: cls._parser = etree.XMLParser(**cls._parser_opts)
86a3cb09569495d76ac3ea45504458f43e29e618 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3449/86a3cb09569495d76ac3ea45504458f43e29e618/__init__.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 6429, 16, 508, 16, 8337, 16, 18253, 4672, 309, 2028, 6315, 2902, 67, 4821, 353, 599, 471, 2028, 6315, 2902, 67, 4821, 67, 768, 353, 486, 599, 30, 2028, 6315, 2902,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 6429, 16, 508, 16, 8337, 16, 18253, 4672, 309, 2028, 6315, 2902, 67, 4821, 353, 599, 471, 2028, 6315, 2902, 67, 4821, 67, 768, 353, 486, 599, 30, 2028, 6315, 2902,...
elif data['form']['result_selection']=='customer_supplier':
elif data['form']['result_selection'] == 'customer_supplier':
def _get_partners(self, data): if data['form']['result_selection']=='customer': return 'Receivable Accounts' elif data['form']['result_selection']=='supplier': return 'Payable Accounts' elif data['form']['result_selection']=='customer_supplier': return 'Receivable and Payable Accounts' return ''
fbadca8d34f0137a629f314b5e51a2ecd4c769b8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/fbadca8d34f0137a629f314b5e51a2ecd4c769b8/third_party_ledger.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 588, 67, 2680, 9646, 12, 2890, 16, 501, 4672, 309, 501, 3292, 687, 21712, 2088, 67, 10705, 3546, 18920, 10061, 4278, 327, 296, 4779, 427, 429, 30849, 11, 1327, 501, 3292, 687, 21712...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 588, 67, 2680, 9646, 12, 2890, 16, 501, 4672, 309, 501, 3292, 687, 21712, 2088, 67, 10705, 3546, 18920, 10061, 4278, 327, 296, 4779, 427, 429, 30849, 11, 1327, 501, 3292, 687, 21712...
return [item[1] for item in items]
if len(items) < num: TestFailed("Error: could not find %d words with Spam=%s - only found %d" % (num, get_spam, len(items))) ret = {} for n, word, info in items[:num]: ret[word]=copy.copy(info) return ret
def FindTopWords(bayes, num, get_spam): items = [] try: bayes.db # bsddb style extractor = DBExtractor except AttributeError: extractor = DictExtractor for word, info in extractor(bayes): if ":" in word: continue if get_spam: if info.hamcount==0: items.append((info.spamcount, word)) else: if info.spamcount==0: items.append((info.hamcount, word)) items.sort() return [item[1] for item in items]
1906eaa7cce41636b5054be689664fc95342a8ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6126/1906eaa7cce41636b5054be689664fc95342a8ca/tester.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4163, 3401, 7363, 12, 70, 528, 281, 16, 818, 16, 336, 67, 1752, 301, 4672, 1516, 273, 5378, 775, 30, 324, 528, 281, 18, 1966, 468, 7081, 449, 70, 2154, 14031, 273, 2383, 10958, 1335, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4163, 3401, 7363, 12, 70, 528, 281, 16, 818, 16, 336, 67, 1752, 301, 4672, 1516, 273, 5378, 775, 30, 324, 528, 281, 18, 1966, 468, 7081, 449, 70, 2154, 14031, 273, 2383, 10958, 1335, ...
error(t.lineno, "syntax error at '%s'" % t.value)
error(t.lexer.lineno, "syntax error at '%s'" % t.value)
def p_error(t): if t: error(t.lineno, "syntax error at '%s'" % t.value) else: error(0, "unknown syntax error", True)
009df5ff1e19276433975fddd43a306ff653a20e /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/6787/009df5ff1e19276433975fddd43a306ff653a20e/isa_parser.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 293, 67, 1636, 12, 88, 4672, 309, 268, 30, 555, 12, 88, 18, 31731, 18, 17782, 16, 315, 14308, 555, 622, 1995, 87, 4970, 738, 268, 18, 1132, 13, 469, 30, 555, 12, 20, 16, 315, 8172,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 293, 67, 1636, 12, 88, 4672, 309, 268, 30, 555, 12, 88, 18, 31731, 18, 17782, 16, 315, 14308, 555, 622, 1995, 87, 4970, 738, 268, 18, 1132, 13, 469, 30, 555, 12, 20, 16, 315, 8172,...
module_type = DynamicModule
pass
def load_module(self, fullname): if fullname in sys.modules: return sys.modules[name]
a644edf0515c26ed027522891ccf02aceac764e8 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8659/a644edf0515c26ed027522891ccf02aceac764e8/importer.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1262, 67, 2978, 12, 2890, 16, 13321, 4672, 309, 13321, 316, 2589, 18, 6400, 30, 327, 2589, 18, 6400, 63, 529, 65, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1262, 67, 2978, 12, 2890, 16, 13321, 4672, 309, 13321, 316, 2589, 18, 6400, 30, 327, 2589, 18, 6400, 63, 529, 65, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
self.assertEqual(None, Everything.test_gslist_null_out()) self.assertEqual(None, Everything.test_glist_null_out())
self.assertEqual([], Everything.test_gslist_null_out()) self.assertEqual([], Everything.test_glist_null_out())
def test_out_nullable_list(self): self.assertEqual(None, Everything.test_gslist_null_out()) self.assertEqual(None, Everything.test_glist_null_out())
4b369f8aca980fc6a582094d6648f40fe4af5e9f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8659/4b369f8aca980fc6a582094d6648f40fe4af5e9f/test_everything.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 659, 67, 20247, 67, 1098, 12, 2890, 4672, 365, 18, 11231, 5812, 12, 7036, 16, 26553, 18, 3813, 67, 564, 1098, 67, 2011, 67, 659, 10756, 365, 18, 11231, 5812, 12, 7036, 16, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 659, 67, 20247, 67, 1098, 12, 2890, 4672, 365, 18, 11231, 5812, 12, 7036, 16, 26553, 18, 3813, 67, 564, 1098, 67, 2011, 67, 659, 10756, 365, 18, 11231, 5812, 12, 7036, 16, ...
if os.path.isfile( stopAgentFile ): try: os.unlink( stopAgentFile ) except: pass
agent.am_removeStopAgentFile()
def __checkControlDir( self ): for agentName in self.__agentModules: if not self.__agentModules[ agentName ][ 'running' ]: continue agent = self.__agentModules[ agentName ][ 'instance' ] stopAgentFile = os.path.join( agent.am_getControlDirectory(), 'stop_agent' )
c40ec108aecee6b600974319bf5b570dfaee37d2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/c40ec108aecee6b600974319bf5b570dfaee37d2/AgentReactor.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 1893, 3367, 1621, 12, 365, 262, 30, 364, 4040, 461, 316, 365, 16186, 5629, 7782, 30, 309, 486, 365, 16186, 5629, 7782, 63, 4040, 461, 308, 63, 296, 8704, 11, 308, 30, 1324, 4040,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 1893, 3367, 1621, 12, 365, 262, 30, 364, 4040, 461, 316, 365, 16186, 5629, 7782, 30, 309, 486, 365, 16186, 5629, 7782, 63, 4040, 461, 308, 63, 296, 8704, 11, 308, 30, 1324, 4040,...
self.session.currentDialog.instance.setTitle(_("Downloadale new plugins"))
self.session.currentDialog.instance.setTitle(_("Downloadable new plugins"))
def setTitle(self): if self.type == self.DOWNLOAD: self.session.currentDialog.instance.setTitle(_("Downloadale new plugins")) elif self.type == self.REMOVE: self.session.currentDialog.instance.setTitle(_("Remove plugins"))
398f5f7ff4b4d37ef55e9d49c178133e970fdac4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6652/398f5f7ff4b4d37ef55e9d49c178133e970fdac4/PluginBrowser.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 14109, 12, 2890, 4672, 309, 365, 18, 723, 422, 365, 18, 12711, 7783, 30, 365, 18, 3184, 18, 2972, 6353, 18, 1336, 18, 542, 4247, 24899, 2932, 7109, 429, 394, 4799, 6, 3719, 1327, 365, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 14109, 12, 2890, 4672, 309, 365, 18, 723, 422, 365, 18, 12711, 7783, 30, 365, 18, 3184, 18, 2972, 6353, 18, 1336, 18, 542, 4247, 24899, 2932, 7109, 429, 394, 4799, 6, 3719, 1327, 365, ...
try: config = ic.IC() except ic.error: return {}
kSCPropNetProxiesHTTPEnable = sc.CFStringCreateWithCString(0, "HTTPEnable", 0) kSCPropNetProxiesHTTPProxy = sc.CFStringCreateWithCString(0, "HTTPProxy", 0) kSCPropNetProxiesHTTPPort = sc.CFStringCreateWithCString(0, "HTTPPort", 0) kSCPropNetProxiesHTTPSEnable = sc.CFStringCreateWithCString(0, "HTTPSEnable", 0) kSCPropNetProxiesHTTPSProxy = sc.CFStringCreateWithCString(0, "HTTPSProxy", 0) kSCPropNetProxiesHTTPSPort = sc.CFStringCreateWithCString(0, "HTTPSPort", 0) kSCPropNetProxiesFTPEnable = sc.CFStringCreateWithCString(0, "FTPEnable", 0) kSCPropNetProxiesFTPPassive = sc.CFStringCreateWithCString(0, "FTPPassive", 0) kSCPropNetProxiesFTPPort = sc.CFStringCreateWithCString(0, "FTPPort", 0) kSCPropNetProxiesFTPProxy = sc.CFStringCreateWithCString(0, "FTPProxy", 0) kSCPropNetProxiesGopherEnable = sc.CFStringCreateWithCString(0, "GopherEnable", 0) kSCPropNetProxiesGopherPort = sc.CFStringCreateWithCString(0, "GopherPort", 0) kSCPropNetProxiesGopherProxy = sc.CFStringCreateWithCString(0, "GopherProxy", 0)
def getproxies_internetconfig(): """Return a dictionary of scheme -> proxy server URL mappings.
e5522816be355104a6f7f078bb1538571ac0bcb0 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8125/e5522816be355104a6f7f078bb1538571ac0bcb0/urllib.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 20314, 606, 67, 267, 14726, 1425, 13332, 3536, 990, 279, 3880, 434, 4355, 317, 2889, 1438, 1976, 7990, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 20314, 606, 67, 267, 14726, 1425, 13332, 3536, 990, 279, 3880, 434, 4355, 317, 2889, 1438, 1976, 7990, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
vereq(x.a, None)
verify(not hasattr(x, "a"))
def slots(): if verbose: print "Testing __slots__..." class C0(object): __slots__ = [] x = C0() verify(not hasattr(x, "__dict__")) verify(not hasattr(x, "foo")) class C1(object): __slots__ = ['a'] x = C1() verify(not hasattr(x, "__dict__")) vereq(x.a, None) x.a = 1 vereq(x.a, 1) del x.a vereq(x.a, None) class C3(object): __slots__ = ['a', 'b', 'c'] x = C3() verify(not hasattr(x, "__dict__")) verify(x.a is None) verify(x.b is None) verify(x.c is None) x.a = 1 x.b = 2 x.c = 3 vereq(x.a, 1) vereq(x.b, 2) vereq(x.c, 3)
6b70599450777a8b911f0eff44b18cd22f1c1e1e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/6b70599450777a8b911f0eff44b18cd22f1c1e1e/test_descr.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 12169, 13332, 309, 3988, 30, 1172, 315, 22218, 1001, 18875, 972, 7070, 667, 385, 20, 12, 1612, 4672, 1001, 18875, 972, 273, 5378, 619, 273, 385, 20, 1435, 3929, 12, 902, 3859, 12, 92, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 12169, 13332, 309, 3988, 30, 1172, 315, 22218, 1001, 18875, 972, 7070, 667, 385, 20, 12, 1612, 4672, 1001, 18875, 972, 273, 5378, 619, 273, 385, 20, 1435, 3929, 12, 902, 3859, 12, 92, ...
d = sess.query(Derived).get(1)
d = sess.query(Derived).get('uid1')
def __init__(self, uid, comment): self.uid = uid self.comment = comment
645fa5255d899431d489b89d1c567bce96c1bb4d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1074/645fa5255d899431d489b89d1c567bce96c1bb4d/assorted_eager.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 4555, 16, 2879, 4672, 365, 18, 1911, 273, 4555, 365, 18, 3469, 273, 2879, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 4555, 16, 2879, 4672, 365, 18, 1911, 273, 4555, 365, 18, 3469, 273, 2879, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
self.app = app
self.rootapp = rootapp self.apps = apps
def __init__(self, app): trace("ISAPISimpleHandler.__init__") self.app = app
e499829a1254ecfb9720f04072a7b33abb8a9c05 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5367/e499829a1254ecfb9720f04072a7b33abb8a9c05/isapi_wsgi.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 595, 4672, 2606, 2932, 5127, 2557, 5784, 1503, 16186, 2738, 972, 7923, 365, 18, 2910, 273, 595, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 595, 4672, 2606, 2932, 5127, 2557, 5784, 1503, 16186, 2738, 972, 7923, 365, 18, 2910, 273, 595, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -1...
text = "%s\n%s" % (url, bug.get_description())
text = "%s\n\n%s" % (url, bug.get_description())
def task_added_cb(self, task): url = task.get_title() r = urlparse(url) if r.hostname is None: return
3ae6d4def4c9a9f155c3fd926bb057f545ddf7dd /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7036/3ae6d4def4c9a9f155c3fd926bb057f545ddf7dd/bugzilla.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1562, 67, 9665, 67, 7358, 12, 2890, 16, 1562, 4672, 880, 273, 1562, 18, 588, 67, 2649, 1435, 436, 273, 15185, 12, 718, 13, 309, 436, 18, 10358, 353, 599, 30, 327, 2, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1562, 67, 9665, 67, 7358, 12, 2890, 16, 1562, 4672, 880, 273, 1562, 18, 588, 67, 2649, 1435, 436, 273, 15185, 12, 718, 13, 309, 436, 18, 10358, 353, 599, 30, 327, 2, -100, -100, -100...
b,a = self.canonical_label(proof=True) d,c = other.canonical_label(proof=True)
b,a = self.canonical_label(proof=True, verbosity=verbosity) d,c = other.canonical_label(proof=True, verbosity=verbosity)
def is_isomorphic(self, other, proof=False): """ Tests for isomorphism between self and other. INPUT: proof -- if True, then output is (a,b), where a is a boolean and b is either a map or None. EXAMPLES: sage: D = graphs.DodecahedralGraph() sage: E = D.copy() sage: gamma = SymmetricGroup(20).random_element() sage: E.relabel(gamma) sage: D.is_isomorphic(E) True
53ee1136334ccef3d22fb07eba25731064a97664 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9890/53ee1136334ccef3d22fb07eba25731064a97664/graph.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 353, 67, 291, 362, 18435, 12, 2890, 16, 1308, 16, 14601, 33, 8381, 4672, 3536, 7766, 87, 364, 353, 362, 7657, 6228, 3086, 365, 471, 1308, 18, 225, 12943, 30, 14601, 1493, 309, 1053, 16...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 353, 67, 291, 362, 18435, 12, 2890, 16, 1308, 16, 14601, 33, 8381, 4672, 3536, 7766, 87, 364, 353, 362, 7657, 6228, 3086, 365, 471, 1308, 18, 225, 12943, 30, 14601, 1493, 309, 1053, 16...
self._debug(NIGHTMARE, "JS: jsScript", ver, `script`)
def jsScript (self, script, ver, item): """execute given script with javascript version ver""" self._debug(NIGHTMARE, "JS: jsScript", ver, `script`) assert self.state == 'parse' assert len(self.buf) >= 2 self.js_output = 0 self.js_env.attachListener(self) # start recursive html filter (used by jsProcessData) self.js_html = FilterHtmlParser(self.rules, self.pics, self.url, comments=self.comments, javascript=self.js_filter, level=self.level+1) # execute self.js_env.executeScript(unescape_js(script), ver) self.js_env.detachListener(self) # wait for recursive filter to finish self.jsEndScript(item)
5827a6b2fdf9c1202486a8f36e257a6eab010228 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3948/5827a6b2fdf9c1202486a8f36e257a6eab010228/HtmlParser.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3828, 3651, 261, 2890, 16, 2728, 16, 1924, 16, 761, 4672, 3536, 8837, 864, 2728, 598, 11341, 1177, 1924, 8395, 225, 1815, 365, 18, 2019, 422, 296, 2670, 11, 1815, 562, 12, 2890, 18, 43...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3828, 3651, 261, 2890, 16, 2728, 16, 1924, 16, 761, 4672, 3536, 8837, 864, 2728, 598, 11341, 1177, 1924, 8395, 225, 1815, 365, 18, 2019, 422, 296, 2670, 11, 1815, 562, 12, 2890, 18, 43...
one recipient. Otherwise it will throw an exception (either SMTPSenderRefused, SMTPRecipientsRefused, or SMTPDataError) That is, if this method does not throw an exception, then someone should get your mail. If this method does not throw an exception, it returns a dictionary, with one entry for each recipient that was refused.
one recipient. It returns a dictionary, with one entry for each recipient that was refused. Each entry contains a tuple of the SMTP error code and the accompanying error message sent by the server. This method may raise the following exceptions: SMTPHeloError The server didn't reply properly to the helo greeting. SMTPRecipientsRefused The server rejected for ALL recipients (no mail was sent). SMTPSenderRefused The server didn't accept the from_addr. SMTPDataError The server replied with an unexpected error code (other than a refusal of a recipient). Note: the connection will be open even after an exception is raised.
def sendmail(self, from_addr, to_addrs, msg, mail_options=[], rcpt_options=[]): """This command performs an entire mail transaction.
7ce512fd70a64eb7cf9563bb31dd4291a44254dd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2120/7ce512fd70a64eb7cf9563bb31dd4291a44254dd/smtplib.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1366, 4408, 12, 2890, 16, 628, 67, 4793, 16, 358, 67, 23305, 16, 1234, 16, 4791, 67, 2116, 22850, 6487, 4519, 337, 67, 2116, 33, 8526, 4672, 3536, 2503, 1296, 11199, 392, 7278, 4791, 2...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1366, 4408, 12, 2890, 16, 628, 67, 4793, 16, 358, 67, 23305, 16, 1234, 16, 4791, 67, 2116, 22850, 6487, 4519, 337, 67, 2116, 33, 8526, 4672, 3536, 2503, 1296, 11199, 392, 7278, 4791, 2...
return sgtbx.space_group_info(group=space_group)
elif cmd == "SFAC": return sgtbx.space_group_info(group=space_group)
def convert(file_object): space_group = None for line in file_object: l = line.rstrip().split("!")[0] if (l.startswith("LATT ")): assert space_group is None latt = read_shelx_latt(l) space_group = sgtbx.space_group() if (latt.centric): space_group.expand_inv(sgtbx.tr_vec((0,0,0))) space_group.expand_conventional_centring_type(latt.z) elif (l.startswith("SYMM ")): assert space_group is not None s = sgtbx.rt_mx(l[5:]) space_group.expand_smx(s) return sgtbx.space_group_info(group=space_group)
7c3e85389af6ef4140425b8dde9e68755dda23c9 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/696/7c3e85389af6ef4140425b8dde9e68755dda23c9/shelx_latt_sym_to_space_group_symbol.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1765, 12, 768, 67, 1612, 4672, 3476, 67, 1655, 273, 599, 364, 980, 316, 585, 67, 1612, 30, 328, 273, 980, 18, 86, 6406, 7675, 4939, 2932, 4442, 25146, 20, 65, 309, 261, 80, 18, 17514...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1765, 12, 768, 67, 1612, 4672, 3476, 67, 1655, 273, 599, 364, 980, 316, 585, 67, 1612, 30, 328, 273, 980, 18, 86, 6406, 7675, 4939, 2932, 4442, 25146, 20, 65, 309, 261, 80, 18, 17514...
if translate and boolstr_to_bool(translate):
if not translate or boolstr_to_bool(translate):
def handle_cleaner_option_label(self, label): """<label> element under <option>""" self.option_name = _(getText(label.childNodes)) translate = label.getAttribute('translate') if translate and boolstr_to_bool(translate): self.xlate_cb(self.option_name)
fbd4e92f74df9f7cafdceed376f0e03a64cc63db /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7853/fbd4e92f74df9f7cafdceed376f0e03a64cc63db/CleanerML.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1640, 67, 6200, 264, 67, 3482, 67, 1925, 12, 2890, 16, 1433, 4672, 3536, 32, 1925, 34, 930, 3613, 411, 3482, 2984, 3660, 365, 18, 3482, 67, 529, 273, 389, 12, 588, 1528, 12, 1925, 18...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1640, 67, 6200, 264, 67, 3482, 67, 1925, 12, 2890, 16, 1433, 4672, 3536, 32, 1925, 34, 930, 3613, 411, 3482, 2984, 3660, 365, 18, 3482, 67, 529, 273, 389, 12, 588, 1528, 12, 1925, 18...
thr = threading.Thread(target=self.__scale_pixbuf) thr.setDaemon(True) thr.start()
gobject.idle_add(self.__scale_pixbuf)
def __close(self, loader, *data): if self.stop_loading: return
cbf4721e646898ffb269b94d5b0b6d9b390d9802 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/4764/cbf4721e646898ffb269b94d5b0b6d9b390d9802/albumart.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 4412, 12, 2890, 16, 4088, 16, 380, 892, 4672, 309, 365, 18, 5681, 67, 15174, 30, 327, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 4412, 12, 2890, 16, 4088, 16, 380, 892, 4672, 309, 365, 18, 5681, 67, 15174, 30, 327, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
data, response = self.site().postForm(address, predata, sysop = True) if not response:
response, data = self.site().postForm(address, predata, sysop=True) if response.status == 302 and not data:
def protect(self, edit='sysop', move='sysop', create='sysop', unprotect=False, reason=None, prompt=True, throttle=True): """(Un)protect a wiki page. Requires administrator status.
8c9c334e0fbc791f48effe8c0a3bfb447f814cae /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/4404/8c9c334e0fbc791f48effe8c0a3bfb447f814cae/wikipedia.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 17151, 12, 2890, 16, 3874, 2218, 9499, 556, 2187, 3635, 2218, 9499, 556, 2187, 752, 2218, 9499, 556, 2187, 640, 11815, 386, 33, 8381, 16, 3971, 33, 7036, 16, 6866, 33, 5510, 16, 18304, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 17151, 12, 2890, 16, 3874, 2218, 9499, 556, 2187, 3635, 2218, 9499, 556, 2187, 752, 2218, 9499, 556, 2187, 640, 11815, 386, 33, 8381, 16, 3971, 33, 7036, 16, 6866, 33, 5510, 16, 18304, ...
return _open_with_retry(urllib.request.urlretrieve, *args)
resource = args[0] with support.transient_internet(resource): return urllib.request.urlretrieve(*args)
def urlretrieve(self, *args): return _open_with_retry(urllib.request.urlretrieve, *args)
ee2538beefa10847ed8a190b25a18a5840e334c3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8546/ee2538beefa10847ed8a190b25a18a5840e334c3/test_urllibnet.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 880, 17466, 12, 2890, 16, 380, 1968, 4672, 327, 389, 3190, 67, 1918, 67, 9620, 12, 718, 2941, 18, 2293, 18, 718, 17466, 16, 380, 1968, 13, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 880, 17466, 12, 2890, 16, 380, 1968, 4672, 327, 389, 3190, 67, 1918, 67, 9620, 12, 718, 2941, 18, 2293, 18, 718, 17466, 16, 380, 1968, 13, 2, -100, -100, -100, -100, -100, -100, -100, ...
d = ifp.read() ofp.write(d)
while 1: d = ifp.read(128000) if not d: break ofp.write(d)
def hexbin(inp, out): """(infilename, outfilename) - Decode binhexed file""" ifp = HexBin(inp) finfo = ifp.FInfo if not out: out = ifp.FName if os.name == 'mac': ofss = macfs.FSSpec(out) out = ofss.as_pathname() ofp = open(out, 'wb') # XXXX Do translation on non-mac systems d = ifp.read() ofp.write(d) ofp.close() ifp.close_data() d = ifp.read_rsrc() if d: ofp = openrsrc(out, 'wb') ofp.write(d) ofp.close() if os.name == 'mac': nfinfo = ofss.GetFInfo() nfinfo.Creator = finfo.Creator nfinfo.Type = finfo.Type nfinfo.Flags = finfo.Flags ofss.SetFInfo(nfinfo) ifp.close()
d9300e7ae7f704ad263013ec13346104b4bba7ba /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/d9300e7ae7f704ad263013ec13346104b4bba7ba/binhex.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3827, 4757, 12, 31647, 16, 596, 4672, 3536, 12, 267, 3459, 16, 596, 3459, 13, 300, 6209, 4158, 7118, 329, 585, 8395, 309, 84, 273, 15734, 9913, 12, 31647, 13, 15190, 273, 309, 84, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3827, 4757, 12, 31647, 16, 596, 4672, 3536, 12, 267, 3459, 16, 596, 3459, 13, 300, 6209, 4158, 7118, 329, 585, 8395, 309, 84, 273, 15734, 9913, 12, 31647, 13, 15190, 273, 309, 84, 18, ...