rem
stringlengths
2
226k
add
stringlengths
0
227k
context
stringlengths
8
228k
meta
stringlengths
156
215
input_ids
list
attention_mask
list
labels
list
warn("Tables rebuilt because of dependencies updated: '" \ +"', '".join(dependingTables) + "'")
if not self.quiet: warn("Tables rebuilt because of dependencies updated: '" \ +"', '".join(dependingTables) + "'")
def build(self, tables): """ Builds the given tables.
dec8906e07681ee925caef6e895971c9e8464f07 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11128/dec8906e07681ee925caef6e895971c9e8464f07/build.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1361, 12, 2890, 16, 4606, 4672, 3536, 3998, 87, 326, 864, 4606, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1361, 12, 2890, 16, 4606, 4672, 3536, 3998, 87, 326, 864, 4606, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
from Numeric import array self.atoms.SetCartesianPositions(array(atoms.get_positions()))
from Numeric import array, Float self.atoms.SetCartesianPositions(array(atoms.get_positions(), Float))
def get_forces(self, atoms): from Numeric import array self.atoms.SetCartesianPositions(array(atoms.get_positions())) return npy.array(self.calc.GetCartesianForces())
00836fe7ba7dc98ee8997299b7e06ef0e82cd0bc /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5735/00836fe7ba7dc98ee8997299b7e06ef0e82cd0bc/old.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 1884, 764, 12, 2890, 16, 9006, 4672, 628, 16980, 1930, 526, 16, 5450, 365, 18, 14937, 18, 694, 13006, 13404, 11024, 12, 1126, 12, 14937, 18, 588, 67, 12388, 9334, 5450, 3719, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 1884, 764, 12, 2890, 16, 9006, 4672, 628, 16980, 1930, 526, 16, 5450, 365, 18, 14937, 18, 694, 13006, 13404, 11024, 12, 1126, 12, 14937, 18, 588, 67, 12388, 9334, 5450, 3719, ...
return self._action(*troveSpecs, ActionClass = GroupFindAction)
return self._action(ActionClass = GroupFindAction, *troveSpecs)
def find(self, *troveSpecs): return self._action(*troveSpecs, ActionClass = GroupFindAction)
d5cd416fce497abb2065fa1182b9d65b8242d794 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8747/d5cd416fce497abb2065fa1182b9d65b8242d794/groupsetrecipe.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1104, 12, 2890, 16, 380, 88, 303, 537, 15999, 4672, 327, 365, 6315, 1128, 30857, 88, 303, 537, 15999, 16, 4382, 797, 273, 3756, 3125, 1803, 13, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1104, 12, 2890, 16, 380, 88, 303, 537, 15999, 4672, 327, 365, 6315, 1128, 30857, 88, 303, 537, 15999, 16, 4382, 797, 273, 3756, 3125, 1803, 13, 2, -100, -100, -100, -100, -100, -100, -...
'bytes'."""
'bytes'. 'zinfo_or_arcname' is either a ZipInfo instance or the name of the file in the archive.""" if not isinstance(zinfo_or_arcname, ZipInfo): zinfo = ZipInfo(filename=zinfo_or_arcname, date_time=time.localtime(time.time())) zinfo.compress_type = self.compression else: zinfo = zinfo_or_arcname
def writestr(self, zinfo, bytes): """Write a file into the archive. The contents is the string 'bytes'.""" self._writecheck(zinfo) zinfo.file_size = len(bytes) # Uncompressed size zinfo.CRC = binascii.crc32(bytes) # CRC-32 checksum if zinfo.compress_type == ZIP_DEFLATED: co = zlib.compressobj(zlib.Z_DEFAULT_COMPRESSION, zlib.DEFLATED, -15) bytes = co.compress(bytes) + co.flush() zinfo.compress_size = len(bytes) # Compressed size else: zinfo.compress_size = zinfo.file_size zinfo.header_offset = self.fp.tell() # Start of header bytes self.fp.write(zinfo.FileHeader()) zinfo.file_offset = self.fp.tell() # Start of file bytes self.fp.write(bytes) if zinfo.flag_bits & 0x08: # Write CRC and file sizes after the file data self.fp.write(struct.pack("<lll", zinfo.CRC, zinfo.compress_size, zinfo.file_size)) self.filelist.append(zinfo) self.NameToInfo[zinfo.filename] = zinfo
ea70053892533503a724420f7e62e3c30e16ed64 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/ea70053892533503a724420f7e62e3c30e16ed64/zipfile.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7262, 313, 12, 2890, 16, 998, 1376, 16, 1731, 4672, 3536, 3067, 279, 585, 1368, 326, 5052, 18, 225, 1021, 2939, 353, 326, 533, 296, 3890, 10332, 225, 296, 94, 1376, 67, 280, 67, 11828,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7262, 313, 12, 2890, 16, 998, 1376, 16, 1731, 4672, 3536, 3067, 279, 585, 1368, 326, 5052, 18, 225, 1021, 2939, 353, 326, 533, 296, 3890, 10332, 225, 296, 94, 1376, 67, 280, 67, 11828,...
self.error('Driver letter\n')
self.error('Driver letter required\n')
def do_run(self, options, args): windows = platform == "Windows" if options.unmount:
541006a76a12343a29852dec14beab20966e9909 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5578/541006a76a12343a29852dec14beab20966e9909/fsmount.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 741, 67, 2681, 12, 2890, 16, 702, 16, 833, 4672, 225, 9965, 273, 4072, 422, 315, 10399, 6, 225, 309, 702, 18, 318, 4778, 30, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 741, 67, 2681, 12, 2890, 16, 702, 16, 833, 4672, 225, 9965, 273, 4072, 422, 315, 10399, 6, 225, 309, 702, 18, 318, 4778, 30, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
extra_params = warning_level
extra_params = [warning_level, "-tt"]
def do_matrix_check() : python_versions=("2.3","2.4","2.5","2.6","2.7","3.1","3.2") berkeleydb_versions=("4.1","4.2","4.3","4.4","4.5","4.6","4.7","4.8","5.0") warning_level=(["-Wdefault"], ["-Werror"])[1] import subprocess for py in python_versions : for bdb in berkeleydb_versions : print print "*** Testing bindings for Python %s and Berkeley DB %s" %(py,bdb) extra_params = warning_level # Extra flags for 3.x extra_params += [] if float(py)<=2.999 else ["-bb -tt"] # Extra flags for >=2.6 extra_params += [] if ((float(py)<=2.599) or (float(py)>=2.999)) else ["-3 -tt"] params = extra_params + ["setup.py", "-q", \ "--berkeley-db=/usr/local/BerkeleyDB."+bdb,"build", "-f"] params = ["/usr/local/bin/python"+py] + params print "EXECUTING:", " ".join(params) ret=subprocess.call(params) if ret : print print ">>> Testsuite skipped" print else : params = ["/usr/local/bin/python"+py] + extra_params + ["test.py","-p"] print "EXECUTING:", " ".join(params) subprocess.call(params)
470ffe42527e6b62b77e9d3a4541c3702aa1a497 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1929/470ffe42527e6b62b77e9d3a4541c3702aa1a497/test-full_prerelease.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 741, 67, 5667, 67, 1893, 1435, 294, 5790, 67, 10169, 33, 2932, 22, 18, 23, 15937, 22, 18, 24, 15937, 22, 18, 25, 15937, 22, 18, 26, 15937, 22, 18, 27, 15937, 23, 18, 21, 15937, 23,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 741, 67, 5667, 67, 1893, 1435, 294, 5790, 67, 10169, 33, 2932, 22, 18, 23, 15937, 22, 18, 24, 15937, 22, 18, 25, 15937, 22, 18, 26, 15937, 22, 18, 27, 15937, 23, 18, 21, 15937, 23,...
move_ids.append(act[1])
if isinstance(act[1], (int, long)): move_ids.append(act[1]) else: move_ids.extend(act[1])
def set_inventory_moves(self, cursor, user, shipment_id, name, value, arg, context=None): move_obj = self.pool.get('stock.move')
67c254956f284f804a6e62bd66d638cb1b74207e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9298/67c254956f284f804a6e62bd66d638cb1b74207e/shipment.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 444, 67, 26024, 67, 81, 10829, 12, 2890, 16, 3347, 16, 729, 16, 26990, 67, 350, 16, 508, 16, 460, 16, 1501, 16, 819, 33, 7036, 4672, 3635, 67, 2603, 273, 365, 18, 6011, 18, 588, 26...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 444, 67, 26024, 67, 81, 10829, 12, 2890, 16, 3347, 16, 729, 16, 26990, 67, 350, 16, 508, 16, 460, 16, 1501, 16, 819, 33, 7036, 4672, 3635, 67, 2603, 273, 365, 18, 6011, 18, 588, 26...
self.assertRaises(POSException.VersionError, self._storage.load, oid, 'bogus')
data, revid = self._storage.load(oid, 'bogus') assert pickle.loads(data) == 11
def checkVersionedLoadErrors(self): oid = self._storage.new_oid() version = 'test-version' revid = self._dostore(oid, data=11) revid = self._dostore(oid, revid=revid, data=12, version=version) # Try to load a bogus oid self.assertRaises(KeyError, self._storage.load, self._storage.new_oid(), '') # Try to load a bogus version string self.assertRaises(POSException.VersionError, self._storage.load, oid, 'bogus')
f1ee732285102e57cecb53cb7466f9b967cd3af2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10048/f1ee732285102e57cecb53cb7466f9b967cd3af2/VersionStorage.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 866, 23676, 2563, 4229, 12, 2890, 4672, 7764, 273, 365, 6315, 5697, 18, 2704, 67, 839, 1435, 1177, 273, 296, 3813, 17, 1589, 11, 283, 1246, 273, 365, 6315, 72, 669, 479, 12, 839, 16, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 866, 23676, 2563, 4229, 12, 2890, 4672, 7764, 273, 365, 6315, 5697, 18, 2704, 67, 839, 1435, 1177, 273, 296, 3813, 17, 1589, 11, 283, 1246, 273, 365, 6315, 72, 669, 479, 12, 839, 16, ...
Writes a three dimensional array. The first axis is the step, i.e number of trace, the second is the pixel in that step and the third
Writes a three dimensional array. The first axis is the step, i.e number of trace, the second is the pixel in that step and the third
def __dayplotGetMinMaxValues(self, *args, **kwargs): """ Takes a Stream object and calculates the min and max values for each pixel in the dayplot.
ac0468b3b204e42963e836956a78c9c18fddeffc /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10470/ac0468b3b204e42963e836956a78c9c18fddeffc/waveform.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2881, 4032, 967, 17454, 1972, 12, 2890, 16, 380, 1968, 16, 2826, 4333, 4672, 3536, 23004, 279, 3961, 733, 471, 17264, 326, 1131, 471, 943, 924, 364, 1517, 4957, 316, 326, 2548, 403...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2881, 4032, 967, 17454, 1972, 12, 2890, 16, 380, 1968, 16, 2826, 4333, 4672, 3536, 23004, 279, 3961, 733, 471, 17264, 326, 1131, 471, 943, 924, 364, 1517, 4957, 316, 326, 2548, 403...
>>> check(y.diagonal(0, 1));
>>> check(y.diagonal(0, 0));
def _numarray_tests(): ''' >>> from numpy_ext import * >>> x = new_array() >>> y = x.copy() >>> p = _printer() >>> check = p.check >>> exercise_numarray(x, p) >>> check(y.astype()); >>> check(y.argmax()); >>> check(y.argmax(0)); >>> check(y.argmin()); >>> check(y.argmin(0)); >>> check(y.argsort()); >>> check(y.argsort(1)); >>> y.byteswap(); >>> check(y); >>> check(y.diagonal()); >>> check(y.diagonal(1)); >>> check(y.diagonal(0, 1)); >>> check(y.diagonal(0, 1, 0)); >>> check(y.is_c_array()); >>> check(y.isbyteswapped()); >>> check(y.trace()); >>> check(y.trace(1)); >>> check(y.trace(0, 1)); >>> check(y.trace(0, 1, 0)); >>> check(y.new('D')); >>> y.sort(); >>> check(y); >>> check(y.type()); >>> check(y.factory((1.2, 3.4))); >>> check(y.factory((1.2, 3.4), "Double")); >>> check(y.factory((1.2, 3.4), "Double", (1,2,1))); >>> check(y.factory((1.2, 3.4), "Double", (2,1,1), false)); >>> check(y.factory((1.2, 3.4), "Double", (2,), true, true)); >>> p.results [] >>> del p ''' pass
49d4aac8ec00a0eee4232402a97c1da5cf5560e8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9978/49d4aac8ec00a0eee4232402a97c1da5cf5560e8/numpy.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2107, 1126, 67, 16341, 13332, 9163, 4080, 628, 3972, 67, 408, 1930, 380, 4080, 619, 273, 394, 67, 1126, 1435, 4080, 677, 273, 619, 18, 3530, 1435, 4080, 293, 273, 389, 30439, 1435, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2107, 1126, 67, 16341, 13332, 9163, 4080, 628, 3972, 67, 408, 1930, 380, 4080, 619, 273, 394, 67, 1126, 1435, 4080, 677, 273, 619, 18, 3530, 1435, 4080, 293, 273, 389, 30439, 1435, ...
method='{http://www.opengis.net/wfs}Get'):
srsname=None, method='{http://www.opengis.net/wfs}Get'):
def getfeature(self, typename=None, filter=None, bbox=None, featureid=None, featureversion=None, propertyname=['*'], maxfeatures=None, method='{http://www.opengis.net/wfs}Get'): """Request and return feature data as a file-like object. Parameters ---------- typename : list List of typenames (string) filter : string XML-encoded OGC filter expression. bbox : tuple (left, bottom, right, top) in the feature type's coordinates. featureid : list List of unique feature ids (string) featureversion : string Default is most recent feature version. propertyname : list List of feature property names. '*' matches all. maxfeatures : int Maximum number of features to be returned. method : string Qualified name of the HTTP DCP method to use.
7620dbe8432ef7a70f2e34d4c8f3d2868486374a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11708/7620dbe8432ef7a70f2e34d4c8f3d2868486374a/wfs.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 7238, 12, 2890, 16, 26735, 33, 7036, 16, 1034, 33, 7036, 16, 8472, 33, 7036, 16, 2572, 350, 33, 7036, 16, 2572, 1589, 33, 7036, 16, 1272, 529, 33, 3292, 14, 17337, 943, 7139, 33...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 7238, 12, 2890, 16, 26735, 33, 7036, 16, 1034, 33, 7036, 16, 8472, 33, 7036, 16, 2572, 350, 33, 7036, 16, 2572, 1589, 33, 7036, 16, 1272, 529, 33, 3292, 14, 17337, 943, 7139, 33...
if self.debug: print >>sys.stderr, "
self.debug(1, "
def run(self): self.state=self.initial[0] self.send_sock = l = conf.L3socket() while 1: try: if self.debug: print >>sys.stderr, "## state=[%s]" % self.state
e317d2d22c1e4dfdb0e7e1931d53d122598d79fb /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7311/e317d2d22c1e4dfdb0e7e1931d53d122598d79fb/scapy.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 4672, 365, 18, 2019, 33, 2890, 18, 6769, 63, 20, 65, 365, 18, 4661, 67, 15031, 273, 328, 273, 2195, 18, 48, 23, 7814, 1435, 1323, 404, 30, 775, 30, 309, 365, 18, 41...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 4672, 365, 18, 2019, 33, 2890, 18, 6769, 63, 20, 65, 365, 18, 4661, 67, 15031, 273, 328, 273, 2195, 18, 48, 23, 7814, 1435, 1323, 404, 30, 775, 30, 309, 365, 18, 41...
f = open(os.path.join(directory, name), 'w')
f = open(os.path.join(directory, name), 'wb')
def write_images(self, directory): for (name, data) in self.IMAGES.items(): f = open(os.path.join(directory, name), 'w') f.write(base64.decodestring(data)) f.close()
2b69e7ac45461732b1622381597218b0dd3a8d58 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/3512/2b69e7ac45461732b1622381597218b0dd3a8d58/html.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1045, 67, 7369, 12, 2890, 16, 1867, 4672, 364, 261, 529, 16, 501, 13, 316, 365, 18, 13603, 55, 18, 3319, 13332, 284, 273, 1696, 12, 538, 18, 803, 18, 5701, 12, 5149, 16, 508, 3631, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1045, 67, 7369, 12, 2890, 16, 1867, 4672, 364, 261, 529, 16, 501, 13, 316, 365, 18, 13603, 55, 18, 3319, 13332, 284, 273, 1696, 12, 538, 18, 803, 18, 5701, 12, 5149, 16, 508, 3631, ...
push(name) need_blank_after_doc = 0
push(self.bold(name)) push('\n')
def _docproperty(self, name, value, mod): results = [] push = results.append
41b8b8a0df4f3e404b0e2be71e8c5ba7ae6ac80c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/41b8b8a0df4f3e404b0e2be71e8c5ba7ae6ac80c/pydoc.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2434, 4468, 12, 2890, 16, 508, 16, 460, 16, 681, 4672, 1686, 273, 5378, 1817, 273, 1686, 18, 6923, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2434, 4468, 12, 2890, 16, 508, 16, 460, 16, 681, 4672, 1686, 273, 5378, 1817, 273, 1686, 18, 6923, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
if rox.app_dir.find(d)==0:
if app_dir.find(d)==0:
def _run_by_injector(): """Internal function.""" try: from zeroinstall.injector import basedir for d in basedir.xdg_cache_dirs: if rox.app_dir.find(d)==0: # Applicaion is in a cache dir return True elif rox._roxlib_dir.find(d)==0: # ROX-Lib is in a cache dir, we are probably being run by the # injector return True except: pass return False
b933633bc6f75861fd45b41140b0d6ccd143145c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/4591/b933633bc6f75861fd45b41140b0d6ccd143145c/mime_handler.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2681, 67, 1637, 67, 28660, 13332, 3536, 3061, 445, 12123, 775, 30, 628, 24910, 885, 3429, 18, 28660, 1930, 15573, 364, 302, 316, 15573, 18, 7669, 75, 67, 2493, 67, 8291, 30, 309, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2681, 67, 1637, 67, 28660, 13332, 3536, 3061, 445, 12123, 775, 30, 628, 24910, 885, 3429, 18, 28660, 1930, 15573, 364, 302, 316, 15573, 18, 7669, 75, 67, 2493, 67, 8291, 30, 309, ...
pass
if not (refID in self.pendingSockets): return canceltimer(self.pendingSockets[refID][2]) self.pendingSockets[refID][2] = None if frame.content == STATUS_CONFIRMED: self.pendingSockets[refID][0] = True try: self.pendingSockets[refID][1].release() except: pass
def _pending_client(self, frame, refID): pass
aa293112179760621e4e879489565b9632bd205e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7995/aa293112179760621e4e879489565b9632bd205e/Multiplexer.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 9561, 67, 2625, 12, 2890, 16, 2623, 16, 1278, 734, 4672, 225, 309, 486, 261, 1734, 734, 316, 365, 18, 9561, 4534, 87, 4672, 327, 282, 3755, 12542, 12, 2890, 18, 9561, 4534, 87, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 9561, 67, 2625, 12, 2890, 16, 2623, 16, 1278, 734, 4672, 225, 309, 486, 261, 1734, 734, 316, 365, 18, 9561, 4534, 87, 4672, 327, 282, 3755, 12542, 12, 2890, 18, 9561, 4534, 87, ...
return re.sub(r'((\\[\\abfnrtv\'"]|\\x..|\\u....)+)',
return re.sub(r'((\\[\\abfnrtv\'"]|\\[0-9]..|\\x..|\\u....)+)',
def repr_string(self, x, level): test = cram(x, self.maxstring) testrepr = repr(test) if '\\' in test and '\\' not in replace(testrepr, (r'\\', '')): # Backslashes are only literal in the string and are never # needed to make any special characters, so show a raw string. return 'r' + testrepr[0] + self.escape(test) + testrepr[0] return re.sub(r'((\\[\\abfnrtv\'"]|\\x..|\\u....)+)', r'<font color="#c040c0">\1</font>', self.escape(testrepr))
a2ebd065552d2e5c9c6624a04e40faa9148962e9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/a2ebd065552d2e5c9c6624a04e40faa9148962e9/pydoc.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8480, 67, 1080, 12, 2890, 16, 619, 16, 1801, 4672, 1842, 273, 276, 1940, 12, 92, 16, 365, 18, 1896, 1080, 13, 1842, 12715, 273, 8480, 12, 3813, 13, 309, 5282, 316, 1842, 471, 5282, 4...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8480, 67, 1080, 12, 2890, 16, 619, 16, 1801, 4672, 1842, 273, 276, 1940, 12, 92, 16, 365, 18, 1896, 1080, 13, 1842, 12715, 273, 8480, 12, 3813, 13, 309, 5282, 316, 1842, 471, 5282, 4...
print("\nWarning: %s already exists. Will not be " "overwritten...\n" % confpath) else: try: open(confpath, "w").write(confdata) os.chmod(confpath, 0600) except Exception, e: print("Error %s occured while trying to write configuration " "file to '%s'\n" % (e, confpath)) raise SystemExit(1)
result = raw_input("\nWarning: %s already exists. " "Overwrite? [y/N]: " % confpath) if result not in ['Y', 'y']: print("Leaving %s unchanged" % confpath) return try: open(confpath, "w").write(confdata) os.chmod(confpath, 0600) except Exception, e: print("Error %s occured while trying to write configuration " "file to '%s'\n" % (e, confpath)) raise SystemExit(1)
def create_conf(confpath, confdata): # don't overwrite existing bcfg2.conf file if os.path.exists(confpath): print("\nWarning: %s already exists. Will not be " "overwritten...\n" % confpath) else: try: open(confpath, "w").write(confdata) os.chmod(confpath, 0600) except Exception, e: print("Error %s occured while trying to write configuration " "file to '%s'\n" % (e, confpath)) raise SystemExit(1)
4b866e4d510f83404425a5ffad3b6149627b273a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11867/4b866e4d510f83404425a5ffad3b6149627b273a/Init.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 67, 3923, 12, 3923, 803, 16, 2195, 892, 4672, 468, 2727, 1404, 6156, 2062, 6533, 2137, 22, 18, 3923, 585, 309, 1140, 18, 803, 18, 1808, 12, 3923, 803, 4672, 563, 273, 1831, 67, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 67, 3923, 12, 3923, 803, 16, 2195, 892, 4672, 468, 2727, 1404, 6156, 2062, 6533, 2137, 22, 18, 3923, 585, 309, 1140, 18, 803, 18, 1808, 12, 3923, 803, 4672, 563, 273, 1831, 67, ...
res = self._findFiles([lfn],['FileID'],connection=connection) if not res['Value']['Successful'].has_key(lfn):
res = self._findFiles( [lfn], ['FileID'], connection = connection ) if not res['Value']['Successful'].has_key( lfn ):
def setReplicaHost(self,lfns,connection=False): connection = self._getConnection(connection) """ Set replica host in the catalog """ successful = {} failed = {} for lfn,info in lfns.items(): res = self._checkInfo(info,['SE','NewSE']) if not res['OK']: failed[lfn] = res['Message'] continue newSE = info['NewSE'] se = info['SE'] res = self._findFiles([lfn],['FileID'],connection=connection) if not res['Value']['Successful'].has_key(lfn): failed[lfn] = res['Value']['Failed'][lfn] continue fileID = res['Value']['Successful'][lfn]['FileID'] res = self._setReplicaHost(fileID,se,newSE,connection=connection) if res['OK']: successful[lfn] = res['Value'] else: failed[lfn] = res['Message'] return S_OK({'Successful':successful,'Failed':failed})
26c5008a7adbc1c761e2409bed59b7710e5f2c74 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/26c5008a7adbc1c761e2409bed59b7710e5f2c74/FileManagerBase.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 444, 14222, 2594, 12, 2890, 16, 20850, 2387, 16, 4071, 33, 8381, 4672, 1459, 273, 365, 6315, 588, 1952, 12, 4071, 13, 3536, 1000, 12335, 1479, 316, 326, 6222, 3536, 6873, 273, 2618, 2535...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 444, 14222, 2594, 12, 2890, 16, 20850, 2387, 16, 4071, 33, 8381, 4672, 1459, 273, 365, 6315, 588, 1952, 12, 4071, 13, 3536, 1000, 12335, 1479, 316, 326, 6222, 3536, 6873, 273, 2618, 2535...
first_index = middle + 1
first_index = middle
def _search_ts (self, target_ts, first_index, last_index):
4ecbff9abd7e2f864d67e0b05021f56f111de597 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12310/4ecbff9abd7e2f864d67e0b05021f56f111de597/Timeline.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 3072, 67, 3428, 261, 2890, 16, 1018, 67, 3428, 16, 1122, 67, 1615, 16, 1142, 67, 1615, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 3072, 67, 3428, 261, 2890, 16, 1018, 67, 3428, 16, 1122, 67, 1615, 16, 1142, 67, 1615, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
FoundationPlist.writePlist(installinfo,
FoundationPlist.writePlist(installinfo,
def check(id=''): '''Checks for available new or updated managed software, downloading installer items if needed. Returns 1 if there are available updates, 0 if there are no available updates, and -1 if there were errors.''' getMachineFacts() munkicommon.report['MachineInfo'] = machine ManagedInstallDir = munkicommon.pref('ManagedInstallDir') if munkicommon.munkistatusoutput: munkistatus.activate() munkistatus.message("Checking for available updates...") munkistatus.detail("") munkistatus.percent("-1") munkicommon.log("### Beginning managed software check ###") mainmanifestpath = getPrimaryManifest(id) if munkicommon.stopRequested(): return 0 installinfo = {} if mainmanifestpath: # initialize our installinfo record installinfo['managed_installs'] = [] installinfo['removals'] = [] munkicommon.display_detail("**Checking for installs**") installinfo = processManifestForInstalls(mainmanifestpath, installinfo) if munkicommon.stopRequested(): return 0 if munkicommon.munkistatusoutput: # reset progress indicator and detail field munkistatus.message("Checking for additional changes...") munkistatus.percent("-1") munkistatus.detail('') # now generate a list of items to be uninstalled munkicommon.display_detail("**Checking for removals**") if munkicommon.stopRequested(): return 0 installinfo = processManifestForRemovals(mainmanifestpath, installinfo) # filter managed_installs to get items already installed installed_items = [item for item in installinfo.get('managed_installs',[]) if item.get('installed')] # filter managed_installs to get problem items: # not installed, but no installer item problem_items = [item for item in installinfo.get('managed_installs',[]) if item.get('installed') == False and not item.get('installer_item')] # filter removals to get items already removed (or never installed) removed_items = [item for item in installinfo.get('removals',[]) if item.get('installed') == False] # filter managed_installs and removals lists # so they have only items that need action installinfo['managed_installs'] = \ [item for item in installinfo.get('managed_installs',[]) if item.get('installer_item')] installinfo['removals'] = \ [item for item in installinfo.get('removals',[]) if item.get('installed')] munkicommon.report['ManagedInstalls'] = installed_items munkicommon.report['ProblemInstalls'] = problem_items munkicommon.report['RemovedItems'] = removed_items munkicommon.report['ItemsToInstall'] = installinfo['managed_installs'] munkicommon.report['ItemsToRemove'] = installinfo['removals'] # clean up cache dir # remove any item in the cache that isn't scheduled # to be used for an install or removal # this could happen if an item is downloaded on one # updatecheck run, but later removed from the manifest # before it is installed or removed - so the cached item # is no longer needed. cache_list = [item["installer_item"] for item in installinfo.get('managed_installs',[])] cache_list.extend([item["uninstaller_item"] for item in installinfo.get('removals',[]) if item.get('uninstaller_item')]) cachedir = os.path.join(ManagedInstallDir, "Cache") for item in os.listdir(cachedir): if item.endswith(".download"): # we have a partial download here # remove the ".download" from the end of the filename fullitem = os.path.splitext(item)[0] if os.path.exists(os.path.join(cachedir, fullitem)): # we have a partial and a full download # for the same item. (This shouldn't happen.) # remove the partial download. os.unlink(os.path.join(cachedir, item)) elif problem_items == []: # problem items is our list of items # that need to be installed but are missing # the installer_item; these might be partial # downloads. So if we have no problem items, it's # OK to get rid of any partial downloads hanging around. os.unlink(os.path.join(cachedir, item)) elif item not in cache_list: munkicommon.display_detail("Removing %s from cache" % item) os.unlink(os.path.join(cachedir, item)) # write out install list so our installer # can use it to install things in the right order installinfochanged = True installinfopath = os.path.join(ManagedInstallDir, "InstallInfo.plist") if os.path.exists(installinfopath): oldinstallinfo = FoundationPlist.readPlist(installinfopath) if oldinstallinfo == installinfo: installinfochanged = False munkicommon.display_detail("No change in InstallInfo.") if installinfochanged: FoundationPlist.writePlist(installinfo, os.path.join(ManagedInstallDir, "InstallInfo.plist")) else: # couldn't get a primary manifest. Check to see if we have a valid # install/remove list from an earlier run. munkicommon.display_error( "Could not retrieve managed install primary manifest.") installinfopath = os.path.join(ManagedInstallDir, "InstallInfo.plist") if os.path.exists(installinfopath): try: installinfo = FoundationPlist.readPlist(installinfopath) except FoundationPlist.NSPropertyListSerializationException: installinfo = {} else: munkicommon.report['ItemsToInstall'] = \ installinfo.get('managed_installs',[]) munkicommon.report['ItemsToRemove'] = \ installinfo.get('removals',[]) installcount = len(installinfo.get("managed_installs",[])) removalcount = len(installinfo.get("removals",[])) munkicommon.log("") if installcount: munkicommon.display_info( "The following items will be installed or upgraded:") for item in installinfo.get('managed_installs',[]): if item.get('installer_item'): munkicommon.display_info(" + %s-%s" % (item.get('name',''), item.get('version_to_install',''))) if item.get('description'): munkicommon.display_info(" %s" % item['description']) if item.get('RestartAction') == 'RequireRestart' or \ item.get('RestartAction') == 'RecommendRestart': munkicommon.display_info(" *Restart required") munkicommon.report['RestartRequired'] = True if item.get('RestartAction') == 'RequireLogout': munkicommon.display_info(" *Logout required") munkicommon.report['LogoutRequired'] = True if removalcount: munkicommon.display_info("The following items will be removed:") for item in installinfo.get('removals',[]): if item.get('installed'): munkicommon.display_info(" - %s" % item.get('name')) if item.get('RestartAction') == 'RequireRestart' or \ item.get('RestartAction') == 'RecommendRestart': munkicommon.display_info(" *Restart required") munkicommon.report['RestartRequired'] = True if item.get('RestartAction') == 'RequireLogout': munkicommon.display_info(" *Logout required") munkicommon.report['LogoutRequired'] = True if installcount == 0 and removalcount == 0: munkicommon.display_info( "No changes to managed software are available.") munkicommon.savereport() munkicommon.log("### End managed software check ###") if installcount or removalcount: return 1 else: return 0
5350042f4ddfb098468c8d75b15dfa28757b9321 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6241/5350042f4ddfb098468c8d75b15dfa28757b9321/updatecheck.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 866, 12, 350, 2218, 11, 4672, 9163, 4081, 364, 2319, 394, 578, 3526, 7016, 17888, 16, 23742, 16050, 1516, 309, 3577, 18, 2860, 404, 309, 1915, 854, 2319, 4533, 16, 374, 309, 1915, 854, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 866, 12, 350, 2218, 11, 4672, 9163, 4081, 364, 2319, 394, 578, 3526, 7016, 17888, 16, 23742, 16050, 1516, 309, 3577, 18, 2860, 404, 309, 1915, 854, 2319, 4533, 16, 374, 309, 1915, 854, ...
speedrace(hdlr, pct, pct+pct_step, circs_per_node, save_every, out_dir)
speedrace(hdlr, pct, pct+pct_step, circs_per_node, save_every, out_dir, max_fetch_time)
def main(argv): TorUtil.read_config(argv[1]) (start_pct,stop_pct,nodes_per_slice,save_every, circs_per_node,out_dir) = read_config(argv[1]) try: (c,hdlr) = setup_handler() except Exception, e: plog("WARN", "Can't connect to Tor: "+str(e)) sql_file = os.getcwd()+'/'+out_dir+'/bwauthority.sqlite' hdlr.attach_sql_listener('sqlite:///'+sql_file) # set SOCKS proxy socks.setdefaultproxy(socks.PROXY_TYPE_SOCKS5, tor_host, tor_port) socket.socket = socks.socksocket while True: pct = start_pct plog('INFO', 'Beginning time loop') while pct < stop_pct: pct_step = hdlr.rank_to_percent(nodes_per_slice) hdlr.reset_stats() hdlr.commit() plog('DEBUG', 'Reset stats') speedrace(hdlr, pct, pct+pct_step, circs_per_node, save_every, out_dir) plog('DEBUG', 'speedroced') hdlr.commit() hdlr.close_circuits() lo = str(round(pct,1)) hi = str(round(pct+pct_step,1)) hdlr.write_sql_stats(pct, pct+pct_step, os.getcwd()+'/'+out_dir+'/sql-'+lo+':'+hi+"-done-"+strftime("20%y-%m-%d-%H:%M:%S")) hdlr.write_strm_bws(pct, pct+pct_step, os.getcwd()+'/'+out_dir+'/bws-'+lo+':'+hi+"-done-"+strftime("20%y-%m-%d-%H:%M:%S")) plog('DEBUG', 'Wrote stats') pct += pct_step hdlr.save_sql_file(sql_file, "db-"+str(lo)+":"+str(hi)+"-"+strftime("20%y-%m-%d-%H:%M:%S")+".sqlite")
e8450ad1cd3c23315f682731ad3047861f509350 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3762/e8450ad1cd3c23315f682731ad3047861f509350/bwauthority.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 12, 19485, 4672, 23763, 1304, 18, 896, 67, 1425, 12, 19485, 63, 21, 5717, 261, 1937, 67, 23989, 16, 5681, 67, 23989, 16, 4690, 67, 457, 67, 6665, 16, 5688, 67, 20434, 16, 5886, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 12, 19485, 4672, 23763, 1304, 18, 896, 67, 1425, 12, 19485, 63, 21, 5717, 261, 1937, 67, 23989, 16, 5681, 67, 23989, 16, 4690, 67, 457, 67, 6665, 16, 5688, 67, 20434, 16, 5886, ...
def base_ring(self): """ Return the underlying base ring. """ return self._base_ring
def __repr__(self): return "SpecialCubicQuotientRing over %s with polynomial T = %s" % \ (self._base_ring, PolynomialRing(self._base_ring)( [self._b, self._a, 0, 1]))
c26d6b07795f11a9104ed5b11740f789a4ab67f6 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9890/c26d6b07795f11a9104ed5b11740f789a4ab67f6/monsky_washnitzer.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 12715, 972, 12, 2890, 4672, 327, 315, 12193, 39, 17616, 7678, 1979, 10369, 1879, 738, 87, 598, 16991, 399, 273, 738, 87, 6, 738, 521, 261, 2890, 6315, 1969, 67, 8022, 16, 6730, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 12715, 972, 12, 2890, 4672, 327, 315, 12193, 39, 17616, 7678, 1979, 10369, 1879, 738, 87, 598, 16991, 399, 273, 738, 87, 6, 738, 521, 261, 2890, 6315, 1969, 67, 8022, 16, 6730, 1...
self["medium_label"].setText(_("exceeds dual layer medium!"))
self["medium_label"].setText(_("Exceeds dual layer medium!"))
def updateSize(self): size = self.project.size/(1024*1024) MAX_DL = self.project.MAX_DL-100 MAX_SL = self.project.MAX_SL-100 print "updateSize:", size, "MAX_DL:", MAX_DL, "MAX_SL:", MAX_SL if size > MAX_DL: percent = 100 * size / float(MAX_DL) self["space_label_dual"].text = "%d MB (%.2f%%)" % (size, percent) self["space_bar_dual"].value = int(percent) self["space_bar_single"].value = 100 self["space_label_single"].text = "" self["medium_label"].setText(_("exceeds dual layer medium!")) self["medium_label"].setForegroundColorNum(2) if self.previous_size < MAX_DL: self.session.open(MessageBox,text = _("exceeds dual layer medium!"), type = MessageBox.TYPE_ERROR) elif size > MAX_SL: percent = 100 * size / float(MAX_DL) self["space_label_dual"].text = "%d MB (%.2f%%)" % (size, percent) self["space_bar_dual"].value = int(percent) self["space_bar_single"].value = 100 self["space_label_single"].text = "" self["medium_label"].setText(_("required medium type:") + " " + _("DUAL LAYER DVD") + ", %d MB " % (MAX_DL - size) + _("free")) self["medium_label"].setForegroundColorNum(1) if self.previous_size < MAX_SL: self.session.open(MessageBox, text = _("Your collection exceeds the size of a single layer medium, you will need a blank dual layer DVD!"), timeout = 10, type = MessageBox.TYPE_INFO) elif size < MAX_SL: percent = 100 * size / float(MAX_SL) self["space_label_single"].text = "%d MB (%.2f%%)" % (size, percent) self["space_bar_single"].value = int(percent) self["space_bar_dual"].value = 0 self["space_label_dual"].text = "" self["medium_label"].setText(_("required medium type:") + " " + _("SINGLE LAYER DVD") + ", %d MB " % (MAX_SL - size) + _("free")) self["medium_label"].setForegroundColorNum(0) self.previous_size = size
16e2662171701ce64bf2e789168e8cee54dc8588 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6652/16e2662171701ce64bf2e789168e8cee54dc8588/TitleList.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 1225, 12, 2890, 4672, 963, 273, 365, 18, 4406, 18, 1467, 19, 12, 2163, 3247, 14, 2163, 3247, 13, 4552, 67, 8914, 273, 365, 18, 4406, 18, 6694, 67, 8914, 17, 6625, 4552, 67, 455...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 1225, 12, 2890, 4672, 963, 273, 365, 18, 4406, 18, 1467, 19, 12, 2163, 3247, 14, 2163, 3247, 13, 4552, 67, 8914, 273, 365, 18, 4406, 18, 6694, 67, 8914, 17, 6625, 4552, 67, 455...
license = 'BSD',
license = 'LGPL',
def finalize_options(self): build_ext.finalize_options(self) self.include_dirs.insert(0, 'include') self.pyrex_include_dirs.extend(self.include_dirs)
fb7c0814001a30edd3c24892c3fd6eaf7520360a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12343/fb7c0814001a30edd3c24892c3fd6eaf7520360a/setup.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 12409, 67, 2116, 12, 2890, 4672, 1361, 67, 408, 18, 30343, 67, 2116, 12, 2890, 13, 365, 18, 6702, 67, 8291, 18, 6387, 12, 20, 16, 296, 6702, 6134, 365, 18, 2074, 266, 92, 67, 6702, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 12409, 67, 2116, 12, 2890, 4672, 1361, 67, 408, 18, 30343, 67, 2116, 12, 2890, 13, 365, 18, 6702, 67, 8291, 18, 6387, 12, 20, 16, 296, 6702, 6134, 365, 18, 2074, 266, 92, 67, 6702, ...
def getHash(self, path=None): """Get the hash of a file.
def getHash(self, path): """Get the hash of a file stored in ``path``.
def getHash(self, path=None): """Get the hash of a file.
fb0d438c88490272f659933b5d44bb0582381fa2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9162/fb0d438c88490272f659933b5d44bb0582381fa2/cachemanager.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 16075, 12, 2890, 16, 589, 33, 7036, 4672, 3536, 967, 326, 1651, 434, 279, 585, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 16075, 12, 2890, 16, 589, 33, 7036, 4672, 3536, 967, 326, 1651, 434, 279, 585, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100...
s.bind('', self.port)
s.bind(('', self.port))
def __init__(self, port=None, connection_hook=None): self.connections = [] self.port = port or self.default_port self.connection_hook = connection_hook
9a580c440c4f064a92bbc537ac3a2df1c0998afc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/9a580c440c4f064a92bbc537ac3a2df1c0998afc/protocol.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 1756, 33, 7036, 16, 1459, 67, 4476, 33, 7036, 4672, 365, 18, 13313, 273, 5378, 365, 18, 655, 273, 1756, 578, 365, 18, 1886, 67, 655, 365, 18, 4071, 67,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 1756, 33, 7036, 16, 1459, 67, 4476, 33, 7036, 4672, 365, 18, 13313, 273, 5378, 365, 18, 655, 273, 1756, 578, 365, 18, 1886, 67, 655, 365, 18, 4071, 67,...
return (invalid[index], "kNoError", "GL_INVALID_ENUM") return ("---ERROR1---", "kNoError", "GL_INVALID_ENUM")
return (invalid[index], "kNoError", self.gl_error) return ("---ERROR1---", "kNoError", self.gl_error)
def GetInvalidArg(self, offset, index): """returns an invalid value by index.""" if 'invalid' in self.enum_info: invalid = self.enum_info['invalid'] num_invalid = len(invalid) if index >= num_invalid: index = num_invalid - 1 return (invalid[index], "kNoError", "GL_INVALID_ENUM") return ("---ERROR1---", "kNoError", "GL_INVALID_ENUM")
51330b3a08d845da9815b843e020833e62577828 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9392/51330b3a08d845da9815b843e020833e62577828/build_gles2_cmd_buffer.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 968, 1941, 4117, 12, 2890, 16, 1384, 16, 770, 4672, 3536, 6154, 392, 2057, 460, 635, 770, 12123, 309, 296, 5387, 11, 316, 365, 18, 7924, 67, 1376, 30, 2057, 273, 365, 18, 7924, 67, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 968, 1941, 4117, 12, 2890, 16, 1384, 16, 770, 4672, 3536, 6154, 392, 2057, 460, 635, 770, 12123, 309, 296, 5387, 11, 316, 365, 18, 7924, 67, 1376, 30, 2057, 273, 365, 18, 7924, 67, 1...
M_PROPPATCH = 9
M_PROPPATCH = 9
def init(): """ This function is called by the server at startup time """ return CallBack()
3de9a0d15ba9eb7275476985ecfc3381876d7703 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10002/3de9a0d15ba9eb7275476985ecfc3381876d7703/apache.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1208, 13332, 3536, 1220, 445, 353, 2566, 635, 326, 1438, 622, 11850, 813, 3536, 225, 327, 3049, 2711, 1435, 225, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1208, 13332, 3536, 1220, 445, 353, 2566, 635, 326, 1438, 622, 11850, 813, 3536, 225, 327, 3049, 2711, 1435, 225, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
site = server.Site(root)
rewriter = rewrite.RewriterResource(root, helpimagesrewrite) site = server.Site(rewriter)
def main(): # the start and stop stuff will be handled from the freevo script logfile = '%s/webserver-%s.log' % (config.LOGDIR, os.getuid()) log.startLogging(open(logfile, 'a')) if os.path.isdir(os.path.join(os.environ['FREEVO_PYTHON'], 'www/htdocs')): docRoot = os.path.join(os.environ['FREEVO_PYTHON'], 'www/htdocs') else: docRoot = os.path.join(config.SHARE_DIR, 'htdocs') root = static.File(docRoot) root.processors = { '.rpy': script.ResourceScript, } root.putChild('vhost', vhost.VHostMonsterResource()) site = server.Site(root) application = app.Application('web') application.listenTCP(config.WWW_PORT, site) application.run(save=0)
f520ecdadf2f70a3822e5eb9fee23ac465913844 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11399/f520ecdadf2f70a3822e5eb9fee23ac465913844/webserver.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 468, 326, 787, 471, 2132, 10769, 903, 506, 7681, 628, 326, 4843, 12307, 2728, 225, 15204, 273, 1995, 87, 19, 1814, 2266, 502, 6456, 87, 18, 1330, 11, 738, 261, 1425, 18, 4...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 468, 326, 787, 471, 2132, 10769, 903, 506, 7681, 628, 326, 4843, 12307, 2728, 225, 15204, 273, 1995, 87, 19, 1814, 2266, 502, 6456, 87, 18, 1330, 11, 738, 261, 1425, 18, 4...
dir = os.path.expanduser("~/.Trash") for file in FileUtilities.children_in_directory(dir, False): yield file
dirname = os.path.expanduser("~/.Trash") for filename in FileUtilities.children_in_directory(dirname, False): yield filename
def list_files(self): dir = os.path.expanduser("~/.Trash") for file in FileUtilities.children_in_directory(dir, False): yield file # fixme http://www.ramendik.ru/docs/trashspec.html # http://standards.freedesktop.org/basedir-spec/basedir-spec-0.6.html # GNOME 2.22, Fedora 9 ~/.local/share/Trash dir = os.path.expanduser("~/.local/share/Trash") for file in FileUtilities.children_in_directory(dir, False): yield file
737a941d1a6026b292895c42a857f9ec0bf05cea /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7853/737a941d1a6026b292895c42a857f9ec0bf05cea/CleanerBackend.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 666, 67, 2354, 12, 2890, 4672, 4283, 273, 1140, 18, 803, 18, 12320, 1355, 2932, 98, 15865, 29420, 7923, 364, 1544, 316, 1387, 11864, 18, 5906, 67, 267, 67, 5149, 12, 12287, 16, 1083, 4...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 666, 67, 2354, 12, 2890, 4672, 4283, 273, 1140, 18, 803, 18, 12320, 1355, 2932, 98, 15865, 29420, 7923, 364, 1544, 316, 1387, 11864, 18, 5906, 67, 267, 67, 5149, 12, 12287, 16, 1083, 4...
self.string = list(string)
self.index = 0 self.string = string
def __init__(self, string):
d46b2a42eda94cbac3b610e36d5f97230d629068 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/d46b2a42eda94cbac3b610e36d5f97230d629068/sre_parse.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 533, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 533, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100,...
index = self.bisect(self, item)
index = bisect.bisect(self.list, item)
def __contains__(self, item): if self.__len__() == 0: return False index = self.bisect(self, item) if index == 0: return False if item in self.__getitem__(index-1): return True return False
ceb5841528feafa5a760412ecaeedf36fbba1e7b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/13031/ceb5841528feafa5a760412ecaeedf36fbba1e7b/mirrorlist_server.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 12298, 972, 12, 2890, 16, 761, 4672, 309, 365, 16186, 1897, 972, 1435, 422, 374, 30, 327, 1083, 770, 273, 27031, 18, 70, 291, 386, 12, 2890, 18, 1098, 16, 761, 13, 309, 770, 42...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 12298, 972, 12, 2890, 16, 761, 4672, 309, 365, 16186, 1897, 972, 1435, 422, 374, 30, 327, 1083, 770, 273, 27031, 18, 70, 291, 386, 12, 2890, 18, 1098, 16, 761, 13, 309, 770, 42...
dev=self.selectedDevice(c) dev.onlyChannel=channel
dev = self.selectedDevice(c) dev.onlyChannel = channel
def regulate(self, c, channel, temperature, loadresistor=30000): """Initializes temperature regulation
ca09465af7cd0d53cd2656a51c12b3b3003818f7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7016/ca09465af7cd0d53cd2656a51c12b3b3003818f7/lakeshore370.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 960, 6243, 12, 2890, 16, 276, 16, 1904, 16, 12843, 16, 1262, 455, 376, 280, 33, 23, 2787, 4672, 3536, 9685, 12843, 960, 6234, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 960, 6243, 12, 2890, 16, 276, 16, 1904, 16, 12843, 16, 1262, 455, 376, 280, 33, 23, 2787, 4672, 3536, 9685, 12843, 960, 6234, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
raise ScriptError("Patch %s from bug %s failed to download and apply." % (patch['url'], patch['bug_id']))
raise ScriptError(message="Patch %s from bug %s failed to download and apply." % (patch['url'], patch['bug_id']))
def apply_patch(self, patch, force=False): # It's possible that the patch was not made from the root directory. # We should detect and handle that case. curl_process = subprocess.Popen(['curl', '--location', '--silent', '--show-error', patch['url']], stdout=subprocess.PIPE) args = [self.script_path('svn-apply'), '--reviewer', patch['reviewer']] if force: args.append('--force') patch_apply_process = subprocess.Popen(args, stdin=curl_process.stdout)
cbfef173775a18c163a046efa120abeeff3a6751 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9392/cbfef173775a18c163a046efa120abeeff3a6751/scm.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2230, 67, 2272, 12, 2890, 16, 4729, 16, 2944, 33, 8381, 4672, 468, 2597, 1807, 3323, 716, 326, 4729, 1703, 486, 7165, 628, 326, 1365, 1867, 18, 468, 1660, 1410, 5966, 471, 1640, 716, 6...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2230, 67, 2272, 12, 2890, 16, 4729, 16, 2944, 33, 8381, 4672, 468, 2597, 1807, 3323, 716, 326, 4729, 1703, 486, 7165, 628, 326, 1365, 1867, 18, 468, 1660, 1410, 5966, 471, 1640, 716, 6...
return chain(self.status())
for msg in chain(self.status()): yield msg
def stop(self): if self._status == 'stopped': raise DiscoError("%s already stopped" % self) try: os.kill(self.pid, signal.SIGTERM) while self._status == 'running': pass except Exception: pass return chain(self.status())
ad8c242af148154828b966902f5ab8e49963a98c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/373/ad8c242af148154828b966902f5ab8e49963a98c/disco.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2132, 12, 2890, 4672, 309, 365, 6315, 2327, 422, 296, 24228, 4278, 1002, 3035, 2894, 668, 27188, 87, 1818, 9627, 6, 738, 365, 13, 775, 30, 1140, 18, 16418, 12, 2890, 18, 6610, 16, 4277...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2132, 12, 2890, 4672, 309, 365, 6315, 2327, 422, 296, 24228, 4278, 1002, 3035, 2894, 668, 27188, 87, 1818, 9627, 6, 738, 365, 13, 775, 30, 1140, 18, 16418, 12, 2890, 18, 6610, 16, 4277...
selected_tabs.add(row[0])
selected_tabs.add((1, row[0]))
def get_selection_and_children(self): """This returns the selection and, in the case of parent rows, returns all children, too. This is particularly useful for getting selections that include children of folders.
1f498bda8c507f0b63f8859884457bea2e925013 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12354/1f498bda8c507f0b63f8859884457bea2e925013/tablistmanager.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 10705, 67, 464, 67, 5906, 12, 2890, 4672, 3536, 2503, 1135, 326, 4421, 471, 16, 316, 326, 648, 434, 982, 2595, 16, 1135, 777, 2325, 16, 4885, 18, 225, 1220, 353, 6826, 715, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 10705, 67, 464, 67, 5906, 12, 2890, 4672, 3536, 2503, 1135, 326, 4421, 471, 16, 316, 326, 648, 434, 982, 2595, 16, 1135, 777, 2325, 16, 4885, 18, 225, 1220, 353, 6826, 715, ...
print " abort!" return
raise RuntimeError, "certain not superset of knownKeep"
def split(self, childSize, otherMarkerId, turn=-1): print "split for", self, childSize, otherMarkerId, turn
d3360650260167e99d819d46c099993d6e0f1f9e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3755/d3360650260167e99d819d46c099993d6e0f1f9e/predictsplits.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1416, 12, 2890, 16, 1151, 1225, 16, 1308, 7078, 548, 16, 7005, 29711, 21, 4672, 1172, 315, 4939, 364, 3113, 365, 16, 1151, 1225, 16, 1308, 7078, 548, 16, 7005, 2, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1416, 12, 2890, 16, 1151, 1225, 16, 1308, 7078, 548, 16, 7005, 29711, 21, 4672, 1172, 315, 4939, 364, 3113, 365, 16, 1151, 1225, 16, 1308, 7078, 548, 16, 7005, 2, -100, -100, -100, -10...
'Added interdiff support')
'Improved login form')
def testDashboard5(self): """Testing dashboard view (mine)""" self.client.login(username='doc', password='doc')
d72238b7393eae1f5069fef4eba2d906a7ba3bdf /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1600/d72238b7393eae1f5069fef4eba2d906a7ba3bdf/tests.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 14830, 25, 12, 2890, 4672, 3536, 22218, 11825, 1476, 261, 3081, 15574, 365, 18, 2625, 18, 5819, 12, 5053, 2218, 2434, 2187, 2201, 2218, 2434, 6134, 2, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 14830, 25, 12, 2890, 4672, 3536, 22218, 11825, 1476, 261, 3081, 15574, 365, 18, 2625, 18, 5819, 12, 5053, 2218, 2434, 2187, 2201, 2218, 2434, 6134, 2, -100, -100, -100, -100, -100, ...
deps = create_deps(ext_modules) cython(deps, ext_modules)
hi, med, lo = compile_command_list(ext_modules) execute_list_of_commands(hi + med + lo)
def create_deps(ext_modules): # first we compute the complete graph of dependencies deps = {} for m in ext_modules: m.extra_compile_args += extra_compile_args for i in range(len(m.sources)): f = m.sources[i] if f[-4:] == '.pyx': visited = set() deps_graph(deps, f, visited) base = os.path.splitext(f)[0] f = f[:-4] + '.pxd' if os.path.exists(f): deps_graph(deps, f, visited) return deps
12aff705aa1e2b2dc6e401b763fe2e9e51bdf02d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/12aff705aa1e2b2dc6e401b763fe2e9e51bdf02d/setup.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 67, 14877, 12, 408, 67, 6400, 4672, 468, 1122, 732, 3671, 326, 3912, 2667, 434, 5030, 8740, 273, 2618, 364, 312, 316, 1110, 67, 6400, 30, 312, 18, 7763, 67, 11100, 67, 1968, 1011,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 67, 14877, 12, 408, 67, 6400, 4672, 468, 1122, 732, 3671, 326, 3912, 2667, 434, 5030, 8740, 273, 2618, 364, 312, 316, 1110, 67, 6400, 30, 312, 18, 7763, 67, 11100, 67, 1968, 1011,...
print "Skipping"
print "Skipping exercise(): input file not available."
def exercise () : from iotbx import file_reader pdb_file = libtbx.env.find_in_repositories( relative_path="phenix_regression/pdb/1ywf.pdb", test=os.path.isfile) pdb_file_h = libtbx.env.find_in_repositories( relative_path="phenix_regression/pdb/1ywf_h.pdb", test=os.path.isfile) if pdb_file is None : print "Skipping" return False log = cStringIO.StringIO() pdb_in = file_reader.any_file(pdb_file_h, force_type="pdb").file_object pdb_hierarchy = pdb_in.construct_hierarchy() xray_structure = pdb_in.xray_structure_simple() sec_str_from_pdb_file = pdb_in.extract_secondary_structure() m = manager(pdb_hierarchy=pdb_hierarchy, xray_structure=xray_structure, sec_str_from_pdb_file=sec_str_from_pdb_file) m.find_automatically(log=log) bonds_table = m.get_bonds_table(log=log) assert bonds_table.bonds.size() == 109 m.params.h_bond_restraints.substitute_n_for_h = True bonds_table = m.get_bonds_table(log=log) assert bonds_table.flag_use_bond.count(True) == 106 (frac_alpha, frac_beta) = m.calculate_structure_content() assert ("%.3f" % frac_alpha) == "0.643" assert ("%.3f" % frac_beta) == "0.075" del m # using KSDSSP try : m = manager(pdb_hierarchy=pdb_hierarchy, xray_structure=xray_structure, sec_str_from_pdb_file=None) m.find_automatically(log=log) bonds_table = m.get_bonds_table(log=log) assert bonds_table.bonds.size() == 93 m.params.h_bond_restraints.substitute_n_for_h = True bonds_table = m.get_bonds_table(log=log) assert bonds_table.flag_use_bond.count(True) == 86 (frac_alpha, frac_beta) = m.calculate_structure_content() assert ("%.3f" % frac_alpha) == "0.552" assert ("%.3f" % frac_beta) == "0.066" del m del pdb_hierarchy del xray_structure except RuntimeError : print "skipping KSDSSP test" # without hydrogens pdb_in = file_reader.any_file(pdb_file, force_type="pdb").file_object pdb_hierarchy = pdb_in.construct_hierarchy() xray_structure = pdb_in.xray_structure_simple() sec_str_from_pdb_file = pdb_in.extract_secondary_structure() m = manager(pdb_hierarchy=pdb_hierarchy, xray_structure=xray_structure, sec_str_from_pdb_file=sec_str_from_pdb_file) m.find_automatically(log=log) bonds_table = m.get_bonds_table(log=log) assert bonds_table.bonds.size() == 109 del m # using KSDSSP try : m = manager(pdb_hierarchy=pdb_hierarchy, xray_structure=xray_structure, sec_str_from_pdb_file=None) m.find_automatically(log=log) bonds_table = m.get_bonds_table(log=log) assert bonds_table.bonds.size() == 93 except RuntimeError : print "skipping KSDSSP test" print "OK"
0bc14071d9364def7d33a409fac526a38f23f248 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/696/0bc14071d9364def7d33a409fac526a38f23f248/secondary_structure.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24165, 1832, 294, 628, 30956, 70, 92, 1930, 585, 67, 10530, 10892, 67, 768, 273, 2561, 18587, 92, 18, 3074, 18, 4720, 67, 267, 67, 26626, 12, 3632, 67, 803, 1546, 13819, 697, 67, 1574,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24165, 1832, 294, 628, 30956, 70, 92, 1930, 585, 67, 10530, 10892, 67, 768, 273, 2561, 18587, 92, 18, 3074, 18, 4720, 67, 267, 67, 26626, 12, 3632, 67, 803, 1546, 13819, 697, 67, 1574,...
d = FSPageTemplate.__dict__ d['source.xml'] = d['source.html'] = Src()
source_dot_xml = Src() setattr(FSPageTemplate, 'source.xml', FSPageTemplate.source_dot_xml) setattr(FSPageTemplate, 'source.html', FSPageTemplate.source_dot_xml)
def _exec(self, bound_names, args, kw): """Call a FSPageTemplate""" try: response = self.REQUEST.RESPONSE except AttributeError: response = None # Read file first to get a correct content_type default value. self._updateFromFS()
6aed3d2ffbfbdc919c13c47bad2bba3569ff3d5c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1843/6aed3d2ffbfbdc919c13c47bad2bba3569ff3d5c/FSPageTemplate.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 4177, 12, 2890, 16, 2489, 67, 1973, 16, 833, 16, 5323, 4672, 3536, 1477, 279, 9247, 1964, 2283, 8395, 775, 30, 766, 273, 365, 18, 5519, 18, 14508, 1335, 6394, 30, 766, 273, 599, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 4177, 12, 2890, 16, 2489, 67, 1973, 16, 833, 16, 5323, 4672, 3536, 1477, 279, 9247, 1964, 2283, 8395, 775, 30, 766, 273, 365, 18, 5519, 18, 14508, 1335, 6394, 30, 766, 273, 599, ...
watchpats.append("Forcing node %s to be terminated" % node)
watchpats.append("Node %s will be fenced because termination was requested" % node)
def __call__(self, node): self.incr("calls") if len(self.CM.Env["nodes"]) < 2: return self.skipped()
0aba552579a25f5248eac3b2776812d5f9c7051e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11962/0aba552579a25f5248eac3b2776812d5f9c7051e/CTStests.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 1991, 972, 12, 2890, 16, 756, 4672, 365, 18, 267, 3353, 2932, 12550, 7923, 309, 562, 12, 2890, 18, 9611, 18, 3491, 9614, 4690, 6, 5717, 411, 576, 30, 327, 365, 18, 25346, 1435, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 1991, 972, 12, 2890, 16, 756, 4672, 365, 18, 267, 3353, 2932, 12550, 7923, 309, 562, 12, 2890, 18, 9611, 18, 3491, 9614, 4690, 6, 5717, 411, 576, 30, 327, 365, 18, 25346, 1435, ...
if re.search( table_name+r'[.]', table_param ) is None:
if table_param.find( table_name+'.' ) == -1:
def __init__( self, table_name, table_param, param_ranges_opt, verbose = False ): """ Parse --param-ranges option. Creates self.param which is the table_name and the table_param appended together (with a '.') and self.param_ranges, which is a list of tuples that give the lower parameter value, whether it is an open or closed boundary, and the same for the upper parameter. For example, if table_name is coinc_inspiral, table_param is mchirp and param_ranges_opt is '[2,8);[8,17]' will get: self.param = 'coinc_inspiral.mchirp' self.param_ranges = [ ( ('>=',2.0), ('<',8.0) ), ( ('>=',8.0), ('<=', 17.0) ) ]
aa613f9ca1679aa7fff3f9cfa9b4bb9989c85fd7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/aa613f9ca1679aa7fff3f9cfa9b4bb9989c85fd7/ligolw_sqlutils.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 365, 16, 1014, 67, 529, 16, 1014, 67, 891, 16, 579, 67, 14530, 67, 3838, 16, 3988, 273, 1083, 262, 30, 3536, 2884, 1493, 891, 17, 14530, 1456, 18, 10210, 365, 18...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 365, 16, 1014, 67, 529, 16, 1014, 67, 891, 16, 579, 67, 14530, 67, 3838, 16, 3988, 273, 1083, 262, 30, 3536, 2884, 1493, 891, 17, 14530, 1456, 18, 10210, 365, 18...
regex_find_licenses = re.compile(r'(?<!\{)\{\{(?:[Tt]emplate:|)([^{]*?)[|\n<}]', re.DOTALL)
regex_find_licenses = re.compile(r'(?<!\{)\{\{(?:[Tt]emplate:|)([^{]+?)[|\n<}]', re.DOTALL)
def smartDetection(self): """ The bot instead of checking if there's a simple template in the image's description, checks also if that template is a license or something else. In this sense this type of check is smart. """ self.seems_ok = False self.license_found = None self.hiddentemplates = self.loadHiddenTemplates() self.licenses_found = self.image.getTemplates() self.whiteTemplatesFound = False regex_find_licenses = re.compile(r'(?<!\{)\{\{(?:[Tt]emplate:|)([^{]*?)[|\n<}]', re.DOTALL) templatesInTheImageRaw = regex_find_licenses.findall(self.imageCheckText) self.allLicenses = list() if self.list_licenses == []: raise wikipedia.Error(u'No licenses allowed provided, add that option to the code to make the script working correctly') # Found the templates ONLY in the image's description for template_selected in templatesInTheImageRaw: for templateReal in self.licenses_found: if self.convert_to_url(template_selected).lower().replace('template:', '') == \ self.convert_to_url(templateReal.title().lower().replace('template:', '')): if templateReal not in self.allLicenses: # don't put the same template, twice. self.allLicenses.append(templateReal) if self.licenses_found != []: self.templateInList() if self.license_found == None and self.allLicenses != list(): self.license_found = self.license_selected if not self.seems_ok and self.license_found != None: rep_text_license_fake = u"\n*[[:Image:%s]] seems to have " % self.imageName + \ "a ''fake license'', license detected: <nowiki>%s</nowiki>" % self.license_found regexFakeLicense = r"\* ?\[\[:Image:%s\]\] seems to have " % (re.escape(self.imageName)) + \ "a ''fake license'', license detected: <nowiki>%s</nowiki>$" % (re.escape(self.license_found)) printWithTimeZone(u"%s seems to have a fake license: %s, reporting..." % (self.imageName, self.license_found)) self.report_image(self.imageName, rep_text = rep_text_license_fake, addings = False, regex = regexFakeLicense) elif self.license_found != None: printWithTimeZone(u"%s seems ok, license found: %s..." % (self.imageName, self.license_found)) return (self.license_found, self.whiteTemplatesFound)
c8bd831cf36832667e225d2876619a25779e9888 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/4404/c8bd831cf36832667e225d2876619a25779e9888/checkimages.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13706, 10214, 12, 2890, 4672, 3536, 1021, 2512, 3560, 434, 6728, 309, 1915, 1807, 279, 4143, 1542, 316, 326, 1316, 1807, 2477, 16, 4271, 2546, 309, 716, 1542, 353, 279, 8630, 578, 5943, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13706, 10214, 12, 2890, 4672, 3536, 1021, 2512, 3560, 434, 6728, 309, 1915, 1807, 279, 4143, 1542, 316, 326, 1316, 1807, 2477, 16, 4271, 2546, 309, 716, 1542, 353, 279, 8630, 578, 5943, ...
or self.getType() != u'bittorrent'):
or self.get_type() != u'bittorrent'):
def startUpload(self): if (self.get_state() not in (u'finished', u'uploading-paused') or self.getType() != u'bittorrent'): return self.manualUpload = True if _downloads.has_key(self.dlid): c = command.StartDownloadCommand(RemoteDownloader.dldaemon, self.dlid) c.send() else: self.before_changing_status() self.status['state'] = u'uploading' self.after_changing_status() self.restart() self.signal_change()
f2fa1d771c944ca514cd470008d100ce0c0da511 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12354/f2fa1d771c944ca514cd470008d100ce0c0da511/downloader.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 787, 4777, 12, 2890, 4672, 309, 261, 2890, 18, 588, 67, 2019, 1435, 486, 316, 261, 89, 11, 13527, 2187, 582, 11, 6327, 310, 17, 8774, 3668, 6134, 578, 365, 18, 588, 67, 723, 1435, 48...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 787, 4777, 12, 2890, 4672, 309, 261, 2890, 18, 588, 67, 2019, 1435, 486, 316, 261, 89, 11, 13527, 2187, 582, 11, 6327, 310, 17, 8774, 3668, 6134, 578, 365, 18, 588, 67, 723, 1435, 48...
signal.alarm(max(math.ceil(prevTimeout - passed), 1))
signal.alarm(int(max(math.ceil(prevTimeout - passed), 1)))
def alarmhandler(signum,stackframe): raise alarmExc("timeout expired")
5c228c4a7a2880defc1ea9809ca61b9010176abf /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/34/5c228c4a7a2880defc1ea9809ca61b9010176abf/pycompat.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13721, 4176, 12, 2977, 379, 16, 3772, 3789, 4672, 1002, 13721, 29924, 2932, 4538, 7708, 7923, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13721, 4176, 12, 2977, 379, 16, 3772, 3789, 4672, 1002, 13721, 29924, 2932, 4538, 7708, 7923, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
license_path = os.path.join(path, filename)
if filename.startswith('/'): license_path = os.path.join(os.getcwd(), filename[1:]) else: license_path = os.path.join(path, filename)
def ParseDir(path): """Examine a third_party/foo component and extract its metadata.""" # Parse metadata fields out of README.chromium. # We examine "LICENSE" for the license file by default. metadata = { "License File": "LICENSE", # Relative path to license text. "Name": None, # Short name (for header on about:credits). "URL": None, # Project home page. } if path in SPECIAL_CASES: metadata.update(SPECIAL_CASES[path]) else: # Try to find README.chromium. readme_path = os.path.join(path, 'README.chromium') if not os.path.exists(readme_path): raise LicenseError("missing README.chromium") for line in open(readme_path): line = line.strip() if not line: break for key in metadata.keys(): field = key + ": " if line.startswith(field): metadata[key] = line[len(field):] # Check that all expected metadata is present. for key, value in metadata.iteritems(): if not value: raise LicenseError("couldn't find '" + key + "' line " "in README.chromium or licences.py " "SPECIAL_CASES") # Check that the license file exists. for filename in (metadata["License File"], "COPYING"): license_path = os.path.join(path, filename) if os.path.exists(license_path): metadata["License File"] = filename break license_path = None if not license_path: raise LicenseError("License file not found. " "Either add a file named LICENSE, " "import upstream's COPYING if available, " "or add a 'License File:' line to README.chromium " "with the appropriate path.") return metadata
8283b10cafab170f6dfd7f266163aefc223892fe /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5060/8283b10cafab170f6dfd7f266163aefc223892fe/licenses.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2884, 1621, 12, 803, 4672, 3536, 424, 301, 558, 279, 12126, 67, 21214, 19, 11351, 1794, 471, 2608, 2097, 1982, 12123, 225, 468, 2884, 1982, 1466, 596, 434, 10746, 958, 18, 16298, 5077, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2884, 1621, 12, 803, 4672, 3536, 424, 301, 558, 279, 12126, 67, 21214, 19, 11351, 1794, 471, 2608, 2097, 1982, 12123, 225, 468, 2884, 1982, 1466, 596, 434, 10746, 958, 18, 16298, 5077, 1...
cr = db.cursor()
def list(self): db = sql_db.db_connect('template1') cr = db.cursor() try: try: cr = db.cursor() db_user = tools.config["db_user"] if not db_user and os.name == 'posix': import pwd db_user = pwd.getpwuid(os.getuid())[0] if not db_user: cr.execute("select decode(usename, 'escape') from pg_user where usesysid=(select datdba from pg_database where datname=%s)", (tools.config["db_name"],)) res = cr.fetchone() db_user = res and str(res[0]) if db_user: cr.execute("select decode(datname, 'escape') from pg_database where datdba=(select usesysid from pg_user where usename=%s) and datname not in ('template0', 'template1', 'postgres')", (db_user,)) else: cr.execute("select decode(datname, 'escape') from pg_database where datname not in('template0', 'template1','postgres')") res = [str(name) for (name,) in cr.fetchall()] except: res = [] finally: cr.close() sql_db.close_db('template1') res.sort() return res
dbdd4c582ec44f0cc8ce5522a1b0d6d71d1d6536 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7397/dbdd4c582ec44f0cc8ce5522a1b0d6d71d1d6536/web_services.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 666, 12, 2890, 4672, 1319, 273, 1847, 67, 1966, 18, 1966, 67, 3612, 2668, 3202, 21, 6134, 775, 30, 775, 30, 1319, 67, 1355, 273, 8513, 18, 1425, 9614, 1966, 67, 1355, 11929, 309, 486, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 666, 12, 2890, 4672, 1319, 273, 1847, 67, 1966, 18, 1966, 67, 3612, 2668, 3202, 21, 6134, 775, 30, 775, 30, 1319, 67, 1355, 273, 8513, 18, 1425, 9614, 1966, 67, 1355, 11929, 309, 486, ...
def print_decls(decldict):
def print_decls(decldict, value):
def print_decls(decldict): for t, d in decldict.items(): d = d.keys() if not d: continue d.sort() print 'syn keyword %s%s %s' % (syntax_name, normalize(t), ' '.join(d))
f7d9779be336dd165ada7dd3923f3f7d87d4a728 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6171/f7d9779be336dd165ada7dd3923f3f7d87d4a728/vim-syn-gen.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1172, 67, 323, 6429, 12, 8840, 1576, 16, 460, 4672, 364, 268, 16, 302, 316, 3496, 1576, 18, 3319, 13332, 302, 273, 302, 18, 2452, 1435, 309, 486, 302, 30, 1324, 302, 18, 3804, 1435, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1172, 67, 323, 6429, 12, 8840, 1576, 16, 460, 4672, 364, 268, 16, 302, 316, 3496, 1576, 18, 3319, 13332, 302, 273, 302, 18, 2452, 1435, 309, 486, 302, 30, 1324, 302, 18, 3804, 1435, ...
elif UName == "osf1":
elif UName == "osf1":
def GetVersion(Key): # # Only read the file if an environment variable is "missing" (they # usually all are, ok), and only read it once. # #print("WriteVariablesIntoEnvironment:3") Value = Versions.get(Key) if Value: return Value # # CM3VERSION d5.7.1 # CM3VERSIONNUM 050701 # CM3LASTCHANGED 2009-01-21 # RegExp = re.compile("(" + "|".join(Versions.keys()) + ") (.+)$", re.IGNORECASE) ShFilePath = os.path.join(os.path.dirname(os.path.dirname(os.path.abspath(__file__))), "version") for Line in open(ShFilePath): Match = RegExp.match(Line) if Match: MatchKey = Match.group(1) # # We are here because one of them wasn't found, but we should be # sure only to overwrite what we don't have. # if not Versions[MatchKey]: Value = Match.group(2) Versions[MatchKey] = Value exec("%s = \"%s\"" % (MatchKey, Value), locals(), globals()) # # Make sure we found every key in the file (at least those # not defined in the environment) # MissingKey = None for Item in Versions.iteritems(): #print(Item) if Item[1] is None: MissingKey = Item[0] File = __file__ sys.stderr.write("%(File)s: %(MissingKey)s not found in %(ShFilePath)s\n" % vars()) if MissingKey: sys.exit(1) return Versions.get(Key)
a5382390fc1645f62582e9440bae90e340b3d767 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9328/a5382390fc1645f62582e9440bae90e340b3d767/pylib.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 968, 1444, 12, 653, 4672, 468, 468, 5098, 855, 326, 585, 309, 392, 3330, 2190, 353, 315, 7337, 6, 261, 451, 402, 468, 11234, 777, 854, 16, 1529, 3631, 471, 1338, 855, 518, 3647, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 968, 1444, 12, 653, 4672, 468, 468, 5098, 855, 326, 585, 309, 392, 3330, 2190, 353, 315, 7337, 6, 261, 451, 402, 468, 11234, 777, 854, 16, 1529, 3631, 471, 1338, 855, 518, 3647, 18, ...
self.assertEqual(sys.exc_info()[0], 'OverflowError')
self.assertEqual(sys.exc_info()[0], OverflowError)
def _raise_during_standard_error_message(*args, **kw): raise ZeroDivisionError('testing')
706b2a37f5c717c21c45c27fefdb17762443caf2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9658/706b2a37f5c717c21c45c27fefdb17762443caf2/testSimpleItem.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 11628, 67, 72, 4017, 67, 10005, 67, 1636, 67, 2150, 30857, 1968, 16, 2826, 9987, 4672, 1002, 12744, 28223, 668, 2668, 3813, 310, 6134, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 11628, 67, 72, 4017, 67, 10005, 67, 1636, 67, 2150, 30857, 1968, 16, 2826, 9987, 4672, 1002, 12744, 28223, 668, 2668, 3813, 310, 6134, 2, -100, -100, -100, -100, -100, -100, -100, -...
if data[f] != False:
if data[f]: data[f].sort()
def copy_data(self, cr, uid, id, default=None, context=None): """ Copy given record's data with all its fields values
01f2d0c0db3743d0f4c0b2f6e774549050434cc2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12853/01f2d0c0db3743d0f4c0b2f6e774549050434cc2/orm.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1610, 67, 892, 12, 2890, 16, 4422, 16, 4555, 16, 612, 16, 805, 33, 7036, 16, 819, 33, 7036, 4672, 3536, 5631, 864, 1409, 1807, 501, 598, 777, 2097, 1466, 924, 2, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1610, 67, 892, 12, 2890, 16, 4422, 16, 4555, 16, 612, 16, 805, 33, 7036, 16, 819, 33, 7036, 4672, 3536, 5631, 864, 1409, 1807, 501, 598, 777, 2097, 1466, 924, 2, -100, -100, -100, -1...
console.stdoutln("NOTE: Skipping Jmol-viewer section")
console.stdoutln("NOTE: Skipping Jmol-viewer section")
def html_tls_graph_path(self, chain, ntls): """Generates the HTML table describing the path (set of tls groups) for the given number of segments(h, or ntls) """ ## class HTMLReport() cpartition = chain.partition_collection.get_chain_partition(ntls) if cpartition == None: return None
80aaf2a14dfcc5b684aad61f8ca72e81b353162c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10674/80aaf2a14dfcc5b684aad61f8ca72e81b353162c/html.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1729, 67, 17116, 67, 4660, 67, 803, 12, 2890, 16, 2687, 16, 9513, 3251, 4672, 3536, 6653, 326, 3982, 1014, 16868, 326, 589, 261, 542, 434, 6871, 3252, 13, 364, 326, 864, 1300, 434, 515...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1729, 67, 17116, 67, 4660, 67, 803, 12, 2890, 16, 2687, 16, 9513, 3251, 4672, 3536, 6653, 326, 3982, 1014, 16868, 326, 589, 261, 542, 434, 6871, 3252, 13, 364, 326, 864, 1300, 434, 515...
if calname == "":
if calname is None:
def importProcess(self, text, extension=None, item=None, changes=None, previousView=None, updateCallback=None): # the item parameter is so that a share item can be passed in for us # to populate.
c4fe9d0f1a7be58d33664be3c6eb832fd34a787d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/c4fe9d0f1a7be58d33664be3c6eb832fd34a787d/ICalendar.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1930, 2227, 12, 2890, 16, 977, 16, 2710, 33, 7036, 16, 761, 33, 7036, 16, 3478, 33, 7036, 16, 2416, 1767, 33, 7036, 16, 1089, 2428, 33, 7036, 4672, 468, 326, 761, 1569, 353, 1427, 71...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1930, 2227, 12, 2890, 16, 977, 16, 2710, 33, 7036, 16, 761, 33, 7036, 16, 3478, 33, 7036, 16, 2416, 1767, 33, 7036, 16, 1089, 2428, 33, 7036, 4672, 468, 326, 761, 1569, 353, 1427, 71...
schedIdList += " " + \ str( job.runningJob['schedulerId'] ).strip() command = "glite-wms-job-cancel --json --noint " + schedIdList out, ret = self.ExecuteCommand( self.proxyString + command ) if ret != 0 : raise SchedulerError('error executing glite-wms-job-cancel', out) elif ret == 0 and out.find("result: success") == -1 : raise SchedulerError('error', out)
jobsToKill.append(str( job.runningJob['schedulerId'] ).strip()) chunk = lambda ulist, step: map(lambda i: ulist[i:i+step], xrange(0, len(ulist), step)) lljobs = chunk(jobsToKill, self.killThreshold) for x in lljobs : schedIdList = ' '.join(x) command = "glite-wms-job-cancel --json --noint " + schedIdList out, ret = self.ExecuteCommand( self.proxyString + command ) if ret != 0 : raise SchedulerError('error executing glite-wms-job-cancel', out) elif ret == 0 and out.find("result: success") == -1 : raise SchedulerError('error', out) return 0
def kill( self, obj ): """ kill job """ # the object passed is a job if type(obj) == Job and self.valid( obj.runningJob ): # check for the RunningJob integrity schedIdList = str( obj.runningJob['schedulerId'] ).strip() # the object passed is a Task elif type(obj) == Task : schedIdList = "" for job in obj.jobs: if not self.valid( job.runningJob ): continue schedIdList += " " + \ str( job.runningJob['schedulerId'] ).strip() command = "glite-wms-job-cancel --json --noint " + schedIdList out, ret = self.ExecuteCommand( self.proxyString + command ) if ret != 0 : raise SchedulerError('error executing glite-wms-job-cancel', out) elif ret == 0 and out.find("result: success") == -1 : raise SchedulerError('error', out)
53423e453953da739f4a02830391dc962d257904 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8886/53423e453953da739f4a02830391dc962d257904/SchedulerGLite.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8673, 12, 365, 16, 1081, 262, 30, 3536, 8673, 1719, 3536, 225, 468, 326, 733, 2275, 353, 279, 1719, 309, 618, 12, 2603, 13, 422, 3956, 471, 365, 18, 877, 12, 1081, 18, 8704, 2278, 26...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8673, 12, 365, 16, 1081, 262, 30, 3536, 8673, 1719, 3536, 225, 468, 326, 733, 2275, 353, 279, 1719, 309, 618, 12, 2603, 13, 422, 3956, 471, 365, 18, 877, 12, 1081, 18, 8704, 2278, 26...
f = sqrt(distnorm/sum)
f = sqrt(distnorm/max(sum,1e-6))
def LSMT(self): # optimize the distance transformation # build vector o effect = 0 self.getDistance() o = [] for i in xrange(1,self.n): for j in xrange(i): o.append((self.originalDistances[i,j],(i,j))) o.sort(_mycompare) # find the ties in o, and construct the d vector sorting in order within ties d = [] td = [] uv = [] # numbers of consecutively tied o values (i,j) = o[0][1] distnorm = self.projectedDistances[i,j]*self.projectedDistances[i,j] td = [self.projectedDistances[i,j]] # fetch distance for l in xrange(1,len(o)): # copy now sorted distances in an array # but sort distances within a tied o (i,j) = o[l][1] cd = self.projectedDistances[i,j] distnorm += self.projectedDistances[i,j]*self.projectedDistances[i,j] if o[l][0] != o[l-1][0]: # differing value, flush sum = reduce(lambda x,y:x+y,td)+0.0 d.append([sum,len(td),sum/len(td),td]) td = [] td.append(cd) sum = reduce(lambda x,y:x+y,td)+0.0 d.append([sum,len(td),sum/len(td),td]) #### # keep merging non-monotonous areas in d monotony = 0 while not monotony and len(d) > 1: monotony = 1 pi = 0 # index n = 1 # n-areas nd = [] r = d[0] # current area for i in range(1,len(d)): tr = d[i] if r[2]>=tr[2]: monotony = 0 effect = 1 r[0] += tr[0] r[1] += tr[1] r[2] = tr[0]/tr[1] r[3] += tr[3] else: nd.append(r) r = tr nd.append(r) d = nd # normalizing multiplier sum = 0.0 for i in d: sum += i[2]*i[2]*i[1] f = sqrt(distnorm/sum) # transform O k = 0 for i in d: for j in range(i[1]): (ii,jj) = o[k][1] self.distances[ii,jj] = f*i[2] k += 1 assert(len(o) == k) self.freshD = 0 return effect
c474478c97cb0593e748e02c9775117c045b1180 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6366/c474478c97cb0593e748e02c9775117c045b1180/orngMDS.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 29204, 6152, 12, 2890, 4672, 468, 10979, 326, 3888, 8620, 468, 1361, 3806, 320, 5426, 273, 374, 365, 18, 588, 7200, 1435, 320, 273, 5378, 364, 277, 316, 12314, 12, 21, 16, 2890, 18, 82...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 29204, 6152, 12, 2890, 4672, 468, 10979, 326, 3888, 8620, 468, 1361, 3806, 320, 5426, 273, 374, 365, 18, 588, 7200, 1435, 320, 273, 5378, 364, 277, 316, 12314, 12, 21, 16, 2890, 18, 82...
vbox.pack_start(hbox, gtk.TRUE, gtk.TRUE, 0)
vbox.pack_start(hbox, True, True, 0)
def __init__(self): self.return_keyval = gtk.gdk.keyval_from_name("Return") self.left_keyval = gtk.gdk.keyval_from_name("Left") self.right_keyval = gtk.gdk.keyval_from_name("Right") self.up_keyval = gtk.gdk.keyval_from_name("Up") self.down_keyval = gtk.gdk.keyval_from_name("Down") self.backspace_keyval = gtk.gdk.keyval_from_name("BackSpace") self.delete_keyval = gtk.gdk.keyval_from_name("Delete") self.home_keyval = gtk.gdk.keyval_from_name("Home") self.end_keyval = gtk.gdk.keyval_from_name("End")
2744524bbe73c70a7af8961c0bcea5e15fbb0a13 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3867/2744524bbe73c70a7af8961c0bcea5e15fbb0a13/ipmigui.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 4672, 365, 18, 2463, 67, 856, 1125, 273, 22718, 18, 75, 2883, 18, 856, 1125, 67, 2080, 67, 529, 2932, 990, 7923, 365, 18, 4482, 67, 856, 1125, 273, 22718, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 4672, 365, 18, 2463, 67, 856, 1125, 273, 22718, 18, 75, 2883, 18, 856, 1125, 67, 2080, 67, 529, 2932, 990, 7923, 365, 18, 4482, 67, 856, 1125, 273, 22718, ...
mul1s rax, t1 mulel rax muleh ah, flags=(OF,CF)
mul1s rax, t1, flags=(OF,CF) mulel rax muleh ah
def macroop IMUL_B_M
b64d0bdeda1662091746c3695b4429fcc6f69342 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6787/b64d0bdeda1662091746c3695b4429fcc6f69342/multiply_and_divide.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 11522, 556, 6246, 1506, 67, 38, 67, 49, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 11522, 556, 6246, 1506, 67, 38, 67, 49, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
path = [module.__filename__]
head, tail = os.path.split(module.__filename__) path = [head]
def reload(self, module, path=None): if path is None and hasattr(module, '__filename__'): path = [module.__filename__] return ihooks.ModuleImporter.reload(self, module, path)
9eca6c4d59803ed1f62627fa0024674267283d18 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/9eca6c4d59803ed1f62627fa0024674267283d18/rexec.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7749, 12, 2890, 16, 1605, 16, 589, 33, 7036, 4672, 309, 589, 353, 599, 471, 3859, 12, 2978, 16, 4940, 3459, 7250, 4672, 910, 16, 5798, 273, 1140, 18, 803, 18, 4939, 12, 2978, 16186, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7749, 12, 2890, 16, 1605, 16, 589, 33, 7036, 4672, 309, 589, 353, 599, 471, 3859, 12, 2978, 16, 4940, 3459, 7250, 4672, 910, 16, 5798, 273, 1140, 18, 803, 18, 4939, 12, 2978, 16186, ...
if (sys.platform == "win32"):
if (sys.platform != "win32"): CopyAllHeaders('panda/src/glxdisplay') else:
def CreatePandaVersionFiles(): version1=int(VERSION.split(".")[0]) version2=int(VERSION.split(".")[1]) version3=int(VERSION.split(".")[2]) nversion=version1*1000000+version2*1000+version3 pandaversion_h = PANDAVERSION_H.replace("VERSION1",str(version1)) pandaversion_h = pandaversion_h.replace("VERSION2",str(version2)) pandaversion_h = pandaversion_h.replace("VERSION3",str(version3)) pandaversion_h = pandaversion_h.replace("NVERSION",str(nversion)) checkpandaversion_cxx = CHECKPANDAVERSION_CXX.replace("VERSION1",str(version1)) checkpandaversion_cxx = checkpandaversion_cxx.replace("VERSION2",str(version2)) checkpandaversion_cxx = checkpandaversion_cxx.replace("VERSION3",str(version3)) checkpandaversion_cxx = checkpandaversion_cxx.replace("NVERSION",str(nversion)) checkpandaversion_h = CHECKPANDAVERSION_H.replace("VERSION1",str(version1)) checkpandaversion_h = checkpandaversion_h.replace("VERSION2",str(version2)) checkpandaversion_h = checkpandaversion_h.replace("VERSION3",str(version3)) checkpandaversion_h = checkpandaversion_h.replace("NVERSION",str(nversion)) ConditionalWriteFile('built/include/pandaVersion.h', pandaversion_h) ConditionalWriteFile('built/include/checkPandaVersion.cxx', checkpandaversion_cxx) ConditionalWriteFile('built/include/checkPandaVersion.h', checkpandaversion_h) ConditionalWriteFile("built/tmp/null.cxx","")
f7a6726c5099ebae926d5d09ee18a7f322e81784 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8543/f7a6726c5099ebae926d5d09ee18a7f322e81784/makepanda.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1788, 52, 464, 69, 1444, 2697, 13332, 1177, 21, 33, 474, 12, 5757, 18, 4939, 2932, 1199, 25146, 20, 5717, 1177, 22, 33, 474, 12, 5757, 18, 4939, 2932, 1199, 25146, 21, 5717, 1177, 23, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1788, 52, 464, 69, 1444, 2697, 13332, 1177, 21, 33, 474, 12, 5757, 18, 4939, 2932, 1199, 25146, 20, 5717, 1177, 22, 33, 474, 12, 5757, 18, 4939, 2932, 1199, 25146, 21, 5717, 1177, 23, ...
elif _isfunction(v) or _isclass(v):
elif _isfunction(v) or _isclass(v) or _ismethod(v):
def run__test__(self, d, name): """d, name -> Treat dict d like module.__test__.
5f8b0b1fd4801b4808223077643c24d41130b81e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/5f8b0b1fd4801b4808223077643c24d41130b81e/doctest.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 972, 3813, 972, 12, 2890, 16, 302, 16, 508, 4672, 3536, 72, 16, 508, 317, 26564, 2065, 302, 3007, 1605, 16186, 3813, 25648, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 972, 3813, 972, 12, 2890, 16, 302, 16, 508, 4672, 3536, 72, 16, 508, 317, 26564, 2065, 302, 3007, 1605, 16186, 3813, 25648, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100,...
<unittest.TestResult run=3 errors=0 failures=3>
<unittest.TestResult run=3 errors=0 failures=2>
def test_DocFileSuite(): """We can test tests found in text files using a DocFileSuite. We create a suite by providing the names of one or more text files that include examples: >>> import unittest >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt') >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=3> The test files are looked for in the directory containing the calling module. A package keyword argument can be provided to specify a different relative location. >>> import unittest >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt', ... package='test') >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=3> '/' should be used as a path separator. It will be converted to a native separator at run time: >>> suite = doctest.DocFileSuite('../test/test_doctest.txt') >>> suite.run(unittest.TestResult()) <unittest.TestResult run=1 errors=0 failures=1> If DocFileSuite is used from an interactive session, then files are resolved relative to the directory of sys.argv[0]: >>> import new, os.path, test.test_doctest >>> save_argv = sys.argv >>> sys.argv = [test.test_doctest.__file__] >>> suite = doctest.DocFileSuite('test_doctest.txt', ... package=new.module('__main__')) >>> sys.argv = save_argv By setting `module_relative=False`, os-specific paths may be used (including absolute paths and paths relative to the working directory): >>> # Get the absolute path of the test package. >>> test_doctest_path = os.path.abspath(test.test_doctest.__file__) >>> test_pkg_path = os.path.split(test_doctest_path)[0] >>> # Use it to find the absolute path of test_doctest.txt. >>> test_file = os.path.join(test_pkg_path, 'test_doctest.txt') >>> suite = doctest.DocFileSuite(test_file, module_relative=False) >>> suite.run(unittest.TestResult()) <unittest.TestResult run=1 errors=0 failures=1> It is an error to specify `package` when `module_relative=False`: >>> suite = doctest.DocFileSuite(test_file, module_relative=False, ... package='test') Traceback (most recent call last): ValueError: Package may only be specified for module-relative paths. You can specify initial global variables: >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt', ... globs={'favorite_color': 'blue'}) >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=2> In this case, we supplied a missing favorite color. You can provide doctest options: >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt', ... optionflags=doctest.DONT_ACCEPT_BLANKLINE, ... globs={'favorite_color': 'blue'}) >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=3> And, you can provide setUp and tearDown functions: You can supply setUp and teatDoen functions: >>> def setUp(t): ... import test.test_doctest ... test.test_doctest.sillySetup = True >>> def tearDown(t): ... import test.test_doctest ... del test.test_doctest.sillySetup Here, we installed a silly variable that the test expects: >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt', ... setUp=setUp, tearDown=tearDown) >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=2> But the tearDown restores sanity: >>> import test.test_doctest >>> test.test_doctest.sillySetup Traceback (most recent call last): ... AttributeError: 'module' object has no attribute 'sillySetup' The setUp and tearDown funtions are passed test objects. Here, we'll use a setUp function to set the favorite color in test_doctest.txt: >>> def setUp(test): ... test.globs['favorite_color'] = 'blue' >>> suite = doctest.DocFileSuite('test_doctest.txt', setUp=setUp) >>> suite.run(unittest.TestResult()) <unittest.TestResult run=1 errors=0 failures=0> Here, we didn't need to use a tearDown function because we modified the test globals. The test globals are automatically cleared for us after a test. Tests in a file run using `DocFileSuite` can also access the `__file__` global, which is set to the name of the file containing the tests: >>> suite = doctest.DocFileSuite('test_doctest3.txt') >>> suite.run(unittest.TestResult()) <unittest.TestResult run=1 errors=0 failures=0> If the tests contain non-ASCII characters, we have to specify which encoding the file is encoded with. We do so by using the `encoding` parameter: >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt', ... encoding='utf-8') >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=2> """
b1a9f2772ba21f268a8164908714e5a106be50d1 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8546/b1a9f2772ba21f268a8164908714e5a106be50d1/test_doctest.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 1759, 812, 13587, 13332, 3536, 3218, 848, 1842, 7434, 1392, 316, 977, 1390, 1450, 279, 3521, 812, 13587, 18, 225, 1660, 752, 279, 11371, 635, 17721, 326, 1257, 434, 1245, 578, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 1759, 812, 13587, 13332, 3536, 3218, 848, 1842, 7434, 1392, 316, 977, 1390, 1450, 279, 3521, 812, 13587, 18, 225, 1660, 752, 279, 11371, 635, 17721, 326, 1257, 434, 1245, 578, ...
path = os.path.join(self.configpath, LAYER_DIRS[layer])
path = os.path.join(self.mapServ_inst.configpath, LAYER_DIRS[layer])
def coord_to_path(self, tile_coord, layer): self.lock.acquire() path = os.path.join(self.configpath, LAYER_DIRS[layer]) path = fileUtils.check_dir(path) path = fileUtils.check_dir(path, '%d' % tile_coord[2]) path = fileUtils.check_dir(path, "%d" % (tile_coord[0] / 1024)) path = fileUtils.check_dir(path, "%d" % (tile_coord[0] % 1024)) path = fileUtils.check_dir(path, "%d" % (tile_coord[1] / 1024)) self.lock.release() return os.path.join(path, "%d.png" % (tile_coord[1] % 1024))
e47ed51cf02f70923536b0b3bfe9e89d6c1e5fbf /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5373/e47ed51cf02f70923536b0b3bfe9e89d6c1e5fbf/tilesRepoFS.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2745, 67, 869, 67, 803, 12, 2890, 16, 4769, 67, 5732, 16, 3018, 4672, 365, 18, 739, 18, 1077, 1039, 1435, 589, 273, 1140, 18, 803, 18, 5701, 12, 2890, 18, 1458, 827, 90, 67, 8591, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2745, 67, 869, 67, 803, 12, 2890, 16, 4769, 67, 5732, 16, 3018, 4672, 365, 18, 739, 18, 1077, 1039, 1435, 589, 273, 1140, 18, 803, 18, 5701, 12, 2890, 18, 1458, 827, 90, 67, 8591, ...
return self.tk.call(self._w, 'info', 'right', index)
return self.tk.call(self._w, 'info', 'right', index)
def info_right(self, index):
0049f0036898e656085d5c214ba07642a9c25a6e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/0049f0036898e656085d5c214ba07642a9c25a6e/Tix.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1123, 67, 4083, 12, 2890, 16, 770, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1123, 67, 4083, 12, 2890, 16, 770, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
r""" A simple periodic sequence. INPUT: n -- non negative integer OUTPUT: integer -- function value EXAMPLES: sage: a = sloane.A000035;a
def __init__(self): r"""
def _eval(self, n): return Integer(2*n)
69a199217d612e1c590af73e16003812c85b93ec /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/69a199217d612e1c590af73e16003812c85b93ec/sloane_functions.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 8622, 12, 2890, 16, 290, 4672, 327, 2144, 12, 22, 14, 82, 13, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 8622, 12, 2890, 16, 290, 4672, 327, 2144, 12, 22, 14, 82, 13, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
outputs[-1:] = \
outputs.extend (
def copy_tree (src, dst, preserve_mode=1, preserve_times=1, preserve_symlinks=0, update=0, verbose=0, dry_run=0): """Copy an entire directory tree 'src' to a new location 'dst'. Both 'src' and 'dst' must be directory names. If 'src' is not a directory, raise DistutilsFileError. If 'dst' does not exist, it is created with 'mkpath()'. The end result of the copy is that every file in 'src' is copied to 'dst', and directories under 'src' are recursively copied to 'dst'. Return the list of files copied (under their output names) -- note that if 'update' is true, this might be less than the list of files considered. Return value is not affected by 'dry_run'. 'preserve_mode' and 'preserve_times' are the same as for 'copy_file'; note that they only apply to regular files, not to directories. If 'preserve_symlinks' is true, symlinks will be copied as symlinks (on platforms that support them!); otherwise (the default), the destination of the symlink will be copied. 'update' and 'verbose' are the same as for 'copy_file'.""" if not dry_run and not os.path.isdir (src): raise DistutilsFileError, \ "cannot copy tree %s: not a directory" % src try: names = os.listdir (src) except os.error, (errno, errstr): if dry_run: names = [] else: raise DistutilsFileError, \ "error listing files in %s: %s" % (src, errstr) if not dry_run: mkpath (dst, verbose=verbose) outputs = [] for n in names: src_name = os.path.join (src, n) dst_name = os.path.join (dst, n) if preserve_symlinks and os.path.islink (src_name): link_dest = os.readlink (src_name) if verbose: print "linking %s -> %s" % (dst_name, link_dest) if not dry_run: os.symlink (link_dest, dst_name) outputs.append (dst_name) elif os.path.isdir (src_name): outputs[-1:] = \ copy_tree (src_name, dst_name, preserve_mode, preserve_times, preserve_symlinks, update, verbose, dry_run) else: if (copy_file (src_name, dst_name, preserve_mode, preserve_times, update, verbose, dry_run)): outputs.append (dst_name) return outputs
f49afccb1be8414afed28dee04df060fdc754715 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/f49afccb1be8414afed28dee04df060fdc754715/util.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1610, 67, 3413, 261, 4816, 16, 3046, 16, 9420, 67, 3188, 33, 21, 16, 9420, 67, 8293, 33, 21, 16, 9420, 67, 21278, 87, 33, 20, 16, 1089, 33, 20, 16, 3988, 33, 20, 16, 10299, 67, 2...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1610, 67, 3413, 261, 4816, 16, 3046, 16, 9420, 67, 3188, 33, 21, 16, 9420, 67, 8293, 33, 21, 16, 9420, 67, 21278, 87, 33, 20, 16, 1089, 33, 20, 16, 3988, 33, 20, 16, 10299, 67, 2...
process_list = get_cmd_output('ps -e -o comm,user=').strip().split('\n')
process_list = commands.getoutput('ps -e -o comm,user=').strip().split('\n')
def testing_status() : server_name = 'rcssserver' server_user = '' process_list = get_cmd_output('ps -e -o comm,user=').strip().split('\n') process_list.pop(0) team_count_map = {} cmd_count_map = {} matched_cmds = {} for process in process_list : info = process.split() (cmd, user) = (info[0], info[1]) cmd_count_map[cmd] = 1 + cmd_count_map.get(cmd, 0) for pattern in team_name_map.keys() : if pattern.match(cmd) : matched_cmds[cmd] = 1 team_name = team_name_map[pattern] team_count_map[team_name] = 1 + team_count_map.get(team_name, 0) break if not server_user and cmd == server_name : server_user = user message = ' #rcssserver: ' if server_user : server_count = cmd_count_map[server_name] if len(team_count_map) <= 1 : count_cmd_map = {} sucess = False for cmd in cmd_count_map.keys() : count_cmd_map.setdefault(cmd_count_map[cmd], []).append(cmd) count_list = count_cmd_map.keys() count_list.sort() count_list.reverse() for count in count_list : for cmd in count_cmd_map[count] : if not matched_cmds.has_key(cmd) : team_count_map['[' + cmd + ']'] = cmd_count_map[cmd] if len(team_count_map) >= 2 : sucess = True break if sucess : break message += '%d, %s, (' % (server_count, server_user) for team in sorted(team_count_map.keys()) : message += '%s x %d, ' % (team, team_count_map[team]) message = message.rstrip(', ') + ')' else : message += '0' return message
004ed3b935e239e61feea17b4ad35f4c4ed06166 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/228/004ed3b935e239e61feea17b4ad35f4c4ed06166/client.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7769, 67, 2327, 1435, 294, 1438, 67, 529, 273, 296, 1310, 1049, 3567, 11, 1438, 67, 1355, 273, 875, 225, 1207, 67, 1098, 273, 4364, 18, 588, 2844, 2668, 1121, 300, 73, 300, 83, 1543, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7769, 67, 2327, 1435, 294, 1438, 67, 529, 273, 296, 1310, 1049, 3567, 11, 1438, 67, 1355, 273, 875, 225, 1207, 67, 1098, 273, 4364, 18, 588, 2844, 2668, 1121, 300, 73, 300, 83, 1543, ...
elif text.find('
elif text.find('
#enddef - closes a scope
0ae3bc473d70b671840d2febe0a653a31a4589b5 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9355/0ae3bc473d70b671840d2febe0a653a31a4589b5/wmliterator.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 468, 409, 536, 300, 13445, 279, 2146, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 468, 409, 536, 300, 13445, 279, 2146, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10...
self.enableCmdBtns() self.enableHistBtns()
self.enableCmdButtons() self.enableHistButtons()
def __init__(self, master, actor,
d1909c3750f356969ac52bbe45612907850f17e7 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6354/d1909c3750f356969ac52bbe45612907850f17e7/GuideWdg.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 4171, 16, 8327, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 4171, 16, 8327, 16, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
connection = MTurkConnection(operation.action)
connection = MTurkConnection(action)
def operation_execute(key): operation = datastore.get(key) if not operation.completed and not operation.error: connection = MTurkConnection(operation.action) try: operation.execute(connection) self.completed = datetime.now() except (BotoClientError, BotoServerError), response: self.error = response_error(response) self.put()
b4cc6abefee1ce6aef654df5be65743d01c92c4c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14392/b4cc6abefee1ce6aef654df5be65743d01c92c4c/main.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1674, 67, 8837, 12, 856, 4672, 1674, 273, 9290, 18, 588, 12, 856, 13, 225, 309, 486, 1674, 18, 13615, 471, 486, 1674, 18, 1636, 30, 1459, 273, 22068, 20628, 1952, 12, 1128, 13, 225, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1674, 67, 8837, 12, 856, 4672, 1674, 273, 9290, 18, 588, 12, 856, 13, 225, 309, 486, 1674, 18, 13615, 471, 486, 1674, 18, 1636, 30, 1459, 273, 22068, 20628, 1952, 12, 1128, 13, 225, ...
loader = importer.find_module(modname) if hasattr(loader,'get_source'):
try: loader = importer.find_module(modname) except SyntaxError: continue if hasattr(loader, 'get_source'):
def run(self, callback, key=None, completer=None, onerror=None): if key: key = key.lower() self.quit = False seen = {}
126c879b406f24363bdfccffa16e62acc695aea1 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8546/126c879b406f24363bdfccffa16e62acc695aea1/pydoc.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 16, 1348, 16, 498, 33, 7036, 16, 31848, 33, 7036, 16, 20991, 33, 7036, 4672, 309, 498, 30, 498, 273, 498, 18, 8167, 1435, 365, 18, 27176, 273, 1083, 5881, 273, 2618, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 16, 1348, 16, 498, 33, 7036, 16, 31848, 33, 7036, 16, 20991, 33, 7036, 4672, 309, 498, 30, 498, 273, 498, 18, 8167, 1435, 365, 18, 27176, 273, 1083, 5881, 273, 2618, ...
raise ValueError, 'You must provide a source address for this\
raise ValueError, 'You must provide a source address for this \
def __call__(self): recipients = [str(mail.strip()) for mail in \ self.element.recipients.split(',')] mailhost = getToolByName(aq_inner(self.context), "MailHost") if not mailhost: raise ComponentLookupError, 'You must have a Mailhost utility to \
d21e9ee976da48d9609790851a11196e38318d7a /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12184/d21e9ee976da48d9609790851a11196e38318d7a/mail.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 1991, 972, 12, 2890, 4672, 12045, 273, 306, 701, 12, 4408, 18, 6406, 10756, 364, 4791, 316, 521, 365, 18, 2956, 18, 27925, 18, 4939, 12, 2187, 6134, 65, 4791, 2564, 273, 336, 636...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 1991, 972, 12, 2890, 4672, 12045, 273, 306, 701, 12, 4408, 18, 6406, 10756, 364, 4791, 316, 521, 365, 18, 2956, 18, 27925, 18, 4939, 12, 2187, 6134, 65, 4791, 2564, 273, 336, 636...
and job.runningJob['status'] != 'SD': :
and job.runningJob['status'] != 'SD':
def query(self, obj, objType='node') : """ query status and eventually other scheduler related information """
c39edd21061a0424fe50fad47d6984792fe4bfcf /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8886/c39edd21061a0424fe50fad47d6984792fe4bfcf/Scheduler.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 843, 12, 2890, 16, 1081, 16, 30078, 2218, 2159, 6134, 294, 3536, 843, 1267, 471, 18011, 1308, 8129, 3746, 1779, 3536, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 843, 12, 2890, 16, 1081, 16, 30078, 2218, 2159, 6134, 294, 3536, 843, 1267, 471, 18011, 1308, 8129, 3746, 1779, 3536, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
print 'called Break'
def trapit(what): print 'TRAPPED', what what.trap(juice.UnhandledRemoteJuiceError)
e56dcf564dc442d95c85eb44f592b6672bad21ad /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8962/e56dcf564dc442d95c85eb44f592b6672bad21ad/test_q2q.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 23034, 305, 12, 23770, 4672, 1172, 296, 4349, 2203, 12120, 2187, 4121, 4121, 18, 27170, 12, 78, 89, 1812, 18, 26952, 5169, 46, 89, 1812, 668, 13, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 23034, 305, 12, 23770, 4672, 1172, 296, 4349, 2203, 12120, 2187, 4121, 4121, 18, 27170, 12, 78, 89, 1812, 18, 26952, 5169, 46, 89, 1812, 668, 13, 2, -100, -100, -100, -100, -100, -100, ...
self.isBusy = False
def onCompleted(self, links): self.sayTo(self.employer, "Build complete.") self.isBusy = False for k in links.keys(): self.sayTo(self.employer, k + ": " + links[k])
f0130d66ba28b8a109656ac4672968689759b937 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10496/f0130d66ba28b8a109656ac4672968689759b937/buildbot.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 603, 9556, 12, 2890, 16, 4716, 4672, 365, 18, 87, 528, 774, 12, 2890, 18, 351, 2817, 264, 16, 315, 3116, 3912, 1199, 13, 364, 417, 316, 4716, 18, 2452, 13332, 365, 18, 87, 528, 774, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 603, 9556, 12, 2890, 16, 4716, 4672, 365, 18, 87, 528, 774, 12, 2890, 18, 351, 2817, 264, 16, 315, 3116, 3912, 1199, 13, 364, 417, 316, 4716, 18, 2452, 13332, 365, 18, 87, 528, 774, ...
pass
raise NotImplementedError
def index_shuffle(self, ind): """Change the indices.
34282af38d23e10b103d48f60c68963923bb99f3 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5735/34282af38d23e10b103d48f60c68963923bb99f3/constraints.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 770, 67, 31980, 12, 2890, 16, 1547, 4672, 3536, 3043, 326, 4295, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 770, 67, 31980, 12, 2890, 16, 1547, 4672, 3536, 3043, 326, 4295, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
path = path.replace('/', os.sep)
path = self._absnorm(path)
def wait_until_removed(self, path, timeout='1 minute'): """Waits until the given file or directory is removed.
4e92df2860c270b97129fb5577b1caec2faab8a2 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/6988/4e92df2860c270b97129fb5577b1caec2faab8a2/OperatingSystem.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2529, 67, 12198, 67, 14923, 12, 2890, 16, 589, 16, 2021, 2218, 21, 8044, 11, 4672, 3536, 26153, 3180, 326, 864, 585, 578, 1867, 353, 3723, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2529, 67, 12198, 67, 14923, 12, 2890, 16, 589, 16, 2021, 2218, 21, 8044, 11, 4672, 3536, 26153, 3180, 326, 864, 585, 578, 1867, 353, 3723, 18, 2, -100, -100, -100, -100, -100, -100, -1...
out += ch
out += chr(ord(ch))
def PyUnicode_EncodeUTF7(s, size, encodeSetO, encodeWhiteSpace, errors):
0e540b97b49394c1d38eed66d9b3eec5969d6c9a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6934/0e540b97b49394c1d38eed66d9b3eec5969d6c9a/unicodecodec.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4707, 16532, 67, 5509, 5159, 27, 12, 87, 16, 963, 16, 2017, 694, 51, 16, 2017, 23108, 16, 1334, 4672, 225, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4707, 16532, 67, 5509, 5159, 27, 12, 87, 16, 963, 16, 2017, 694, 51, 16, 2017, 23108, 16, 1334, 4672, 225, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -1...
semi = string.rfind(type, ';')
semi = type.rfind(';')
def open_data(self, url, data=None): """Use "data" URL.""" # ignore POSTed data # # syntax of data URLs: # dataurl := "data:" [ mediatype ] [ ";base64" ] "," data # mediatype := [ type "/" subtype ] *( ";" parameter ) # data := *urlchar # parameter := attribute "=" value import StringIO, mimetools, time try: [type, data] = string.split(url, ',', 1) except ValueError: raise IOError, ('data error', 'bad data URL') if not type: type = 'text/plain;charset=US-ASCII' semi = string.rfind(type, ';') if semi >= 0 and '=' not in type[semi:]: encoding = type[semi+1:] type = type[:semi] else: encoding = '' msg = [] msg.append('Date: %s'%time.strftime('%a, %d %b %Y %T GMT', time.gmtime(time.time()))) msg.append('Content-type: %s' % type) if encoding == 'base64': import base64 data = base64.decodestring(data) else: data = unquote(data) msg.append('Content-length: %d' % len(data)) msg.append('') msg.append(data) msg = string.join(msg, '\n') f = StringIO.StringIO(msg) headers = mimetools.Message(f, 0) f.fileno = None # needed for addinfourl return addinfourl(f, headers, url)
2a3596c7d04c9635576a712f85db07ce4de4bff2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/2a3596c7d04c9635576a712f85db07ce4de4bff2/urllib.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1696, 67, 892, 12, 2890, 16, 880, 16, 501, 33, 7036, 4672, 3536, 3727, 315, 892, 6, 1976, 12123, 468, 2305, 5485, 329, 501, 468, 468, 6279, 434, 501, 10414, 30, 468, 501, 718, 282, 5...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1696, 67, 892, 12, 2890, 16, 880, 16, 501, 33, 7036, 4672, 3536, 3727, 315, 892, 6, 1976, 12123, 468, 2305, 5485, 329, 501, 468, 468, 6279, 434, 501, 10414, 30, 468, 501, 718, 282, 5...
return False, "Cannot find block device %s" % (disk,)
_Fail("Cannot find block device %s", disk)
def BlockdevGrow(disk, amount): """Grow a stack of block devices. This function is called recursively, with the childrens being the first ones to resize. @type disk: L{objects.Disk} @param disk: the disk to be grown @rtype: (status, result) @return: a tuple with the status of the operation (True/False), and the errors message if status is False """ r_dev = _RecursiveFindBD(disk) if r_dev is None: return False, "Cannot find block device %s" % (disk,) try: r_dev.Grow(amount) except errors.BlockDeviceError, err: _Fail("Failed to grow block device: %s", err, exc=True) return True, None
afdc3985644be8d2b91d9f0e37dfda391a30d036 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7542/afdc3985644be8d2b91d9f0e37dfda391a30d036/backend.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3914, 5206, 30948, 12, 10863, 16, 3844, 4672, 3536, 30948, 279, 2110, 434, 1203, 7166, 18, 225, 1220, 445, 353, 2566, 8536, 16, 598, 326, 2325, 87, 3832, 326, 1122, 5945, 358, 7041, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3914, 5206, 30948, 12, 10863, 16, 3844, 4672, 3536, 30948, 279, 2110, 434, 1203, 7166, 18, 225, 1220, 445, 353, 2566, 8536, 16, 598, 326, 2325, 87, 3832, 326, 1122, 5945, 358, 7041, 18, ...
def __init__(self, toplot, xvals=None, yvals=None,
def __init__(self, data, xvals=None, yvals=None,
def __init__(self, toplot, xvals=None, yvals=None, binary=1, inline=_unset, **keyw): """GridData constructor.
57b7d171447e091edf5aff623b42289284f0cd78 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9569/57b7d171447e091edf5aff623b42289284f0cd78/__init__.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 501, 16, 619, 4524, 33, 7036, 16, 677, 4524, 33, 7036, 16, 3112, 33, 21, 16, 6370, 33, 67, 18579, 16, 2826, 856, 91, 4672, 3536, 6313, 751, 3885, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 501, 16, 619, 4524, 33, 7036, 16, 677, 4524, 33, 7036, 16, 3112, 33, 21, 16, 6370, 33, 67, 18579, 16, 2826, 856, 91, 4672, 3536, 6313, 751, 3885, 18, ...
iid = model.InvoiceItem(description='Penguin Dinner ticket (included in registraion)',
iid = model.InvoiceItem(description='Penguin Dinner ticket (included in registration)',
def _build_invoice(self): person = c.registration.person r = c.registration
4564a7b8c923eccde8041311a7f7d56b61a99834 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12856/4564a7b8c923eccde8041311a7f7d56b61a99834/registration.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 3510, 67, 16119, 12, 2890, 4672, 6175, 273, 276, 18, 14170, 18, 12479, 436, 273, 276, 18, 14170, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 3510, 67, 16119, 12, 2890, 4672, 6175, 273, 276, 18, 14170, 18, 12479, 436, 273, 276, 18, 14170, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10...
for base in (_HKEY_CLASSES_ROOT, _HKEY_LOCAL_MACHINE, _HKEY_CURRENT_USER, _HKEY_USERS):
for base in (HKEY_CLASSES_ROOT, HKEY_LOCAL_MACHINE, HKEY_CURRENT_USER, HKEY_USERS):
def get_msvc_paths (path, version='6.0', platform='x86'): """Get a list of devstudio directories (include, lib or path). Return a list of strings; will be empty list if unable to access the registry or appropriate registry keys not found.""" if not _can_read_reg: return [] L = [] if path=='lib': path= 'Library' path = string.upper(path + ' Dirs') K = ('Software\\Microsoft\\Devstudio\\%s\\' + 'Build System\\Components\\Platforms\\Win32 (%s)\\Directories') % \ (version,platform) for base in (_HKEY_CLASSES_ROOT, _HKEY_LOCAL_MACHINE, _HKEY_CURRENT_USER, _HKEY_USERS): try: k = _RegOpenKeyEx(base,K) i = 0 while 1: try: (p,v,t) = _RegEnumValue(k,i) if string.upper(p) == path: V = string.split(v,';') for v in V: if v == '' or v in L: continue L.append(v) break i = i + 1 except _RegError: break except _RegError: pass return L
f524abb944b7edf65acd200dfe2f4076696b42e3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/f524abb944b7edf65acd200dfe2f4076696b42e3/msvccompiler.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 959, 4227, 67, 4481, 261, 803, 16, 1177, 2218, 26, 18, 20, 2187, 4072, 2218, 92, 5292, 11, 4672, 3536, 967, 279, 666, 434, 4461, 334, 4484, 6402, 261, 6702, 16, 2561, 578, 5...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 959, 4227, 67, 4481, 261, 803, 16, 1177, 2218, 26, 18, 20, 2187, 4072, 2218, 92, 5292, 11, 4672, 3536, 967, 279, 666, 434, 4461, 334, 4484, 6402, 261, 6702, 16, 2561, 578, 5...
line = line[m.end() - 1:]
line = line[m.end(1):] elif empty: line = line[m.end(1):]
def subconvert(line, ofp, table, discards, autoclosing, knownempty, endchar=None): stack = [] while line: if line[0] == endchar and not stack: return line[1:] m = _comment_rx.match(line) if m: text = m.group(1) if text: ofp.write("(COMMENT\n") ofp.write("-%s\n" % encode(text)) ofp.write(")COMMENT\n") ofp.write("-\\n\n") else: ofp.write("-\\n\n") line = line[m.end():] continue m = _begin_env_rx.match(line) if m: # re-write to use the macro handler line = r"\%s%s" % (m.group(1), line[m.end():]) continue m =_end_env_rx.match(line) if m: # end of environment envname = m.group(1) if envname == "document": # special magic for n in stack[1:]: if n not in autoclosing: raise LaTeXFormatError("open element on stack: " + `n`) # should be more careful, but this is easier to code: stack = [] ofp.write(")document\n") elif envname == stack[-1]: ofp.write(")%s\n" % envname) del stack[-1] else:
fd9c2ca53cfbd9d2df1c1c71382345041d3849ae /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/fd9c2ca53cfbd9d2df1c1c71382345041d3849ae/latex2esis.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 720, 6283, 12, 1369, 16, 434, 84, 16, 1014, 16, 10388, 87, 16, 23173, 10910, 16, 4846, 5531, 16, 679, 3001, 33, 7036, 4672, 2110, 273, 5378, 1323, 980, 30, 309, 980, 63, 20, 65, 422,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 720, 6283, 12, 1369, 16, 434, 84, 16, 1014, 16, 10388, 87, 16, 23173, 10910, 16, 4846, 5531, 16, 679, 3001, 33, 7036, 4672, 2110, 273, 5378, 1323, 980, 30, 309, 980, 63, 20, 65, 422,...
if not type(obj) in [str, unicode]:
if not _is_in(type(obj), [str, unicode]):
def _require_string(obj): if not type(obj) in [str, unicode]: raise NamespaceRequirementError
3c551ef3827ddde6b2f4571bc3357a59857633a1 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7995/3c551ef3827ddde6b2f4571bc3357a59857633a1/namespace.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 6528, 67, 1080, 12, 2603, 4672, 309, 486, 389, 291, 67, 267, 12, 723, 12, 2603, 3631, 306, 701, 16, 5252, 65, 4672, 1002, 6005, 18599, 668, 565, 2, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 6528, 67, 1080, 12, 2603, 4672, 309, 486, 389, 291, 67, 267, 12, 723, 12, 2603, 3631, 306, 701, 16, 5252, 65, 4672, 1002, 6005, 18599, 668, 565, 2, -100, -100, -100, -100, -100, ...
row.coinc_event_id = coinc_event_id
cid = lsctables.CoincTable.get_next_id() row.coinc_event_id = cid
def populate_inspiral_tables(MBTA_frame, UID, set_keys = MBTA_set_keys, \ process_id = 'process:process_id:0', \ event_id_dict = insp_event_id_dict, \ coinc_event_id_base=InspiralCoincIdBase): """ create xml file and populate the SnglInspiral and CoincInspiral tables from a coinc .gwf file from MBTA xmldoc: xml file to append the tables to MBTA_frame: frame file to get info about triggers from set_keys: columns in the SnglInspiral Table to set process_id: process_id event_id_dict: {ifo:event_id} dictionary to assign event_id's coinc_event_id: coinc_event_id detectors: detectors participating in the coinc returns xmldoc and contents of the comment field """ #initialize xml document xmldoc = ligolw.Document() xmldoc.appendChild(ligolw.LIGO_LW()) #dictionaries to store about individual triggers end_time_s = {} end_time_ns = {} snr = {} mass1 = {} mass2 = {} Deff = {} mchirp = {} eta = {} #extract the information from the frame file events = Fr.frgetevent(MBTA_frame) #get the ifos from the event name for event in events: if 'MbtaHLV' in event['name']: detectors = H1L1V1_detlist elif 'MbtaHL' in event['name']: detectors = H1L1_detlist elif 'MbtaHV' in event['name']: detectors = H1V1_detlist elif 'MbtaH' in event['name']: detectors = H1_detlist elif 'MbtaLV' in event['name']: detectors = L1V1_detlist elif 'MbtaL' in event['name']: detectors = L1_detlist elif 'MbtaV' in event['name']: detectors = V1_detlist else: raise ValueError, "Invalid FrEvent name" log_data = event['comment'] + '\n' far = [line.split(':')[1].split()[0] for line in log_data.splitlines() if \ 'False Alarm Rate' in line][0] for ifo in detectors: end_time_s[ifo], end_time_ns[ifo] = str(event[ifo+':end_time']).split('.') snr[ifo] = float(event[ifo+':SNR']) mass1[ifo] = float(event[ifo+':mass1']) mass2[ifo] = float(event[ifo+':mass2']) mchirp[ifo], eta[ifo] = compute_mchirp_eta(mass1[ifo],mass2[ifo]) Deff[ifo] = float(event[ifo+':eff_distance']) #fill the SnglInspiralTable sin_table = lsctables.New(lsctables.SnglInspiralTable) xmldoc.childNodes[0].appendChild(sin_table) for ifo in detectors: row = sin_table.RowType() row.ifo = ifo row.search = 'MBTA' row.end_time = int(end_time_s[ifo]) row.end_time_ns = int(end_time_ns[ifo]) row.mass1 = mass1[ifo] row.mass2 = mass2[ifo] row.mchirp = mchirp[ifo] row.mtotal = mass1[ifo] + mass2[ifo] row.eta = eta[ifo] row.snr = snr[ifo] row.eff_distance = Deff[ifo] row.event_id = event_id_dict[ifo] row.process_id = process_id row.channel = '' #zero out the rest of the columns #should work in chi2 and chi2cut for key in sin_table.validcolumns.keys(): if key not in set_keys: setattr(row,key,None) sin_table.append(row) #CoincInspiralTable #using the conventions found in: #https://www.lsc-group.phys.uwm.edu/ligovirgo/cbcnote/S6Plan/ #090505160219S6PlanningNotebookCoinc_and_Experiment_Tables_ihope_implementation? #highlight=%28coinc%29|%28table%29 if len(detectors) < 2: return xmldoc, log_data, detectors coinc_event_id = coinc_event_id_base + str(UID) cin_table = lsctables.New(lsctables.CoincInspiralTable) xmldoc.childNodes[0].appendChild(cin_table) row = cin_table.RowType() row.set_ifos(detectors) row.coinc_event_id = coinc_event_id row.end_time = int(end_time_s['H1']) row.end_time_ns = int(end_time_ns['H1']) row.mass = (sum(mass1.values()) + sum(mass2.values()))/3 row.mchirp = sum(mchirp.values())/3 #the snr here is really the snr NOT effective snr row.snr = pow(sum([x*x for x in snr.values()]),0.5) #far is triggers/day row.false_alarm_rate = float(far) row.combined_far = 0 cin_table.append(row) return xmldoc, log_data, detectors
af257218e83ac728902869fa75f174902babf41f /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5758/af257218e83ac728902869fa75f174902babf41f/utils.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6490, 67, 267, 1752, 481, 287, 67, 9373, 12, 7969, 9833, 67, 3789, 16, 10034, 16, 444, 67, 2452, 273, 17897, 9833, 67, 542, 67, 2452, 16, 521, 1207, 67, 350, 273, 296, 2567, 30, 2567...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6490, 67, 267, 1752, 481, 287, 67, 9373, 12, 7969, 9833, 67, 3789, 16, 10034, 16, 444, 67, 2452, 273, 17897, 9833, 67, 542, 67, 2452, 16, 521, 1207, 67, 350, 273, 296, 2567, 30, 2567...
obj = self.parse_header(data)
obj = self._parse_realmedia_file(data)
def parse_objects(self, stream):
47877dfb08c6f43c07b0e458250fb48397a0d7e6 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/5551/47877dfb08c6f43c07b0e458250fb48397a0d7e6/realmedia.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1109, 67, 6911, 12, 2890, 16, 1407, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1109, 67, 6911, 12, 2890, 16, 1407, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100,...
serializer.start_keyword(self._keyword) serializer.end_keyword(self._keyword)
if self._keyword: serializer.start_keyword(self._keyword) serializer.end_keyword(self._keyword)
def serialize(self, serializer): serializer.start_keyword(self._keyword) serializer.end_keyword(self._keyword)
2aaf76c987baf2de8e2b740de02e1f6dd44f4b75 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6988/2aaf76c987baf2de8e2b740de02e1f6dd44f4b75/fixture.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4472, 12, 2890, 16, 6340, 4672, 6340, 18, 1937, 67, 11041, 12, 2890, 6315, 11041, 13, 6340, 18, 409, 67, 11041, 12, 2890, 6315, 11041, 13, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4472, 12, 2890, 16, 6340, 4672, 6340, 18, 1937, 67, 11041, 12, 2890, 6315, 11041, 13, 6340, 18, 409, 67, 11041, 12, 2890, 6315, 11041, 13, 2, -100, -100, -100, -100, -100, -100, -100, ...
types_callProxyMethod = [StringType,ListType,DictionaryType]
types_callProxyMethod = [StringType,TupleType,DictionaryType]
def initializeLcgFileCatalogProxyHandler(serviceInfo): global lfcCombinedClient lfcCombinedClient = LcgFileCatalogCombinedClient() return S_OK()
1b93f09d0e7ef9a769bbec95dee9da3c61cca28b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12864/1b93f09d0e7ef9a769bbec95dee9da3c61cca28b/LcgFileCatalogProxyHandler.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4046, 48, 26275, 812, 9769, 3886, 1503, 12, 3278, 966, 4672, 2552, 328, 7142, 27994, 1227, 328, 7142, 27994, 1227, 273, 511, 26275, 812, 9769, 27994, 1227, 1435, 327, 348, 67, 3141, 1435, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4046, 48, 26275, 812, 9769, 3886, 1503, 12, 3278, 966, 4672, 2552, 328, 7142, 27994, 1227, 328, 7142, 27994, 1227, 273, 511, 26275, 812, 9769, 27994, 1227, 1435, 327, 348, 67, 3141, 1435, ...
H = G._list_of_elements_in_H()
H = [ int(x) for x in G._list_of_elements_in_H() ]
def _coset_reduction_data_first_coord(G): """ Compute data used for determining the canonical coset representative of an element of SL_2(Z) modulo G. This function specfically returns data needed for the first part of the reduction step (the first coordinate).
eb301dff111c888467560bbc7545e7b8e22d5329 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/eb301dff111c888467560bbc7545e7b8e22d5329/congroup.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 14445, 278, 67, 1118, 4062, 67, 892, 67, 3645, 67, 5732, 12, 43, 4672, 3536, 8155, 501, 1399, 364, 23789, 326, 7378, 4987, 278, 23174, 434, 392, 930, 434, 348, 48, 67, 22, 12, 6...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 14445, 278, 67, 1118, 4062, 67, 892, 67, 3645, 67, 5732, 12, 43, 4672, 3536, 8155, 501, 1399, 364, 23789, 326, 7378, 4987, 278, 23174, 434, 392, 930, 434, 348, 48, 67, 22, 12, 6...
def finalize_product(self): "if requested, make bonds and/or join units into one part" desc = " (N = %d)" % self.ncopies self.final_msg_accum = "%s making %s%s: " % (self.msg_modename.split()[0], self.product_type, desc) msg0 = "leaving mode, finalizing product..." self.status_msg(self.final_msg_accum + msg0) print "fyi: extrude params not mentioned in statusbar: offset = %r, tol = %r" % (self.offset, self.bond_tolerance) if self.whendone_make_bonds:
def finalize_product(self, cancelling = 0): "if requested, make bonds and/or join units into one part; cancelling = 1 means just do cleanup, use diff msgs" if not cancelling: desc = " (N = %d)" % self.ncopies self.final_msg_accum = "%s making %s%s: " % (self.msg_modename.split()[0], self.product_type, desc) msg0 = "leaving mode, finalizing product..." self.status_msg(self.final_msg_accum + msg0) print "fyi: extrude params not mentioned in statusbar: offset = %r, tol = %r" % (self.offset, self.bond_tolerance) else: msg = "%s cancelled (alpha warning: might not fully restore initial state)" % (self.msg_modename.split()[0],) self.status_msg( msg) if self.whendone_make_bonds and not cancelling:
def StateDone(self): ## self.update_from_controls() #k 041017 night - will this help or hurt? since hard to know, not adding it now. # restore normal appearance for mol in self.molcopies: try: del mol._colorfunc mol.changeapp(0) except: pass self.finalize_product() # ... and emit status message about it return None
93def9cf394f749ee4f6f85c71a4645a60945878 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/93def9cf394f749ee4f6f85c71a4645a60945878/extrudeMode.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3287, 7387, 12, 2890, 4672, 7541, 365, 18, 2725, 67, 2080, 67, 24350, 1435, 468, 79, 374, 9803, 1611, 27, 290, 750, 300, 903, 333, 2809, 578, 366, 295, 88, 35, 3241, 7877, 358, 5055, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3287, 7387, 12, 2890, 4672, 7541, 365, 18, 2725, 67, 2080, 67, 24350, 1435, 468, 79, 374, 9803, 1611, 27, 290, 750, 300, 903, 333, 2809, 578, 366, 295, 88, 35, 3241, 7877, 358, 5055, ...
print "*** FINISHED ***"
print "***_FINISHED_***"
def updateMonitoringInformation( jobs ):
f28943d46a1275727d8c070b84ce3598227b3c8b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1488/f28943d46a1275727d8c070b84ce3598227b3c8b/Remote.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 18410, 5369, 12, 6550, 262, 30, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 18410, 5369, 12, 6550, 262, 30, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10...
assert descr in node.curfields
fieldnode = node.curfields[descr] resbox = op.result opt.assertval(resbox, fieldnode.source)
def getfield_gc(self, opt, op): node = opt.spec.getnode(op.args[0]) descr = op.descr assert isinstance(descr, AbstractValue) if node.virtual: assert descr in node.curfields return None return op
2a4f036953ba7f84f56cdc173003fb1c30722732 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6934/2a4f036953ba7f84f56cdc173003fb1c30722732/optimize3.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 1518, 67, 13241, 12, 2890, 16, 2153, 16, 1061, 4672, 756, 273, 2153, 18, 2793, 18, 588, 2159, 12, 556, 18, 1968, 63, 20, 5717, 18426, 273, 1061, 18, 28313, 1815, 1549, 12, 28313, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 1518, 67, 13241, 12, 2890, 16, 2153, 16, 1061, 4672, 756, 273, 2153, 18, 2793, 18, 588, 2159, 12, 556, 18, 1968, 63, 20, 5717, 18426, 273, 1061, 18, 28313, 1815, 1549, 12, 28313, ...
self._stop = False
self._stop = self._closed = False
def __init__(self, name, env=None, prefix='zodb_', config=None): """Create a new storage.
ea82bc70eae6ac2afa78935c67947cd34432ac5f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9658/ea82bc70eae6ac2afa78935c67947cd34432ac5f/BerkeleyBase.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 508, 16, 1550, 33, 7036, 16, 1633, 2218, 94, 10605, 67, 2187, 642, 33, 7036, 4672, 3536, 1684, 279, 394, 2502, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 508, 16, 1550, 33, 7036, 16, 1633, 2218, 94, 10605, 67, 2187, 642, 33, 7036, 4672, 3536, 1684, 279, 394, 2502, 18, 2, -100, -100, -100, -100, -100, -100,...
"%s cannot be child of %s" % (repr(node), repr(self) )
"%s cannot be child of %s" % (repr(node), repr(self))
def appendChild(self, node): if node.nodeType not in self.childNodeTypes: raise HierarchyRequestErr, \ "%s cannot be child of %s" % (repr(node), repr(self) ) if node.parentNode is not None: node.parentNode.removeChild(node) if self.childNodes: last = self.lastChild node.previousSibling = last last.nextSibling = node else: node.previousSibling = None node.nextSibling = None self.childNodes.append(node) if self._makeParentNodes: node.parentNode = self return node
3d9b92654f18f440a68a8e80ed41265487140c54 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/3d9b92654f18f440a68a8e80ed41265487140c54/minidom.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5833, 12, 2890, 16, 756, 4672, 309, 756, 18, 2159, 559, 486, 316, 365, 18, 3624, 907, 2016, 30, 1002, 670, 6000, 691, 2524, 16, 521, 2213, 87, 2780, 506, 1151, 434, 738, 87, 6, 738, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5833, 12, 2890, 16, 756, 4672, 309, 756, 18, 2159, 559, 486, 316, 365, 18, 3624, 907, 2016, 30, 1002, 670, 6000, 691, 2524, 16, 521, 2213, 87, 2780, 506, 1151, 434, 738, 87, 6, 738, ...
'</text:list-item>' '</text:list>')
'<text:list-item>' '<text:list>' '<text:list-item>' '<text:p>eggs</text:p>' '</text:list-item>' '</text:list>' '</text:list-item>' '</text:list>' '</text:list-item>' '</text:list>')
def test_append_sub_item(self): # XXX Fail until a true odf_list element is implemented breakfast = odf_create_list() spam = odf_create_list_item(u'spam') ham = odf_create_list_item(u'ham') eggs = odf_create_list_item(u'eggs')
c85637406aa6ba3e15ed3c5f3443dfe819814624 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10612/c85637406aa6ba3e15ed3c5f3443dfe819814624/test_content.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 6923, 67, 1717, 67, 1726, 12, 2890, 4672, 468, 11329, 8911, 3180, 279, 638, 320, 2180, 67, 1098, 930, 353, 8249, 898, 8076, 273, 320, 2180, 67, 2640, 67, 1098, 1435, 24824, 2...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 6923, 67, 1717, 67, 1726, 12, 2890, 4672, 468, 11329, 8911, 3180, 279, 638, 320, 2180, 67, 1098, 930, 353, 8249, 898, 8076, 273, 320, 2180, 67, 2640, 67, 1098, 1435, 24824, 2...
rcr t1, t1, imm
rcri t1, t1, imm
def macroop RCR_M_I
66911a1fab9900768db67610346585245a484ef4 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/6787/66911a1fab9900768db67610346585245a484ef4/rotate.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 11522, 556, 534, 5093, 67, 49, 67, 45, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 11522, 556, 534, 5093, 67, 49, 67, 45, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
self.liststore = liststore = gtk.ListStore(str, str)
self.liststore = liststore = gtk.ListStore(str)
def __init__(self, *arg): gtk.Entry.__init__(self, *arg) self.set_editable(False) self.hwindow = gtk.Window() self.hwindow.set_position(gtk.WIN_POS_MOUSE) self.hwindow.set_decorated(False) vbox = gtk.VBox() self.liststore = liststore = gtk.ListStore(str, str) self.treeview = treeview = gtk.TreeView(liststore) self.hwindow.set_default_size(200, 300)
52a18dc9abc5f0d68a536f1ffff67bac53cc3297 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2456/52a18dc9abc5f0d68a536f1ffff67bac53cc3297/config.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 380, 3175, 4672, 22718, 18, 1622, 16186, 2738, 972, 12, 2890, 16, 380, 3175, 13, 365, 18, 542, 67, 19653, 12, 8381, 13, 225, 365, 18, 76, 5668, 273, 22...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 380, 3175, 4672, 22718, 18, 1622, 16186, 2738, 972, 12, 2890, 16, 380, 3175, 13, 365, 18, 542, 67, 19653, 12, 8381, 13, 225, 365, 18, 76, 5668, 273, 22...
"""
"""% d
def c_code(self, node, name, (img2d, filtersflipped), (z, ), sub): if node.inputs[0].type.dtype != node.inputs[1].type.dtype: raise NotImplementedError() assert node.inputs[0].type.dtype == node.inputs[1].type.dtype d=locals() d.update(sub)
6ef4fe09e7b91d6db497a10ee67d728eda46182a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12438/6ef4fe09e7b91d6db497a10ee67d728eda46182a/conv.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 276, 67, 710, 12, 2890, 16, 756, 16, 508, 16, 261, 6081, 22, 72, 16, 3415, 12357, 1845, 3631, 261, 94, 16, 262, 16, 720, 4672, 309, 756, 18, 10029, 63, 20, 8009, 723, 18, 8972, 480...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 276, 67, 710, 12, 2890, 16, 756, 16, 508, 16, 261, 6081, 22, 72, 16, 3415, 12357, 1845, 3631, 261, 94, 16, 262, 16, 720, 4672, 309, 756, 18, 10029, 63, 20, 8009, 723, 18, 8972, 480...