rem stringlengths 2 226k | add stringlengths 0 227k | context stringlengths 8 228k | meta stringlengths 156 215 | input_ids list | attention_mask list | labels list |
|---|---|---|---|---|---|---|
warn("Tables rebuilt because of dependencies updated: '" \ +"', '".join(dependingTables) + "'") | if not self.quiet: warn("Tables rebuilt because of dependencies updated: '" \ +"', '".join(dependingTables) + "'") | def build(self, tables): """ Builds the given tables. | dec8906e07681ee925caef6e895971c9e8464f07 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11128/dec8906e07681ee925caef6e895971c9e8464f07/build.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1361,
12,
2890,
16,
4606,
4672,
3536,
3998,
87,
326,
864,
4606,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1361,
12,
2890,
16,
4606,
4672,
3536,
3998,
87,
326,
864,
4606,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
from Numeric import array self.atoms.SetCartesianPositions(array(atoms.get_positions())) | from Numeric import array, Float self.atoms.SetCartesianPositions(array(atoms.get_positions(), Float)) | def get_forces(self, atoms): from Numeric import array self.atoms.SetCartesianPositions(array(atoms.get_positions())) return npy.array(self.calc.GetCartesianForces()) | 00836fe7ba7dc98ee8997299b7e06ef0e82cd0bc /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5735/00836fe7ba7dc98ee8997299b7e06ef0e82cd0bc/old.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
1884,
764,
12,
2890,
16,
9006,
4672,
628,
16980,
1930,
526,
16,
5450,
365,
18,
14937,
18,
694,
13006,
13404,
11024,
12,
1126,
12,
14937,
18,
588,
67,
12388,
9334,
5450,
3719,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
1884,
764,
12,
2890,
16,
9006,
4672,
628,
16980,
1930,
526,
16,
5450,
365,
18,
14937,
18,
694,
13006,
13404,
11024,
12,
1126,
12,
14937,
18,
588,
67,
12388,
9334,
5450,
3719,
... |
return self._action(*troveSpecs, ActionClass = GroupFindAction) | return self._action(ActionClass = GroupFindAction, *troveSpecs) | def find(self, *troveSpecs): return self._action(*troveSpecs, ActionClass = GroupFindAction) | d5cd416fce497abb2065fa1182b9d65b8242d794 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8747/d5cd416fce497abb2065fa1182b9d65b8242d794/groupsetrecipe.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1104,
12,
2890,
16,
380,
88,
303,
537,
15999,
4672,
327,
365,
6315,
1128,
30857,
88,
303,
537,
15999,
16,
4382,
797,
273,
3756,
3125,
1803,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1104,
12,
2890,
16,
380,
88,
303,
537,
15999,
4672,
327,
365,
6315,
1128,
30857,
88,
303,
537,
15999,
16,
4382,
797,
273,
3756,
3125,
1803,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
'bytes'.""" | 'bytes'. 'zinfo_or_arcname' is either a ZipInfo instance or the name of the file in the archive.""" if not isinstance(zinfo_or_arcname, ZipInfo): zinfo = ZipInfo(filename=zinfo_or_arcname, date_time=time.localtime(time.time())) zinfo.compress_type = self.compression else: zinfo = zinfo_or_arcname | def writestr(self, zinfo, bytes): """Write a file into the archive. The contents is the string 'bytes'.""" self._writecheck(zinfo) zinfo.file_size = len(bytes) # Uncompressed size zinfo.CRC = binascii.crc32(bytes) # CRC-32 checksum if zinfo.compress_type == ZIP_DEFLATED: co = zlib.compressobj(zlib.Z_DEFAULT_COMPRESSION, zlib.DEFLATED, -15) bytes = co.compress(bytes) + co.flush() zinfo.compress_size = len(bytes) # Compressed size else: zinfo.compress_size = zinfo.file_size zinfo.header_offset = self.fp.tell() # Start of header bytes self.fp.write(zinfo.FileHeader()) zinfo.file_offset = self.fp.tell() # Start of file bytes self.fp.write(bytes) if zinfo.flag_bits & 0x08: # Write CRC and file sizes after the file data self.fp.write(struct.pack("<lll", zinfo.CRC, zinfo.compress_size, zinfo.file_size)) self.filelist.append(zinfo) self.NameToInfo[zinfo.filename] = zinfo | ea70053892533503a724420f7e62e3c30e16ed64 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/ea70053892533503a724420f7e62e3c30e16ed64/zipfile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7262,
313,
12,
2890,
16,
998,
1376,
16,
1731,
4672,
3536,
3067,
279,
585,
1368,
326,
5052,
18,
225,
1021,
2939,
353,
326,
533,
296,
3890,
10332,
225,
296,
94,
1376,
67,
280,
67,
11828,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7262,
313,
12,
2890,
16,
998,
1376,
16,
1731,
4672,
3536,
3067,
279,
585,
1368,
326,
5052,
18,
225,
1021,
2939,
353,
326,
533,
296,
3890,
10332,
225,
296,
94,
1376,
67,
280,
67,
11828,... |
self.error('Driver letter\n') | self.error('Driver letter required\n') | def do_run(self, options, args): windows = platform == "Windows" if options.unmount: | 541006a76a12343a29852dec14beab20966e9909 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5578/541006a76a12343a29852dec14beab20966e9909/fsmount.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
2681,
12,
2890,
16,
702,
16,
833,
4672,
225,
9965,
273,
4072,
422,
315,
10399,
6,
225,
309,
702,
18,
318,
4778,
30,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
2681,
12,
2890,
16,
702,
16,
833,
4672,
225,
9965,
273,
4072,
422,
315,
10399,
6,
225,
309,
702,
18,
318,
4778,
30,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
extra_params = warning_level | extra_params = [warning_level, "-tt"] | def do_matrix_check() : python_versions=("2.3","2.4","2.5","2.6","2.7","3.1","3.2") berkeleydb_versions=("4.1","4.2","4.3","4.4","4.5","4.6","4.7","4.8","5.0") warning_level=(["-Wdefault"], ["-Werror"])[1] import subprocess for py in python_versions : for bdb in berkeleydb_versions : print print "*** Testing bindings for Python %s and Berkeley DB %s" %(py,bdb) extra_params = warning_level # Extra flags for 3.x extra_params += [] if float(py)<=2.999 else ["-bb -tt"] # Extra flags for >=2.6 extra_params += [] if ((float(py)<=2.599) or (float(py)>=2.999)) else ["-3 -tt"] params = extra_params + ["setup.py", "-q", \ "--berkeley-db=/usr/local/BerkeleyDB."+bdb,"build", "-f"] params = ["/usr/local/bin/python"+py] + params print "EXECUTING:", " ".join(params) ret=subprocess.call(params) if ret : print print ">>> Testsuite skipped" print else : params = ["/usr/local/bin/python"+py] + extra_params + ["test.py","-p"] print "EXECUTING:", " ".join(params) subprocess.call(params) | 470ffe42527e6b62b77e9d3a4541c3702aa1a497 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1929/470ffe42527e6b62b77e9d3a4541c3702aa1a497/test-full_prerelease.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
5667,
67,
1893,
1435,
294,
5790,
67,
10169,
33,
2932,
22,
18,
23,
15937,
22,
18,
24,
15937,
22,
18,
25,
15937,
22,
18,
26,
15937,
22,
18,
27,
15937,
23,
18,
21,
15937,
23,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
5667,
67,
1893,
1435,
294,
5790,
67,
10169,
33,
2932,
22,
18,
23,
15937,
22,
18,
24,
15937,
22,
18,
25,
15937,
22,
18,
26,
15937,
22,
18,
27,
15937,
23,
18,
21,
15937,
23,... |
move_ids.append(act[1]) | if isinstance(act[1], (int, long)): move_ids.append(act[1]) else: move_ids.extend(act[1]) | def set_inventory_moves(self, cursor, user, shipment_id, name, value, arg, context=None): move_obj = self.pool.get('stock.move') | 67c254956f284f804a6e62bd66d638cb1b74207e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9298/67c254956f284f804a6e62bd66d638cb1b74207e/shipment.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
67,
26024,
67,
81,
10829,
12,
2890,
16,
3347,
16,
729,
16,
26990,
67,
350,
16,
508,
16,
460,
16,
1501,
16,
819,
33,
7036,
4672,
3635,
67,
2603,
273,
365,
18,
6011,
18,
588,
26... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
67,
26024,
67,
81,
10829,
12,
2890,
16,
3347,
16,
729,
16,
26990,
67,
350,
16,
508,
16,
460,
16,
1501,
16,
819,
33,
7036,
4672,
3635,
67,
2603,
273,
365,
18,
6011,
18,
588,
26... |
self.assertRaises(POSException.VersionError, self._storage.load, oid, 'bogus') | data, revid = self._storage.load(oid, 'bogus') assert pickle.loads(data) == 11 | def checkVersionedLoadErrors(self): oid = self._storage.new_oid() version = 'test-version' revid = self._dostore(oid, data=11) revid = self._dostore(oid, revid=revid, data=12, version=version) # Try to load a bogus oid self.assertRaises(KeyError, self._storage.load, self._storage.new_oid(), '') # Try to load a bogus version string self.assertRaises(POSException.VersionError, self._storage.load, oid, 'bogus') | f1ee732285102e57cecb53cb7466f9b967cd3af2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10048/f1ee732285102e57cecb53cb7466f9b967cd3af2/VersionStorage.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
23676,
2563,
4229,
12,
2890,
4672,
7764,
273,
365,
6315,
5697,
18,
2704,
67,
839,
1435,
1177,
273,
296,
3813,
17,
1589,
11,
283,
1246,
273,
365,
6315,
72,
669,
479,
12,
839,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
23676,
2563,
4229,
12,
2890,
4672,
7764,
273,
365,
6315,
5697,
18,
2704,
67,
839,
1435,
1177,
273,
296,
3813,
17,
1589,
11,
283,
1246,
273,
365,
6315,
72,
669,
479,
12,
839,
16,
... |
Writes a three dimensional array. The first axis is the step, i.e number of trace, the second is the pixel in that step and the third | Writes a three dimensional array. The first axis is the step, i.e number of trace, the second is the pixel in that step and the third | def __dayplotGetMinMaxValues(self, *args, **kwargs): """ Takes a Stream object and calculates the min and max values for each pixel in the dayplot. | ac0468b3b204e42963e836956a78c9c18fddeffc /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10470/ac0468b3b204e42963e836956a78c9c18fddeffc/waveform.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2881,
4032,
967,
17454,
1972,
12,
2890,
16,
380,
1968,
16,
2826,
4333,
4672,
3536,
23004,
279,
3961,
733,
471,
17264,
326,
1131,
471,
943,
924,
364,
1517,
4957,
316,
326,
2548,
403... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2881,
4032,
967,
17454,
1972,
12,
2890,
16,
380,
1968,
16,
2826,
4333,
4672,
3536,
23004,
279,
3961,
733,
471,
17264,
326,
1131,
471,
943,
924,
364,
1517,
4957,
316,
326,
2548,
403... |
>>> check(y.diagonal(0, 1)); | >>> check(y.diagonal(0, 0)); | def _numarray_tests(): ''' >>> from numpy_ext import * >>> x = new_array() >>> y = x.copy() >>> p = _printer() >>> check = p.check >>> exercise_numarray(x, p) >>> check(y.astype()); >>> check(y.argmax()); >>> check(y.argmax(0)); >>> check(y.argmin()); >>> check(y.argmin(0)); >>> check(y.argsort()); >>> check(y.argsort(1)); >>> y.byteswap(); >>> check(y); >>> check(y.diagonal()); >>> check(y.diagonal(1)); >>> check(y.diagonal(0, 1)); >>> check(y.diagonal(0, 1, 0)); >>> check(y.is_c_array()); >>> check(y.isbyteswapped()); >>> check(y.trace()); >>> check(y.trace(1)); >>> check(y.trace(0, 1)); >>> check(y.trace(0, 1, 0)); >>> check(y.new('D')); >>> y.sort(); >>> check(y); >>> check(y.type()); >>> check(y.factory((1.2, 3.4))); >>> check(y.factory((1.2, 3.4), "Double")); >>> check(y.factory((1.2, 3.4), "Double", (1,2,1))); >>> check(y.factory((1.2, 3.4), "Double", (2,1,1), false)); >>> check(y.factory((1.2, 3.4), "Double", (2,), true, true)); >>> p.results [] >>> del p ''' pass | 49d4aac8ec00a0eee4232402a97c1da5cf5560e8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9978/49d4aac8ec00a0eee4232402a97c1da5cf5560e8/numpy.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2107,
1126,
67,
16341,
13332,
9163,
4080,
628,
3972,
67,
408,
1930,
380,
4080,
619,
273,
394,
67,
1126,
1435,
4080,
677,
273,
619,
18,
3530,
1435,
4080,
293,
273,
389,
30439,
1435,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2107,
1126,
67,
16341,
13332,
9163,
4080,
628,
3972,
67,
408,
1930,
380,
4080,
619,
273,
394,
67,
1126,
1435,
4080,
677,
273,
619,
18,
3530,
1435,
4080,
293,
273,
389,
30439,
1435,
... |
method='{http://www.opengis.net/wfs}Get'): | srsname=None, method='{http://www.opengis.net/wfs}Get'): | def getfeature(self, typename=None, filter=None, bbox=None, featureid=None, featureversion=None, propertyname=['*'], maxfeatures=None, method='{http://www.opengis.net/wfs}Get'): """Request and return feature data as a file-like object. Parameters ---------- typename : list List of typenames (string) filter : string XML-encoded OGC filter expression. bbox : tuple (left, bottom, right, top) in the feature type's coordinates. featureid : list List of unique feature ids (string) featureversion : string Default is most recent feature version. propertyname : list List of feature property names. '*' matches all. maxfeatures : int Maximum number of features to be returned. method : string Qualified name of the HTTP DCP method to use. | 7620dbe8432ef7a70f2e34d4c8f3d2868486374a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11708/7620dbe8432ef7a70f2e34d4c8f3d2868486374a/wfs.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
7238,
12,
2890,
16,
26735,
33,
7036,
16,
1034,
33,
7036,
16,
8472,
33,
7036,
16,
2572,
350,
33,
7036,
16,
2572,
1589,
33,
7036,
16,
1272,
529,
33,
3292,
14,
17337,
943,
7139,
33... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
7238,
12,
2890,
16,
26735,
33,
7036,
16,
1034,
33,
7036,
16,
8472,
33,
7036,
16,
2572,
350,
33,
7036,
16,
2572,
1589,
33,
7036,
16,
1272,
529,
33,
3292,
14,
17337,
943,
7139,
33... |
if self.debug: print >>sys.stderr, " | self.debug(1, " | def run(self): self.state=self.initial[0] self.send_sock = l = conf.L3socket() while 1: try: if self.debug: print >>sys.stderr, "## state=[%s]" % self.state | e317d2d22c1e4dfdb0e7e1931d53d122598d79fb /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7311/e317d2d22c1e4dfdb0e7e1931d53d122598d79fb/scapy.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
365,
18,
2019,
33,
2890,
18,
6769,
63,
20,
65,
365,
18,
4661,
67,
15031,
273,
328,
273,
2195,
18,
48,
23,
7814,
1435,
1323,
404,
30,
775,
30,
309,
365,
18,
41... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
365,
18,
2019,
33,
2890,
18,
6769,
63,
20,
65,
365,
18,
4661,
67,
15031,
273,
328,
273,
2195,
18,
48,
23,
7814,
1435,
1323,
404,
30,
775,
30,
309,
365,
18,
41... |
f = open(os.path.join(directory, name), 'w') | f = open(os.path.join(directory, name), 'wb') | def write_images(self, directory): for (name, data) in self.IMAGES.items(): f = open(os.path.join(directory, name), 'w') f.write(base64.decodestring(data)) f.close() | 2b69e7ac45461732b1622381597218b0dd3a8d58 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/3512/2b69e7ac45461732b1622381597218b0dd3a8d58/html.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
67,
7369,
12,
2890,
16,
1867,
4672,
364,
261,
529,
16,
501,
13,
316,
365,
18,
13603,
55,
18,
3319,
13332,
284,
273,
1696,
12,
538,
18,
803,
18,
5701,
12,
5149,
16,
508,
3631,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
67,
7369,
12,
2890,
16,
1867,
4672,
364,
261,
529,
16,
501,
13,
316,
365,
18,
13603,
55,
18,
3319,
13332,
284,
273,
1696,
12,
538,
18,
803,
18,
5701,
12,
5149,
16,
508,
3631,
... |
push(name) need_blank_after_doc = 0 | push(self.bold(name)) push('\n') | def _docproperty(self, name, value, mod): results = [] push = results.append | 41b8b8a0df4f3e404b0e2be71e8c5ba7ae6ac80c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/41b8b8a0df4f3e404b0e2be71e8c5ba7ae6ac80c/pydoc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2434,
4468,
12,
2890,
16,
508,
16,
460,
16,
681,
4672,
1686,
273,
5378,
1817,
273,
1686,
18,
6923,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2434,
4468,
12,
2890,
16,
508,
16,
460,
16,
681,
4672,
1686,
273,
5378,
1817,
273,
1686,
18,
6923,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
if rox.app_dir.find(d)==0: | if app_dir.find(d)==0: | def _run_by_injector(): """Internal function.""" try: from zeroinstall.injector import basedir for d in basedir.xdg_cache_dirs: if rox.app_dir.find(d)==0: # Applicaion is in a cache dir return True elif rox._roxlib_dir.find(d)==0: # ROX-Lib is in a cache dir, we are probably being run by the # injector return True except: pass return False | b933633bc6f75861fd45b41140b0d6ccd143145c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/4591/b933633bc6f75861fd45b41140b0d6ccd143145c/mime_handler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2681,
67,
1637,
67,
28660,
13332,
3536,
3061,
445,
12123,
775,
30,
628,
24910,
885,
3429,
18,
28660,
1930,
15573,
364,
302,
316,
15573,
18,
7669,
75,
67,
2493,
67,
8291,
30,
309,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2681,
67,
1637,
67,
28660,
13332,
3536,
3061,
445,
12123,
775,
30,
628,
24910,
885,
3429,
18,
28660,
1930,
15573,
364,
302,
316,
15573,
18,
7669,
75,
67,
2493,
67,
8291,
30,
309,
... |
pass | if not (refID in self.pendingSockets): return canceltimer(self.pendingSockets[refID][2]) self.pendingSockets[refID][2] = None if frame.content == STATUS_CONFIRMED: self.pendingSockets[refID][0] = True try: self.pendingSockets[refID][1].release() except: pass | def _pending_client(self, frame, refID): pass | aa293112179760621e4e879489565b9632bd205e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7995/aa293112179760621e4e879489565b9632bd205e/Multiplexer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
9561,
67,
2625,
12,
2890,
16,
2623,
16,
1278,
734,
4672,
225,
309,
486,
261,
1734,
734,
316,
365,
18,
9561,
4534,
87,
4672,
327,
282,
3755,
12542,
12,
2890,
18,
9561,
4534,
87,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
9561,
67,
2625,
12,
2890,
16,
2623,
16,
1278,
734,
4672,
225,
309,
486,
261,
1734,
734,
316,
365,
18,
9561,
4534,
87,
4672,
327,
282,
3755,
12542,
12,
2890,
18,
9561,
4534,
87,
... |
return re.sub(r'((\\[\\abfnrtv\'"]|\\x..|\\u....)+)', | return re.sub(r'((\\[\\abfnrtv\'"]|\\[0-9]..|\\x..|\\u....)+)', | def repr_string(self, x, level): test = cram(x, self.maxstring) testrepr = repr(test) if '\\' in test and '\\' not in replace(testrepr, (r'\\', '')): # Backslashes are only literal in the string and are never # needed to make any special characters, so show a raw string. return 'r' + testrepr[0] + self.escape(test) + testrepr[0] return re.sub(r'((\\[\\abfnrtv\'"]|\\x..|\\u....)+)', r'<font color="#c040c0">\1</font>', self.escape(testrepr)) | a2ebd065552d2e5c9c6624a04e40faa9148962e9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/a2ebd065552d2e5c9c6624a04e40faa9148962e9/pydoc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8480,
67,
1080,
12,
2890,
16,
619,
16,
1801,
4672,
1842,
273,
276,
1940,
12,
92,
16,
365,
18,
1896,
1080,
13,
1842,
12715,
273,
8480,
12,
3813,
13,
309,
5282,
316,
1842,
471,
5282,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8480,
67,
1080,
12,
2890,
16,
619,
16,
1801,
4672,
1842,
273,
276,
1940,
12,
92,
16,
365,
18,
1896,
1080,
13,
1842,
12715,
273,
8480,
12,
3813,
13,
309,
5282,
316,
1842,
471,
5282,
4... |
print("\nWarning: %s already exists. Will not be " "overwritten...\n" % confpath) else: try: open(confpath, "w").write(confdata) os.chmod(confpath, 0600) except Exception, e: print("Error %s occured while trying to write configuration " "file to '%s'\n" % (e, confpath)) raise SystemExit(1) | result = raw_input("\nWarning: %s already exists. " "Overwrite? [y/N]: " % confpath) if result not in ['Y', 'y']: print("Leaving %s unchanged" % confpath) return try: open(confpath, "w").write(confdata) os.chmod(confpath, 0600) except Exception, e: print("Error %s occured while trying to write configuration " "file to '%s'\n" % (e, confpath)) raise SystemExit(1) | def create_conf(confpath, confdata): # don't overwrite existing bcfg2.conf file if os.path.exists(confpath): print("\nWarning: %s already exists. Will not be " "overwritten...\n" % confpath) else: try: open(confpath, "w").write(confdata) os.chmod(confpath, 0600) except Exception, e: print("Error %s occured while trying to write configuration " "file to '%s'\n" % (e, confpath)) raise SystemExit(1) | 4b866e4d510f83404425a5ffad3b6149627b273a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11867/4b866e4d510f83404425a5ffad3b6149627b273a/Init.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
3923,
12,
3923,
803,
16,
2195,
892,
4672,
468,
2727,
1404,
6156,
2062,
6533,
2137,
22,
18,
3923,
585,
309,
1140,
18,
803,
18,
1808,
12,
3923,
803,
4672,
563,
273,
1831,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
3923,
12,
3923,
803,
16,
2195,
892,
4672,
468,
2727,
1404,
6156,
2062,
6533,
2137,
22,
18,
3923,
585,
309,
1140,
18,
803,
18,
1808,
12,
3923,
803,
4672,
563,
273,
1831,
67,
... |
res = self._findFiles([lfn],['FileID'],connection=connection) if not res['Value']['Successful'].has_key(lfn): | res = self._findFiles( [lfn], ['FileID'], connection = connection ) if not res['Value']['Successful'].has_key( lfn ): | def setReplicaHost(self,lfns,connection=False): connection = self._getConnection(connection) """ Set replica host in the catalog """ successful = {} failed = {} for lfn,info in lfns.items(): res = self._checkInfo(info,['SE','NewSE']) if not res['OK']: failed[lfn] = res['Message'] continue newSE = info['NewSE'] se = info['SE'] res = self._findFiles([lfn],['FileID'],connection=connection) if not res['Value']['Successful'].has_key(lfn): failed[lfn] = res['Value']['Failed'][lfn] continue fileID = res['Value']['Successful'][lfn]['FileID'] res = self._setReplicaHost(fileID,se,newSE,connection=connection) if res['OK']: successful[lfn] = res['Value'] else: failed[lfn] = res['Message'] return S_OK({'Successful':successful,'Failed':failed}) | 26c5008a7adbc1c761e2409bed59b7710e5f2c74 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/26c5008a7adbc1c761e2409bed59b7710e5f2c74/FileManagerBase.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
14222,
2594,
12,
2890,
16,
20850,
2387,
16,
4071,
33,
8381,
4672,
1459,
273,
365,
6315,
588,
1952,
12,
4071,
13,
3536,
1000,
12335,
1479,
316,
326,
6222,
3536,
6873,
273,
2618,
2535... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
14222,
2594,
12,
2890,
16,
20850,
2387,
16,
4071,
33,
8381,
4672,
1459,
273,
365,
6315,
588,
1952,
12,
4071,
13,
3536,
1000,
12335,
1479,
316,
326,
6222,
3536,
6873,
273,
2618,
2535... |
first_index = middle + 1 | first_index = middle | def _search_ts (self, target_ts, first_index, last_index): | 4ecbff9abd7e2f864d67e0b05021f56f111de597 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12310/4ecbff9abd7e2f864d67e0b05021f56f111de597/Timeline.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
3072,
67,
3428,
261,
2890,
16,
1018,
67,
3428,
16,
1122,
67,
1615,
16,
1142,
67,
1615,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
3072,
67,
3428,
261,
2890,
16,
1018,
67,
3428,
16,
1122,
67,
1615,
16,
1142,
67,
1615,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
FoundationPlist.writePlist(installinfo, | FoundationPlist.writePlist(installinfo, | def check(id=''): '''Checks for available new or updated managed software, downloading installer items if needed. Returns 1 if there are available updates, 0 if there are no available updates, and -1 if there were errors.''' getMachineFacts() munkicommon.report['MachineInfo'] = machine ManagedInstallDir = munkicommon.pref('ManagedInstallDir') if munkicommon.munkistatusoutput: munkistatus.activate() munkistatus.message("Checking for available updates...") munkistatus.detail("") munkistatus.percent("-1") munkicommon.log("### Beginning managed software check ###") mainmanifestpath = getPrimaryManifest(id) if munkicommon.stopRequested(): return 0 installinfo = {} if mainmanifestpath: # initialize our installinfo record installinfo['managed_installs'] = [] installinfo['removals'] = [] munkicommon.display_detail("**Checking for installs**") installinfo = processManifestForInstalls(mainmanifestpath, installinfo) if munkicommon.stopRequested(): return 0 if munkicommon.munkistatusoutput: # reset progress indicator and detail field munkistatus.message("Checking for additional changes...") munkistatus.percent("-1") munkistatus.detail('') # now generate a list of items to be uninstalled munkicommon.display_detail("**Checking for removals**") if munkicommon.stopRequested(): return 0 installinfo = processManifestForRemovals(mainmanifestpath, installinfo) # filter managed_installs to get items already installed installed_items = [item for item in installinfo.get('managed_installs',[]) if item.get('installed')] # filter managed_installs to get problem items: # not installed, but no installer item problem_items = [item for item in installinfo.get('managed_installs',[]) if item.get('installed') == False and not item.get('installer_item')] # filter removals to get items already removed (or never installed) removed_items = [item for item in installinfo.get('removals',[]) if item.get('installed') == False] # filter managed_installs and removals lists # so they have only items that need action installinfo['managed_installs'] = \ [item for item in installinfo.get('managed_installs',[]) if item.get('installer_item')] installinfo['removals'] = \ [item for item in installinfo.get('removals',[]) if item.get('installed')] munkicommon.report['ManagedInstalls'] = installed_items munkicommon.report['ProblemInstalls'] = problem_items munkicommon.report['RemovedItems'] = removed_items munkicommon.report['ItemsToInstall'] = installinfo['managed_installs'] munkicommon.report['ItemsToRemove'] = installinfo['removals'] # clean up cache dir # remove any item in the cache that isn't scheduled # to be used for an install or removal # this could happen if an item is downloaded on one # updatecheck run, but later removed from the manifest # before it is installed or removed - so the cached item # is no longer needed. cache_list = [item["installer_item"] for item in installinfo.get('managed_installs',[])] cache_list.extend([item["uninstaller_item"] for item in installinfo.get('removals',[]) if item.get('uninstaller_item')]) cachedir = os.path.join(ManagedInstallDir, "Cache") for item in os.listdir(cachedir): if item.endswith(".download"): # we have a partial download here # remove the ".download" from the end of the filename fullitem = os.path.splitext(item)[0] if os.path.exists(os.path.join(cachedir, fullitem)): # we have a partial and a full download # for the same item. (This shouldn't happen.) # remove the partial download. os.unlink(os.path.join(cachedir, item)) elif problem_items == []: # problem items is our list of items # that need to be installed but are missing # the installer_item; these might be partial # downloads. So if we have no problem items, it's # OK to get rid of any partial downloads hanging around. os.unlink(os.path.join(cachedir, item)) elif item not in cache_list: munkicommon.display_detail("Removing %s from cache" % item) os.unlink(os.path.join(cachedir, item)) # write out install list so our installer # can use it to install things in the right order installinfochanged = True installinfopath = os.path.join(ManagedInstallDir, "InstallInfo.plist") if os.path.exists(installinfopath): oldinstallinfo = FoundationPlist.readPlist(installinfopath) if oldinstallinfo == installinfo: installinfochanged = False munkicommon.display_detail("No change in InstallInfo.") if installinfochanged: FoundationPlist.writePlist(installinfo, os.path.join(ManagedInstallDir, "InstallInfo.plist")) else: # couldn't get a primary manifest. Check to see if we have a valid # install/remove list from an earlier run. munkicommon.display_error( "Could not retrieve managed install primary manifest.") installinfopath = os.path.join(ManagedInstallDir, "InstallInfo.plist") if os.path.exists(installinfopath): try: installinfo = FoundationPlist.readPlist(installinfopath) except FoundationPlist.NSPropertyListSerializationException: installinfo = {} else: munkicommon.report['ItemsToInstall'] = \ installinfo.get('managed_installs',[]) munkicommon.report['ItemsToRemove'] = \ installinfo.get('removals',[]) installcount = len(installinfo.get("managed_installs",[])) removalcount = len(installinfo.get("removals",[])) munkicommon.log("") if installcount: munkicommon.display_info( "The following items will be installed or upgraded:") for item in installinfo.get('managed_installs',[]): if item.get('installer_item'): munkicommon.display_info(" + %s-%s" % (item.get('name',''), item.get('version_to_install',''))) if item.get('description'): munkicommon.display_info(" %s" % item['description']) if item.get('RestartAction') == 'RequireRestart' or \ item.get('RestartAction') == 'RecommendRestart': munkicommon.display_info(" *Restart required") munkicommon.report['RestartRequired'] = True if item.get('RestartAction') == 'RequireLogout': munkicommon.display_info(" *Logout required") munkicommon.report['LogoutRequired'] = True if removalcount: munkicommon.display_info("The following items will be removed:") for item in installinfo.get('removals',[]): if item.get('installed'): munkicommon.display_info(" - %s" % item.get('name')) if item.get('RestartAction') == 'RequireRestart' or \ item.get('RestartAction') == 'RecommendRestart': munkicommon.display_info(" *Restart required") munkicommon.report['RestartRequired'] = True if item.get('RestartAction') == 'RequireLogout': munkicommon.display_info(" *Logout required") munkicommon.report['LogoutRequired'] = True if installcount == 0 and removalcount == 0: munkicommon.display_info( "No changes to managed software are available.") munkicommon.savereport() munkicommon.log("### End managed software check ###") if installcount or removalcount: return 1 else: return 0 | 5350042f4ddfb098468c8d75b15dfa28757b9321 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6241/5350042f4ddfb098468c8d75b15dfa28757b9321/updatecheck.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
12,
350,
2218,
11,
4672,
9163,
4081,
364,
2319,
394,
578,
3526,
7016,
17888,
16,
23742,
16050,
1516,
309,
3577,
18,
2860,
404,
309,
1915,
854,
2319,
4533,
16,
374,
309,
1915,
854,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
12,
350,
2218,
11,
4672,
9163,
4081,
364,
2319,
394,
578,
3526,
7016,
17888,
16,
23742,
16050,
1516,
309,
3577,
18,
2860,
404,
309,
1915,
854,
2319,
4533,
16,
374,
309,
1915,
854,
... |
speedrace(hdlr, pct, pct+pct_step, circs_per_node, save_every, out_dir) | speedrace(hdlr, pct, pct+pct_step, circs_per_node, save_every, out_dir, max_fetch_time) | def main(argv): TorUtil.read_config(argv[1]) (start_pct,stop_pct,nodes_per_slice,save_every, circs_per_node,out_dir) = read_config(argv[1]) try: (c,hdlr) = setup_handler() except Exception, e: plog("WARN", "Can't connect to Tor: "+str(e)) sql_file = os.getcwd()+'/'+out_dir+'/bwauthority.sqlite' hdlr.attach_sql_listener('sqlite:///'+sql_file) # set SOCKS proxy socks.setdefaultproxy(socks.PROXY_TYPE_SOCKS5, tor_host, tor_port) socket.socket = socks.socksocket while True: pct = start_pct plog('INFO', 'Beginning time loop') while pct < stop_pct: pct_step = hdlr.rank_to_percent(nodes_per_slice) hdlr.reset_stats() hdlr.commit() plog('DEBUG', 'Reset stats') speedrace(hdlr, pct, pct+pct_step, circs_per_node, save_every, out_dir) plog('DEBUG', 'speedroced') hdlr.commit() hdlr.close_circuits() lo = str(round(pct,1)) hi = str(round(pct+pct_step,1)) hdlr.write_sql_stats(pct, pct+pct_step, os.getcwd()+'/'+out_dir+'/sql-'+lo+':'+hi+"-done-"+strftime("20%y-%m-%d-%H:%M:%S")) hdlr.write_strm_bws(pct, pct+pct_step, os.getcwd()+'/'+out_dir+'/bws-'+lo+':'+hi+"-done-"+strftime("20%y-%m-%d-%H:%M:%S")) plog('DEBUG', 'Wrote stats') pct += pct_step hdlr.save_sql_file(sql_file, "db-"+str(lo)+":"+str(hi)+"-"+strftime("20%y-%m-%d-%H:%M:%S")+".sqlite") | e8450ad1cd3c23315f682731ad3047861f509350 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3762/e8450ad1cd3c23315f682731ad3047861f509350/bwauthority.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
19485,
4672,
23763,
1304,
18,
896,
67,
1425,
12,
19485,
63,
21,
5717,
261,
1937,
67,
23989,
16,
5681,
67,
23989,
16,
4690,
67,
457,
67,
6665,
16,
5688,
67,
20434,
16,
5886,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
19485,
4672,
23763,
1304,
18,
896,
67,
1425,
12,
19485,
63,
21,
5717,
261,
1937,
67,
23989,
16,
5681,
67,
23989,
16,
4690,
67,
457,
67,
6665,
16,
5688,
67,
20434,
16,
5886,
... |
def base_ring(self): """ Return the underlying base ring. """ return self._base_ring | def __repr__(self): return "SpecialCubicQuotientRing over %s with polynomial T = %s" % \ (self._base_ring, PolynomialRing(self._base_ring)( [self._b, self._a, 0, 1])) | c26d6b07795f11a9104ed5b11740f789a4ab67f6 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9890/c26d6b07795f11a9104ed5b11740f789a4ab67f6/monsky_washnitzer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
12715,
972,
12,
2890,
4672,
327,
315,
12193,
39,
17616,
7678,
1979,
10369,
1879,
738,
87,
598,
16991,
399,
273,
738,
87,
6,
738,
521,
261,
2890,
6315,
1969,
67,
8022,
16,
6730,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
12715,
972,
12,
2890,
4672,
327,
315,
12193,
39,
17616,
7678,
1979,
10369,
1879,
738,
87,
598,
16991,
399,
273,
738,
87,
6,
738,
521,
261,
2890,
6315,
1969,
67,
8022,
16,
6730,
1... | |
self["medium_label"].setText(_("exceeds dual layer medium!")) | self["medium_label"].setText(_("Exceeds dual layer medium!")) | def updateSize(self): size = self.project.size/(1024*1024) MAX_DL = self.project.MAX_DL-100 MAX_SL = self.project.MAX_SL-100 print "updateSize:", size, "MAX_DL:", MAX_DL, "MAX_SL:", MAX_SL if size > MAX_DL: percent = 100 * size / float(MAX_DL) self["space_label_dual"].text = "%d MB (%.2f%%)" % (size, percent) self["space_bar_dual"].value = int(percent) self["space_bar_single"].value = 100 self["space_label_single"].text = "" self["medium_label"].setText(_("exceeds dual layer medium!")) self["medium_label"].setForegroundColorNum(2) if self.previous_size < MAX_DL: self.session.open(MessageBox,text = _("exceeds dual layer medium!"), type = MessageBox.TYPE_ERROR) elif size > MAX_SL: percent = 100 * size / float(MAX_DL) self["space_label_dual"].text = "%d MB (%.2f%%)" % (size, percent) self["space_bar_dual"].value = int(percent) self["space_bar_single"].value = 100 self["space_label_single"].text = "" self["medium_label"].setText(_("required medium type:") + " " + _("DUAL LAYER DVD") + ", %d MB " % (MAX_DL - size) + _("free")) self["medium_label"].setForegroundColorNum(1) if self.previous_size < MAX_SL: self.session.open(MessageBox, text = _("Your collection exceeds the size of a single layer medium, you will need a blank dual layer DVD!"), timeout = 10, type = MessageBox.TYPE_INFO) elif size < MAX_SL: percent = 100 * size / float(MAX_SL) self["space_label_single"].text = "%d MB (%.2f%%)" % (size, percent) self["space_bar_single"].value = int(percent) self["space_bar_dual"].value = 0 self["space_label_dual"].text = "" self["medium_label"].setText(_("required medium type:") + " " + _("SINGLE LAYER DVD") + ", %d MB " % (MAX_SL - size) + _("free")) self["medium_label"].setForegroundColorNum(0) self.previous_size = size | 16e2662171701ce64bf2e789168e8cee54dc8588 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6652/16e2662171701ce64bf2e789168e8cee54dc8588/TitleList.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
1225,
12,
2890,
4672,
963,
273,
365,
18,
4406,
18,
1467,
19,
12,
2163,
3247,
14,
2163,
3247,
13,
4552,
67,
8914,
273,
365,
18,
4406,
18,
6694,
67,
8914,
17,
6625,
4552,
67,
455... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
1225,
12,
2890,
4672,
963,
273,
365,
18,
4406,
18,
1467,
19,
12,
2163,
3247,
14,
2163,
3247,
13,
4552,
67,
8914,
273,
365,
18,
4406,
18,
6694,
67,
8914,
17,
6625,
4552,
67,
455... |
license = 'BSD', | license = 'LGPL', | def finalize_options(self): build_ext.finalize_options(self) self.include_dirs.insert(0, 'include') self.pyrex_include_dirs.extend(self.include_dirs) | fb7c0814001a30edd3c24892c3fd6eaf7520360a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12343/fb7c0814001a30edd3c24892c3fd6eaf7520360a/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
12409,
67,
2116,
12,
2890,
4672,
1361,
67,
408,
18,
30343,
67,
2116,
12,
2890,
13,
365,
18,
6702,
67,
8291,
18,
6387,
12,
20,
16,
296,
6702,
6134,
365,
18,
2074,
266,
92,
67,
6702,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
12409,
67,
2116,
12,
2890,
4672,
1361,
67,
408,
18,
30343,
67,
2116,
12,
2890,
13,
365,
18,
6702,
67,
8291,
18,
6387,
12,
20,
16,
296,
6702,
6134,
365,
18,
2074,
266,
92,
67,
6702,
... |
def getHash(self, path=None): """Get the hash of a file. | def getHash(self, path): """Get the hash of a file stored in ``path``. | def getHash(self, path=None): """Get the hash of a file. | fb0d438c88490272f659933b5d44bb0582381fa2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9162/fb0d438c88490272f659933b5d44bb0582381fa2/cachemanager.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
16075,
12,
2890,
16,
589,
33,
7036,
4672,
3536,
967,
326,
1651,
434,
279,
585,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
16075,
12,
2890,
16,
589,
33,
7036,
4672,
3536,
967,
326,
1651,
434,
279,
585,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
s.bind('', self.port) | s.bind(('', self.port)) | def __init__(self, port=None, connection_hook=None): self.connections = [] self.port = port or self.default_port self.connection_hook = connection_hook | 9a580c440c4f064a92bbc537ac3a2df1c0998afc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/9a580c440c4f064a92bbc537ac3a2df1c0998afc/protocol.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1756,
33,
7036,
16,
1459,
67,
4476,
33,
7036,
4672,
365,
18,
13313,
273,
5378,
365,
18,
655,
273,
1756,
578,
365,
18,
1886,
67,
655,
365,
18,
4071,
67,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1756,
33,
7036,
16,
1459,
67,
4476,
33,
7036,
4672,
365,
18,
13313,
273,
5378,
365,
18,
655,
273,
1756,
578,
365,
18,
1886,
67,
655,
365,
18,
4071,
67,... |
return (invalid[index], "kNoError", "GL_INVALID_ENUM") return ("---ERROR1---", "kNoError", "GL_INVALID_ENUM") | return (invalid[index], "kNoError", self.gl_error) return ("---ERROR1---", "kNoError", self.gl_error) | def GetInvalidArg(self, offset, index): """returns an invalid value by index.""" if 'invalid' in self.enum_info: invalid = self.enum_info['invalid'] num_invalid = len(invalid) if index >= num_invalid: index = num_invalid - 1 return (invalid[index], "kNoError", "GL_INVALID_ENUM") return ("---ERROR1---", "kNoError", "GL_INVALID_ENUM") | 51330b3a08d845da9815b843e020833e62577828 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9392/51330b3a08d845da9815b843e020833e62577828/build_gles2_cmd_buffer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
1941,
4117,
12,
2890,
16,
1384,
16,
770,
4672,
3536,
6154,
392,
2057,
460,
635,
770,
12123,
309,
296,
5387,
11,
316,
365,
18,
7924,
67,
1376,
30,
2057,
273,
365,
18,
7924,
67,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
1941,
4117,
12,
2890,
16,
1384,
16,
770,
4672,
3536,
6154,
392,
2057,
460,
635,
770,
12123,
309,
296,
5387,
11,
316,
365,
18,
7924,
67,
1376,
30,
2057,
273,
365,
18,
7924,
67,
1... |
M_PROPPATCH = 9 | M_PROPPATCH = 9 | def init(): """ This function is called by the server at startup time """ return CallBack() | 3de9a0d15ba9eb7275476985ecfc3381876d7703 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10002/3de9a0d15ba9eb7275476985ecfc3381876d7703/apache.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1208,
13332,
3536,
1220,
445,
353,
2566,
635,
326,
1438,
622,
11850,
813,
3536,
225,
327,
3049,
2711,
1435,
225,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1208,
13332,
3536,
1220,
445,
353,
2566,
635,
326,
1438,
622,
11850,
813,
3536,
225,
327,
3049,
2711,
1435,
225,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
site = server.Site(root) | rewriter = rewrite.RewriterResource(root, helpimagesrewrite) site = server.Site(rewriter) | def main(): # the start and stop stuff will be handled from the freevo script logfile = '%s/webserver-%s.log' % (config.LOGDIR, os.getuid()) log.startLogging(open(logfile, 'a')) if os.path.isdir(os.path.join(os.environ['FREEVO_PYTHON'], 'www/htdocs')): docRoot = os.path.join(os.environ['FREEVO_PYTHON'], 'www/htdocs') else: docRoot = os.path.join(config.SHARE_DIR, 'htdocs') root = static.File(docRoot) root.processors = { '.rpy': script.ResourceScript, } root.putChild('vhost', vhost.VHostMonsterResource()) site = server.Site(root) application = app.Application('web') application.listenTCP(config.WWW_PORT, site) application.run(save=0) | f520ecdadf2f70a3822e5eb9fee23ac465913844 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11399/f520ecdadf2f70a3822e5eb9fee23ac465913844/webserver.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
468,
326,
787,
471,
2132,
10769,
903,
506,
7681,
628,
326,
4843,
12307,
2728,
225,
15204,
273,
1995,
87,
19,
1814,
2266,
502,
6456,
87,
18,
1330,
11,
738,
261,
1425,
18,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
468,
326,
787,
471,
2132,
10769,
903,
506,
7681,
628,
326,
4843,
12307,
2728,
225,
15204,
273,
1995,
87,
19,
1814,
2266,
502,
6456,
87,
18,
1330,
11,
738,
261,
1425,
18,
4... |
dir = os.path.expanduser("~/.Trash") for file in FileUtilities.children_in_directory(dir, False): yield file | dirname = os.path.expanduser("~/.Trash") for filename in FileUtilities.children_in_directory(dirname, False): yield filename | def list_files(self): dir = os.path.expanduser("~/.Trash") for file in FileUtilities.children_in_directory(dir, False): yield file # fixme http://www.ramendik.ru/docs/trashspec.html # http://standards.freedesktop.org/basedir-spec/basedir-spec-0.6.html # GNOME 2.22, Fedora 9 ~/.local/share/Trash dir = os.path.expanduser("~/.local/share/Trash") for file in FileUtilities.children_in_directory(dir, False): yield file | 737a941d1a6026b292895c42a857f9ec0bf05cea /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7853/737a941d1a6026b292895c42a857f9ec0bf05cea/CleanerBackend.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
666,
67,
2354,
12,
2890,
4672,
4283,
273,
1140,
18,
803,
18,
12320,
1355,
2932,
98,
15865,
29420,
7923,
364,
1544,
316,
1387,
11864,
18,
5906,
67,
267,
67,
5149,
12,
12287,
16,
1083,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
666,
67,
2354,
12,
2890,
4672,
4283,
273,
1140,
18,
803,
18,
12320,
1355,
2932,
98,
15865,
29420,
7923,
364,
1544,
316,
1387,
11864,
18,
5906,
67,
267,
67,
5149,
12,
12287,
16,
1083,
4... |
self.string = list(string) | self.index = 0 self.string = string | def __init__(self, string): | d46b2a42eda94cbac3b610e36d5f97230d629068 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/d46b2a42eda94cbac3b610e36d5f97230d629068/sre_parse.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
533,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
533,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
index = self.bisect(self, item) | index = bisect.bisect(self.list, item) | def __contains__(self, item): if self.__len__() == 0: return False index = self.bisect(self, item) if index == 0: return False if item in self.__getitem__(index-1): return True return False | ceb5841528feafa5a760412ecaeedf36fbba1e7b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/13031/ceb5841528feafa5a760412ecaeedf36fbba1e7b/mirrorlist_server.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
12298,
972,
12,
2890,
16,
761,
4672,
309,
365,
16186,
1897,
972,
1435,
422,
374,
30,
327,
1083,
770,
273,
27031,
18,
70,
291,
386,
12,
2890,
18,
1098,
16,
761,
13,
309,
770,
42... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
12298,
972,
12,
2890,
16,
761,
4672,
309,
365,
16186,
1897,
972,
1435,
422,
374,
30,
327,
1083,
770,
273,
27031,
18,
70,
291,
386,
12,
2890,
18,
1098,
16,
761,
13,
309,
770,
42... |
dev=self.selectedDevice(c) dev.onlyChannel=channel | dev = self.selectedDevice(c) dev.onlyChannel = channel | def regulate(self, c, channel, temperature, loadresistor=30000): """Initializes temperature regulation | ca09465af7cd0d53cd2656a51c12b3b3003818f7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7016/ca09465af7cd0d53cd2656a51c12b3b3003818f7/lakeshore370.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
960,
6243,
12,
2890,
16,
276,
16,
1904,
16,
12843,
16,
1262,
455,
376,
280,
33,
23,
2787,
4672,
3536,
9685,
12843,
960,
6234,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
960,
6243,
12,
2890,
16,
276,
16,
1904,
16,
12843,
16,
1262,
455,
376,
280,
33,
23,
2787,
4672,
3536,
9685,
12843,
960,
6234,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
raise ScriptError("Patch %s from bug %s failed to download and apply." % (patch['url'], patch['bug_id'])) | raise ScriptError(message="Patch %s from bug %s failed to download and apply." % (patch['url'], patch['bug_id'])) | def apply_patch(self, patch, force=False): # It's possible that the patch was not made from the root directory. # We should detect and handle that case. curl_process = subprocess.Popen(['curl', '--location', '--silent', '--show-error', patch['url']], stdout=subprocess.PIPE) args = [self.script_path('svn-apply'), '--reviewer', patch['reviewer']] if force: args.append('--force') patch_apply_process = subprocess.Popen(args, stdin=curl_process.stdout) | cbfef173775a18c163a046efa120abeeff3a6751 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9392/cbfef173775a18c163a046efa120abeeff3a6751/scm.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2230,
67,
2272,
12,
2890,
16,
4729,
16,
2944,
33,
8381,
4672,
468,
2597,
1807,
3323,
716,
326,
4729,
1703,
486,
7165,
628,
326,
1365,
1867,
18,
468,
1660,
1410,
5966,
471,
1640,
716,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2230,
67,
2272,
12,
2890,
16,
4729,
16,
2944,
33,
8381,
4672,
468,
2597,
1807,
3323,
716,
326,
4729,
1703,
486,
7165,
628,
326,
1365,
1867,
18,
468,
1660,
1410,
5966,
471,
1640,
716,
6... |
return chain(self.status()) | for msg in chain(self.status()): yield msg | def stop(self): if self._status == 'stopped': raise DiscoError("%s already stopped" % self) try: os.kill(self.pid, signal.SIGTERM) while self._status == 'running': pass except Exception: pass return chain(self.status()) | ad8c242af148154828b966902f5ab8e49963a98c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/373/ad8c242af148154828b966902f5ab8e49963a98c/disco.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2132,
12,
2890,
4672,
309,
365,
6315,
2327,
422,
296,
24228,
4278,
1002,
3035,
2894,
668,
27188,
87,
1818,
9627,
6,
738,
365,
13,
775,
30,
1140,
18,
16418,
12,
2890,
18,
6610,
16,
4277... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2132,
12,
2890,
4672,
309,
365,
6315,
2327,
422,
296,
24228,
4278,
1002,
3035,
2894,
668,
27188,
87,
1818,
9627,
6,
738,
365,
13,
775,
30,
1140,
18,
16418,
12,
2890,
18,
6610,
16,
4277... |
selected_tabs.add(row[0]) | selected_tabs.add((1, row[0])) | def get_selection_and_children(self): """This returns the selection and, in the case of parent rows, returns all children, too. This is particularly useful for getting selections that include children of folders. | 1f498bda8c507f0b63f8859884457bea2e925013 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12354/1f498bda8c507f0b63f8859884457bea2e925013/tablistmanager.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
10705,
67,
464,
67,
5906,
12,
2890,
4672,
3536,
2503,
1135,
326,
4421,
471,
16,
316,
326,
648,
434,
982,
2595,
16,
1135,
777,
2325,
16,
4885,
18,
225,
1220,
353,
6826,
715,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
10705,
67,
464,
67,
5906,
12,
2890,
4672,
3536,
2503,
1135,
326,
4421,
471,
16,
316,
326,
648,
434,
982,
2595,
16,
1135,
777,
2325,
16,
4885,
18,
225,
1220,
353,
6826,
715,
... |
print " abort!" return | raise RuntimeError, "certain not superset of knownKeep" | def split(self, childSize, otherMarkerId, turn=-1): print "split for", self, childSize, otherMarkerId, turn | d3360650260167e99d819d46c099993d6e0f1f9e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3755/d3360650260167e99d819d46c099993d6e0f1f9e/predictsplits.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1416,
12,
2890,
16,
1151,
1225,
16,
1308,
7078,
548,
16,
7005,
29711,
21,
4672,
1172,
315,
4939,
364,
3113,
365,
16,
1151,
1225,
16,
1308,
7078,
548,
16,
7005,
2,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1416,
12,
2890,
16,
1151,
1225,
16,
1308,
7078,
548,
16,
7005,
29711,
21,
4672,
1172,
315,
4939,
364,
3113,
365,
16,
1151,
1225,
16,
1308,
7078,
548,
16,
7005,
2,
-100,
-100,
-100,
-10... |
'Added interdiff support') | 'Improved login form') | def testDashboard5(self): """Testing dashboard view (mine)""" self.client.login(username='doc', password='doc') | d72238b7393eae1f5069fef4eba2d906a7ba3bdf /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1600/d72238b7393eae1f5069fef4eba2d906a7ba3bdf/tests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
14830,
25,
12,
2890,
4672,
3536,
22218,
11825,
1476,
261,
3081,
15574,
365,
18,
2625,
18,
5819,
12,
5053,
2218,
2434,
2187,
2201,
2218,
2434,
6134,
2,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
14830,
25,
12,
2890,
4672,
3536,
22218,
11825,
1476,
261,
3081,
15574,
365,
18,
2625,
18,
5819,
12,
5053,
2218,
2434,
2187,
2201,
2218,
2434,
6134,
2,
-100,
-100,
-100,
-100,
-100,
... |
deps = create_deps(ext_modules) cython(deps, ext_modules) | hi, med, lo = compile_command_list(ext_modules) execute_list_of_commands(hi + med + lo) | def create_deps(ext_modules): # first we compute the complete graph of dependencies deps = {} for m in ext_modules: m.extra_compile_args += extra_compile_args for i in range(len(m.sources)): f = m.sources[i] if f[-4:] == '.pyx': visited = set() deps_graph(deps, f, visited) base = os.path.splitext(f)[0] f = f[:-4] + '.pxd' if os.path.exists(f): deps_graph(deps, f, visited) return deps | 12aff705aa1e2b2dc6e401b763fe2e9e51bdf02d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/12aff705aa1e2b2dc6e401b763fe2e9e51bdf02d/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
14877,
12,
408,
67,
6400,
4672,
468,
1122,
732,
3671,
326,
3912,
2667,
434,
5030,
8740,
273,
2618,
364,
312,
316,
1110,
67,
6400,
30,
312,
18,
7763,
67,
11100,
67,
1968,
1011,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
14877,
12,
408,
67,
6400,
4672,
468,
1122,
732,
3671,
326,
3912,
2667,
434,
5030,
8740,
273,
2618,
364,
312,
316,
1110,
67,
6400,
30,
312,
18,
7763,
67,
11100,
67,
1968,
1011,... |
print "Skipping" | print "Skipping exercise(): input file not available." | def exercise () : from iotbx import file_reader pdb_file = libtbx.env.find_in_repositories( relative_path="phenix_regression/pdb/1ywf.pdb", test=os.path.isfile) pdb_file_h = libtbx.env.find_in_repositories( relative_path="phenix_regression/pdb/1ywf_h.pdb", test=os.path.isfile) if pdb_file is None : print "Skipping" return False log = cStringIO.StringIO() pdb_in = file_reader.any_file(pdb_file_h, force_type="pdb").file_object pdb_hierarchy = pdb_in.construct_hierarchy() xray_structure = pdb_in.xray_structure_simple() sec_str_from_pdb_file = pdb_in.extract_secondary_structure() m = manager(pdb_hierarchy=pdb_hierarchy, xray_structure=xray_structure, sec_str_from_pdb_file=sec_str_from_pdb_file) m.find_automatically(log=log) bonds_table = m.get_bonds_table(log=log) assert bonds_table.bonds.size() == 109 m.params.h_bond_restraints.substitute_n_for_h = True bonds_table = m.get_bonds_table(log=log) assert bonds_table.flag_use_bond.count(True) == 106 (frac_alpha, frac_beta) = m.calculate_structure_content() assert ("%.3f" % frac_alpha) == "0.643" assert ("%.3f" % frac_beta) == "0.075" del m # using KSDSSP try : m = manager(pdb_hierarchy=pdb_hierarchy, xray_structure=xray_structure, sec_str_from_pdb_file=None) m.find_automatically(log=log) bonds_table = m.get_bonds_table(log=log) assert bonds_table.bonds.size() == 93 m.params.h_bond_restraints.substitute_n_for_h = True bonds_table = m.get_bonds_table(log=log) assert bonds_table.flag_use_bond.count(True) == 86 (frac_alpha, frac_beta) = m.calculate_structure_content() assert ("%.3f" % frac_alpha) == "0.552" assert ("%.3f" % frac_beta) == "0.066" del m del pdb_hierarchy del xray_structure except RuntimeError : print "skipping KSDSSP test" # without hydrogens pdb_in = file_reader.any_file(pdb_file, force_type="pdb").file_object pdb_hierarchy = pdb_in.construct_hierarchy() xray_structure = pdb_in.xray_structure_simple() sec_str_from_pdb_file = pdb_in.extract_secondary_structure() m = manager(pdb_hierarchy=pdb_hierarchy, xray_structure=xray_structure, sec_str_from_pdb_file=sec_str_from_pdb_file) m.find_automatically(log=log) bonds_table = m.get_bonds_table(log=log) assert bonds_table.bonds.size() == 109 del m # using KSDSSP try : m = manager(pdb_hierarchy=pdb_hierarchy, xray_structure=xray_structure, sec_str_from_pdb_file=None) m.find_automatically(log=log) bonds_table = m.get_bonds_table(log=log) assert bonds_table.bonds.size() == 93 except RuntimeError : print "skipping KSDSSP test" print "OK" | 0bc14071d9364def7d33a409fac526a38f23f248 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/696/0bc14071d9364def7d33a409fac526a38f23f248/secondary_structure.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
24165,
1832,
294,
628,
30956,
70,
92,
1930,
585,
67,
10530,
10892,
67,
768,
273,
2561,
18587,
92,
18,
3074,
18,
4720,
67,
267,
67,
26626,
12,
3632,
67,
803,
1546,
13819,
697,
67,
1574,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
24165,
1832,
294,
628,
30956,
70,
92,
1930,
585,
67,
10530,
10892,
67,
768,
273,
2561,
18587,
92,
18,
3074,
18,
4720,
67,
267,
67,
26626,
12,
3632,
67,
803,
1546,
13819,
697,
67,
1574,... |
d = FSPageTemplate.__dict__ d['source.xml'] = d['source.html'] = Src() | source_dot_xml = Src() setattr(FSPageTemplate, 'source.xml', FSPageTemplate.source_dot_xml) setattr(FSPageTemplate, 'source.html', FSPageTemplate.source_dot_xml) | def _exec(self, bound_names, args, kw): """Call a FSPageTemplate""" try: response = self.REQUEST.RESPONSE except AttributeError: response = None # Read file first to get a correct content_type default value. self._updateFromFS() | 6aed3d2ffbfbdc919c13c47bad2bba3569ff3d5c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1843/6aed3d2ffbfbdc919c13c47bad2bba3569ff3d5c/FSPageTemplate.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4177,
12,
2890,
16,
2489,
67,
1973,
16,
833,
16,
5323,
4672,
3536,
1477,
279,
9247,
1964,
2283,
8395,
775,
30,
766,
273,
365,
18,
5519,
18,
14508,
1335,
6394,
30,
766,
273,
599,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4177,
12,
2890,
16,
2489,
67,
1973,
16,
833,
16,
5323,
4672,
3536,
1477,
279,
9247,
1964,
2283,
8395,
775,
30,
766,
273,
365,
18,
5519,
18,
14508,
1335,
6394,
30,
766,
273,
599,
... |
watchpats.append("Forcing node %s to be terminated" % node) | watchpats.append("Node %s will be fenced because termination was requested" % node) | def __call__(self, node): self.incr("calls") if len(self.CM.Env["nodes"]) < 2: return self.skipped() | 0aba552579a25f5248eac3b2776812d5f9c7051e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11962/0aba552579a25f5248eac3b2776812d5f9c7051e/CTStests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
756,
4672,
365,
18,
267,
3353,
2932,
12550,
7923,
309,
562,
12,
2890,
18,
9611,
18,
3491,
9614,
4690,
6,
5717,
411,
576,
30,
327,
365,
18,
25346,
1435,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
756,
4672,
365,
18,
267,
3353,
2932,
12550,
7923,
309,
562,
12,
2890,
18,
9611,
18,
3491,
9614,
4690,
6,
5717,
411,
576,
30,
327,
365,
18,
25346,
1435,
... |
if re.search( table_name+r'[.]', table_param ) is None: | if table_param.find( table_name+'.' ) == -1: | def __init__( self, table_name, table_param, param_ranges_opt, verbose = False ): """ Parse --param-ranges option. Creates self.param which is the table_name and the table_param appended together (with a '.') and self.param_ranges, which is a list of tuples that give the lower parameter value, whether it is an open or closed boundary, and the same for the upper parameter. For example, if table_name is coinc_inspiral, table_param is mchirp and param_ranges_opt is '[2,8);[8,17]' will get: self.param = 'coinc_inspiral.mchirp' self.param_ranges = [ ( ('>=',2.0), ('<',8.0) ), ( ('>=',8.0), ('<=', 17.0) ) ] | aa613f9ca1679aa7fff3f9cfa9b4bb9989c85fd7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/aa613f9ca1679aa7fff3f9cfa9b4bb9989c85fd7/ligolw_sqlutils.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
365,
16,
1014,
67,
529,
16,
1014,
67,
891,
16,
579,
67,
14530,
67,
3838,
16,
3988,
273,
1083,
262,
30,
3536,
2884,
1493,
891,
17,
14530,
1456,
18,
10210,
365,
18... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
365,
16,
1014,
67,
529,
16,
1014,
67,
891,
16,
579,
67,
14530,
67,
3838,
16,
3988,
273,
1083,
262,
30,
3536,
2884,
1493,
891,
17,
14530,
1456,
18,
10210,
365,
18... |
regex_find_licenses = re.compile(r'(?<!\{)\{\{(?:[Tt]emplate:|)([^{]*?)[|\n<}]', re.DOTALL) | regex_find_licenses = re.compile(r'(?<!\{)\{\{(?:[Tt]emplate:|)([^{]+?)[|\n<}]', re.DOTALL) | def smartDetection(self): """ The bot instead of checking if there's a simple template in the image's description, checks also if that template is a license or something else. In this sense this type of check is smart. """ self.seems_ok = False self.license_found = None self.hiddentemplates = self.loadHiddenTemplates() self.licenses_found = self.image.getTemplates() self.whiteTemplatesFound = False regex_find_licenses = re.compile(r'(?<!\{)\{\{(?:[Tt]emplate:|)([^{]*?)[|\n<}]', re.DOTALL) templatesInTheImageRaw = regex_find_licenses.findall(self.imageCheckText) self.allLicenses = list() if self.list_licenses == []: raise wikipedia.Error(u'No licenses allowed provided, add that option to the code to make the script working correctly') # Found the templates ONLY in the image's description for template_selected in templatesInTheImageRaw: for templateReal in self.licenses_found: if self.convert_to_url(template_selected).lower().replace('template:', '') == \ self.convert_to_url(templateReal.title().lower().replace('template:', '')): if templateReal not in self.allLicenses: # don't put the same template, twice. self.allLicenses.append(templateReal) if self.licenses_found != []: self.templateInList() if self.license_found == None and self.allLicenses != list(): self.license_found = self.license_selected if not self.seems_ok and self.license_found != None: rep_text_license_fake = u"\n*[[:Image:%s]] seems to have " % self.imageName + \ "a ''fake license'', license detected: <nowiki>%s</nowiki>" % self.license_found regexFakeLicense = r"\* ?\[\[:Image:%s\]\] seems to have " % (re.escape(self.imageName)) + \ "a ''fake license'', license detected: <nowiki>%s</nowiki>$" % (re.escape(self.license_found)) printWithTimeZone(u"%s seems to have a fake license: %s, reporting..." % (self.imageName, self.license_found)) self.report_image(self.imageName, rep_text = rep_text_license_fake, addings = False, regex = regexFakeLicense) elif self.license_found != None: printWithTimeZone(u"%s seems ok, license found: %s..." % (self.imageName, self.license_found)) return (self.license_found, self.whiteTemplatesFound) | c8bd831cf36832667e225d2876619a25779e9888 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/4404/c8bd831cf36832667e225d2876619a25779e9888/checkimages.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13706,
10214,
12,
2890,
4672,
3536,
1021,
2512,
3560,
434,
6728,
309,
1915,
1807,
279,
4143,
1542,
316,
326,
1316,
1807,
2477,
16,
4271,
2546,
309,
716,
1542,
353,
279,
8630,
578,
5943,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13706,
10214,
12,
2890,
4672,
3536,
1021,
2512,
3560,
434,
6728,
309,
1915,
1807,
279,
4143,
1542,
316,
326,
1316,
1807,
2477,
16,
4271,
2546,
309,
716,
1542,
353,
279,
8630,
578,
5943,
... |
or self.getType() != u'bittorrent'): | or self.get_type() != u'bittorrent'): | def startUpload(self): if (self.get_state() not in (u'finished', u'uploading-paused') or self.getType() != u'bittorrent'): return self.manualUpload = True if _downloads.has_key(self.dlid): c = command.StartDownloadCommand(RemoteDownloader.dldaemon, self.dlid) c.send() else: self.before_changing_status() self.status['state'] = u'uploading' self.after_changing_status() self.restart() self.signal_change() | f2fa1d771c944ca514cd470008d100ce0c0da511 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12354/f2fa1d771c944ca514cd470008d100ce0c0da511/downloader.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
787,
4777,
12,
2890,
4672,
309,
261,
2890,
18,
588,
67,
2019,
1435,
486,
316,
261,
89,
11,
13527,
2187,
582,
11,
6327,
310,
17,
8774,
3668,
6134,
578,
365,
18,
588,
67,
723,
1435,
48... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
787,
4777,
12,
2890,
4672,
309,
261,
2890,
18,
588,
67,
2019,
1435,
486,
316,
261,
89,
11,
13527,
2187,
582,
11,
6327,
310,
17,
8774,
3668,
6134,
578,
365,
18,
588,
67,
723,
1435,
48... |
signal.alarm(max(math.ceil(prevTimeout - passed), 1)) | signal.alarm(int(max(math.ceil(prevTimeout - passed), 1))) | def alarmhandler(signum,stackframe): raise alarmExc("timeout expired") | 5c228c4a7a2880defc1ea9809ca61b9010176abf /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/34/5c228c4a7a2880defc1ea9809ca61b9010176abf/pycompat.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13721,
4176,
12,
2977,
379,
16,
3772,
3789,
4672,
1002,
13721,
29924,
2932,
4538,
7708,
7923,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13721,
4176,
12,
2977,
379,
16,
3772,
3789,
4672,
1002,
13721,
29924,
2932,
4538,
7708,
7923,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
license_path = os.path.join(path, filename) | if filename.startswith('/'): license_path = os.path.join(os.getcwd(), filename[1:]) else: license_path = os.path.join(path, filename) | def ParseDir(path): """Examine a third_party/foo component and extract its metadata.""" # Parse metadata fields out of README.chromium. # We examine "LICENSE" for the license file by default. metadata = { "License File": "LICENSE", # Relative path to license text. "Name": None, # Short name (for header on about:credits). "URL": None, # Project home page. } if path in SPECIAL_CASES: metadata.update(SPECIAL_CASES[path]) else: # Try to find README.chromium. readme_path = os.path.join(path, 'README.chromium') if not os.path.exists(readme_path): raise LicenseError("missing README.chromium") for line in open(readme_path): line = line.strip() if not line: break for key in metadata.keys(): field = key + ": " if line.startswith(field): metadata[key] = line[len(field):] # Check that all expected metadata is present. for key, value in metadata.iteritems(): if not value: raise LicenseError("couldn't find '" + key + "' line " "in README.chromium or licences.py " "SPECIAL_CASES") # Check that the license file exists. for filename in (metadata["License File"], "COPYING"): license_path = os.path.join(path, filename) if os.path.exists(license_path): metadata["License File"] = filename break license_path = None if not license_path: raise LicenseError("License file not found. " "Either add a file named LICENSE, " "import upstream's COPYING if available, " "or add a 'License File:' line to README.chromium " "with the appropriate path.") return metadata | 8283b10cafab170f6dfd7f266163aefc223892fe /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5060/8283b10cafab170f6dfd7f266163aefc223892fe/licenses.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2884,
1621,
12,
803,
4672,
3536,
424,
301,
558,
279,
12126,
67,
21214,
19,
11351,
1794,
471,
2608,
2097,
1982,
12123,
225,
468,
2884,
1982,
1466,
596,
434,
10746,
958,
18,
16298,
5077,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2884,
1621,
12,
803,
4672,
3536,
424,
301,
558,
279,
12126,
67,
21214,
19,
11351,
1794,
471,
2608,
2097,
1982,
12123,
225,
468,
2884,
1982,
1466,
596,
434,
10746,
958,
18,
16298,
5077,
1... |
cr = db.cursor() | def list(self): db = sql_db.db_connect('template1') cr = db.cursor() try: try: cr = db.cursor() db_user = tools.config["db_user"] if not db_user and os.name == 'posix': import pwd db_user = pwd.getpwuid(os.getuid())[0] if not db_user: cr.execute("select decode(usename, 'escape') from pg_user where usesysid=(select datdba from pg_database where datname=%s)", (tools.config["db_name"],)) res = cr.fetchone() db_user = res and str(res[0]) if db_user: cr.execute("select decode(datname, 'escape') from pg_database where datdba=(select usesysid from pg_user where usename=%s) and datname not in ('template0', 'template1', 'postgres')", (db_user,)) else: cr.execute("select decode(datname, 'escape') from pg_database where datname not in('template0', 'template1','postgres')") res = [str(name) for (name,) in cr.fetchall()] except: res = [] finally: cr.close() sql_db.close_db('template1') res.sort() return res | dbdd4c582ec44f0cc8ce5522a1b0d6d71d1d6536 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7397/dbdd4c582ec44f0cc8ce5522a1b0d6d71d1d6536/web_services.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
666,
12,
2890,
4672,
1319,
273,
1847,
67,
1966,
18,
1966,
67,
3612,
2668,
3202,
21,
6134,
775,
30,
775,
30,
1319,
67,
1355,
273,
8513,
18,
1425,
9614,
1966,
67,
1355,
11929,
309,
486,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
666,
12,
2890,
4672,
1319,
273,
1847,
67,
1966,
18,
1966,
67,
3612,
2668,
3202,
21,
6134,
775,
30,
775,
30,
1319,
67,
1355,
273,
8513,
18,
1425,
9614,
1966,
67,
1355,
11929,
309,
486,
... | |
def print_decls(decldict): | def print_decls(decldict, value): | def print_decls(decldict): for t, d in decldict.items(): d = d.keys() if not d: continue d.sort() print 'syn keyword %s%s %s' % (syntax_name, normalize(t), ' '.join(d)) | f7d9779be336dd165ada7dd3923f3f7d87d4a728 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6171/f7d9779be336dd165ada7dd3923f3f7d87d4a728/vim-syn-gen.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1172,
67,
323,
6429,
12,
8840,
1576,
16,
460,
4672,
364,
268,
16,
302,
316,
3496,
1576,
18,
3319,
13332,
302,
273,
302,
18,
2452,
1435,
309,
486,
302,
30,
1324,
302,
18,
3804,
1435,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1172,
67,
323,
6429,
12,
8840,
1576,
16,
460,
4672,
364,
268,
16,
302,
316,
3496,
1576,
18,
3319,
13332,
302,
273,
302,
18,
2452,
1435,
309,
486,
302,
30,
1324,
302,
18,
3804,
1435,
... |
elif UName == "osf1": | elif UName == "osf1": | def GetVersion(Key): # # Only read the file if an environment variable is "missing" (they # usually all are, ok), and only read it once. # #print("WriteVariablesIntoEnvironment:3") Value = Versions.get(Key) if Value: return Value # # CM3VERSION d5.7.1 # CM3VERSIONNUM 050701 # CM3LASTCHANGED 2009-01-21 # RegExp = re.compile("(" + "|".join(Versions.keys()) + ") (.+)$", re.IGNORECASE) ShFilePath = os.path.join(os.path.dirname(os.path.dirname(os.path.abspath(__file__))), "version") for Line in open(ShFilePath): Match = RegExp.match(Line) if Match: MatchKey = Match.group(1) # # We are here because one of them wasn't found, but we should be # sure only to overwrite what we don't have. # if not Versions[MatchKey]: Value = Match.group(2) Versions[MatchKey] = Value exec("%s = \"%s\"" % (MatchKey, Value), locals(), globals()) # # Make sure we found every key in the file (at least those # not defined in the environment) # MissingKey = None for Item in Versions.iteritems(): #print(Item) if Item[1] is None: MissingKey = Item[0] File = __file__ sys.stderr.write("%(File)s: %(MissingKey)s not found in %(ShFilePath)s\n" % vars()) if MissingKey: sys.exit(1) return Versions.get(Key) | a5382390fc1645f62582e9440bae90e340b3d767 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9328/a5382390fc1645f62582e9440bae90e340b3d767/pylib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
1444,
12,
653,
4672,
468,
468,
5098,
855,
326,
585,
309,
392,
3330,
2190,
353,
315,
7337,
6,
261,
451,
402,
468,
11234,
777,
854,
16,
1529,
3631,
471,
1338,
855,
518,
3647,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
1444,
12,
653,
4672,
468,
468,
5098,
855,
326,
585,
309,
392,
3330,
2190,
353,
315,
7337,
6,
261,
451,
402,
468,
11234,
777,
854,
16,
1529,
3631,
471,
1338,
855,
518,
3647,
18,
... |
self.assertEqual(sys.exc_info()[0], 'OverflowError') | self.assertEqual(sys.exc_info()[0], OverflowError) | def _raise_during_standard_error_message(*args, **kw): raise ZeroDivisionError('testing') | 706b2a37f5c717c21c45c27fefdb17762443caf2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9658/706b2a37f5c717c21c45c27fefdb17762443caf2/testSimpleItem.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
11628,
67,
72,
4017,
67,
10005,
67,
1636,
67,
2150,
30857,
1968,
16,
2826,
9987,
4672,
1002,
12744,
28223,
668,
2668,
3813,
310,
6134,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
11628,
67,
72,
4017,
67,
10005,
67,
1636,
67,
2150,
30857,
1968,
16,
2826,
9987,
4672,
1002,
12744,
28223,
668,
2668,
3813,
310,
6134,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
if data[f] != False: | if data[f]: data[f].sort() | def copy_data(self, cr, uid, id, default=None, context=None): """ Copy given record's data with all its fields values | 01f2d0c0db3743d0f4c0b2f6e774549050434cc2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12853/01f2d0c0db3743d0f4c0b2f6e774549050434cc2/orm.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1610,
67,
892,
12,
2890,
16,
4422,
16,
4555,
16,
612,
16,
805,
33,
7036,
16,
819,
33,
7036,
4672,
3536,
5631,
864,
1409,
1807,
501,
598,
777,
2097,
1466,
924,
2,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1610,
67,
892,
12,
2890,
16,
4422,
16,
4555,
16,
612,
16,
805,
33,
7036,
16,
819,
33,
7036,
4672,
3536,
5631,
864,
1409,
1807,
501,
598,
777,
2097,
1466,
924,
2,
-100,
-100,
-100,
-1... |
console.stdoutln("NOTE: Skipping Jmol-viewer section") | console.stdoutln("NOTE: Skipping Jmol-viewer section") | def html_tls_graph_path(self, chain, ntls): """Generates the HTML table describing the path (set of tls groups) for the given number of segments(h, or ntls) """ ## class HTMLReport() cpartition = chain.partition_collection.get_chain_partition(ntls) if cpartition == None: return None | 80aaf2a14dfcc5b684aad61f8ca72e81b353162c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10674/80aaf2a14dfcc5b684aad61f8ca72e81b353162c/html.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1729,
67,
17116,
67,
4660,
67,
803,
12,
2890,
16,
2687,
16,
9513,
3251,
4672,
3536,
6653,
326,
3982,
1014,
16868,
326,
589,
261,
542,
434,
6871,
3252,
13,
364,
326,
864,
1300,
434,
515... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1729,
67,
17116,
67,
4660,
67,
803,
12,
2890,
16,
2687,
16,
9513,
3251,
4672,
3536,
6653,
326,
3982,
1014,
16868,
326,
589,
261,
542,
434,
6871,
3252,
13,
364,
326,
864,
1300,
434,
515... |
if calname == "": | if calname is None: | def importProcess(self, text, extension=None, item=None, changes=None, previousView=None, updateCallback=None): # the item parameter is so that a share item can be passed in for us # to populate. | c4fe9d0f1a7be58d33664be3c6eb832fd34a787d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/c4fe9d0f1a7be58d33664be3c6eb832fd34a787d/ICalendar.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1930,
2227,
12,
2890,
16,
977,
16,
2710,
33,
7036,
16,
761,
33,
7036,
16,
3478,
33,
7036,
16,
2416,
1767,
33,
7036,
16,
1089,
2428,
33,
7036,
4672,
468,
326,
761,
1569,
353,
1427,
71... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1930,
2227,
12,
2890,
16,
977,
16,
2710,
33,
7036,
16,
761,
33,
7036,
16,
3478,
33,
7036,
16,
2416,
1767,
33,
7036,
16,
1089,
2428,
33,
7036,
4672,
468,
326,
761,
1569,
353,
1427,
71... |
schedIdList += " " + \ str( job.runningJob['schedulerId'] ).strip() command = "glite-wms-job-cancel --json --noint " + schedIdList out, ret = self.ExecuteCommand( self.proxyString + command ) if ret != 0 : raise SchedulerError('error executing glite-wms-job-cancel', out) elif ret == 0 and out.find("result: success") == -1 : raise SchedulerError('error', out) | jobsToKill.append(str( job.runningJob['schedulerId'] ).strip()) chunk = lambda ulist, step: map(lambda i: ulist[i:i+step], xrange(0, len(ulist), step)) lljobs = chunk(jobsToKill, self.killThreshold) for x in lljobs : schedIdList = ' '.join(x) command = "glite-wms-job-cancel --json --noint " + schedIdList out, ret = self.ExecuteCommand( self.proxyString + command ) if ret != 0 : raise SchedulerError('error executing glite-wms-job-cancel', out) elif ret == 0 and out.find("result: success") == -1 : raise SchedulerError('error', out) return 0 | def kill( self, obj ): """ kill job """ # the object passed is a job if type(obj) == Job and self.valid( obj.runningJob ): # check for the RunningJob integrity schedIdList = str( obj.runningJob['schedulerId'] ).strip() # the object passed is a Task elif type(obj) == Task : schedIdList = "" for job in obj.jobs: if not self.valid( job.runningJob ): continue schedIdList += " " + \ str( job.runningJob['schedulerId'] ).strip() command = "glite-wms-job-cancel --json --noint " + schedIdList out, ret = self.ExecuteCommand( self.proxyString + command ) if ret != 0 : raise SchedulerError('error executing glite-wms-job-cancel', out) elif ret == 0 and out.find("result: success") == -1 : raise SchedulerError('error', out) | 53423e453953da739f4a02830391dc962d257904 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8886/53423e453953da739f4a02830391dc962d257904/SchedulerGLite.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8673,
12,
365,
16,
1081,
262,
30,
3536,
8673,
1719,
3536,
225,
468,
326,
733,
2275,
353,
279,
1719,
309,
618,
12,
2603,
13,
422,
3956,
471,
365,
18,
877,
12,
1081,
18,
8704,
2278,
26... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8673,
12,
365,
16,
1081,
262,
30,
3536,
8673,
1719,
3536,
225,
468,
326,
733,
2275,
353,
279,
1719,
309,
618,
12,
2603,
13,
422,
3956,
471,
365,
18,
877,
12,
1081,
18,
8704,
2278,
26... |
f = sqrt(distnorm/sum) | f = sqrt(distnorm/max(sum,1e-6)) | def LSMT(self): # optimize the distance transformation # build vector o effect = 0 self.getDistance() o = [] for i in xrange(1,self.n): for j in xrange(i): o.append((self.originalDistances[i,j],(i,j))) o.sort(_mycompare) # find the ties in o, and construct the d vector sorting in order within ties d = [] td = [] uv = [] # numbers of consecutively tied o values (i,j) = o[0][1] distnorm = self.projectedDistances[i,j]*self.projectedDistances[i,j] td = [self.projectedDistances[i,j]] # fetch distance for l in xrange(1,len(o)): # copy now sorted distances in an array # but sort distances within a tied o (i,j) = o[l][1] cd = self.projectedDistances[i,j] distnorm += self.projectedDistances[i,j]*self.projectedDistances[i,j] if o[l][0] != o[l-1][0]: # differing value, flush sum = reduce(lambda x,y:x+y,td)+0.0 d.append([sum,len(td),sum/len(td),td]) td = [] td.append(cd) sum = reduce(lambda x,y:x+y,td)+0.0 d.append([sum,len(td),sum/len(td),td]) #### # keep merging non-monotonous areas in d monotony = 0 while not monotony and len(d) > 1: monotony = 1 pi = 0 # index n = 1 # n-areas nd = [] r = d[0] # current area for i in range(1,len(d)): tr = d[i] if r[2]>=tr[2]: monotony = 0 effect = 1 r[0] += tr[0] r[1] += tr[1] r[2] = tr[0]/tr[1] r[3] += tr[3] else: nd.append(r) r = tr nd.append(r) d = nd # normalizing multiplier sum = 0.0 for i in d: sum += i[2]*i[2]*i[1] f = sqrt(distnorm/sum) # transform O k = 0 for i in d: for j in range(i[1]): (ii,jj) = o[k][1] self.distances[ii,jj] = f*i[2] k += 1 assert(len(o) == k) self.freshD = 0 return effect | c474478c97cb0593e748e02c9775117c045b1180 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6366/c474478c97cb0593e748e02c9775117c045b1180/orngMDS.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
29204,
6152,
12,
2890,
4672,
468,
10979,
326,
3888,
8620,
468,
1361,
3806,
320,
5426,
273,
374,
365,
18,
588,
7200,
1435,
320,
273,
5378,
364,
277,
316,
12314,
12,
21,
16,
2890,
18,
82... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
29204,
6152,
12,
2890,
4672,
468,
10979,
326,
3888,
8620,
468,
1361,
3806,
320,
5426,
273,
374,
365,
18,
588,
7200,
1435,
320,
273,
5378,
364,
277,
316,
12314,
12,
21,
16,
2890,
18,
82... |
vbox.pack_start(hbox, gtk.TRUE, gtk.TRUE, 0) | vbox.pack_start(hbox, True, True, 0) | def __init__(self): self.return_keyval = gtk.gdk.keyval_from_name("Return") self.left_keyval = gtk.gdk.keyval_from_name("Left") self.right_keyval = gtk.gdk.keyval_from_name("Right") self.up_keyval = gtk.gdk.keyval_from_name("Up") self.down_keyval = gtk.gdk.keyval_from_name("Down") self.backspace_keyval = gtk.gdk.keyval_from_name("BackSpace") self.delete_keyval = gtk.gdk.keyval_from_name("Delete") self.home_keyval = gtk.gdk.keyval_from_name("Home") self.end_keyval = gtk.gdk.keyval_from_name("End") | 2744524bbe73c70a7af8961c0bcea5e15fbb0a13 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3867/2744524bbe73c70a7af8961c0bcea5e15fbb0a13/ipmigui.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
4672,
365,
18,
2463,
67,
856,
1125,
273,
22718,
18,
75,
2883,
18,
856,
1125,
67,
2080,
67,
529,
2932,
990,
7923,
365,
18,
4482,
67,
856,
1125,
273,
22718,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
4672,
365,
18,
2463,
67,
856,
1125,
273,
22718,
18,
75,
2883,
18,
856,
1125,
67,
2080,
67,
529,
2932,
990,
7923,
365,
18,
4482,
67,
856,
1125,
273,
22718,
... |
mul1s rax, t1 mulel rax muleh ah, flags=(OF,CF) | mul1s rax, t1, flags=(OF,CF) mulel rax muleh ah | def macroop IMUL_B_M | b64d0bdeda1662091746c3695b4429fcc6f69342 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6787/b64d0bdeda1662091746c3695b4429fcc6f69342/multiply_and_divide.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11522,
556,
6246,
1506,
67,
38,
67,
49,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11522,
556,
6246,
1506,
67,
38,
67,
49,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
path = [module.__filename__] | head, tail = os.path.split(module.__filename__) path = [head] | def reload(self, module, path=None): if path is None and hasattr(module, '__filename__'): path = [module.__filename__] return ihooks.ModuleImporter.reload(self, module, path) | 9eca6c4d59803ed1f62627fa0024674267283d18 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/9eca6c4d59803ed1f62627fa0024674267283d18/rexec.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7749,
12,
2890,
16,
1605,
16,
589,
33,
7036,
4672,
309,
589,
353,
599,
471,
3859,
12,
2978,
16,
4940,
3459,
7250,
4672,
910,
16,
5798,
273,
1140,
18,
803,
18,
4939,
12,
2978,
16186,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7749,
12,
2890,
16,
1605,
16,
589,
33,
7036,
4672,
309,
589,
353,
599,
471,
3859,
12,
2978,
16,
4940,
3459,
7250,
4672,
910,
16,
5798,
273,
1140,
18,
803,
18,
4939,
12,
2978,
16186,
... |
if (sys.platform == "win32"): | if (sys.platform != "win32"): CopyAllHeaders('panda/src/glxdisplay') else: | def CreatePandaVersionFiles(): version1=int(VERSION.split(".")[0]) version2=int(VERSION.split(".")[1]) version3=int(VERSION.split(".")[2]) nversion=version1*1000000+version2*1000+version3 pandaversion_h = PANDAVERSION_H.replace("VERSION1",str(version1)) pandaversion_h = pandaversion_h.replace("VERSION2",str(version2)) pandaversion_h = pandaversion_h.replace("VERSION3",str(version3)) pandaversion_h = pandaversion_h.replace("NVERSION",str(nversion)) checkpandaversion_cxx = CHECKPANDAVERSION_CXX.replace("VERSION1",str(version1)) checkpandaversion_cxx = checkpandaversion_cxx.replace("VERSION2",str(version2)) checkpandaversion_cxx = checkpandaversion_cxx.replace("VERSION3",str(version3)) checkpandaversion_cxx = checkpandaversion_cxx.replace("NVERSION",str(nversion)) checkpandaversion_h = CHECKPANDAVERSION_H.replace("VERSION1",str(version1)) checkpandaversion_h = checkpandaversion_h.replace("VERSION2",str(version2)) checkpandaversion_h = checkpandaversion_h.replace("VERSION3",str(version3)) checkpandaversion_h = checkpandaversion_h.replace("NVERSION",str(nversion)) ConditionalWriteFile('built/include/pandaVersion.h', pandaversion_h) ConditionalWriteFile('built/include/checkPandaVersion.cxx', checkpandaversion_cxx) ConditionalWriteFile('built/include/checkPandaVersion.h', checkpandaversion_h) ConditionalWriteFile("built/tmp/null.cxx","") | f7a6726c5099ebae926d5d09ee18a7f322e81784 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8543/f7a6726c5099ebae926d5d09ee18a7f322e81784/makepanda.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1788,
52,
464,
69,
1444,
2697,
13332,
1177,
21,
33,
474,
12,
5757,
18,
4939,
2932,
1199,
25146,
20,
5717,
1177,
22,
33,
474,
12,
5757,
18,
4939,
2932,
1199,
25146,
21,
5717,
1177,
23,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1788,
52,
464,
69,
1444,
2697,
13332,
1177,
21,
33,
474,
12,
5757,
18,
4939,
2932,
1199,
25146,
20,
5717,
1177,
22,
33,
474,
12,
5757,
18,
4939,
2932,
1199,
25146,
21,
5717,
1177,
23,
... |
elif _isfunction(v) or _isclass(v): | elif _isfunction(v) or _isclass(v) or _ismethod(v): | def run__test__(self, d, name): """d, name -> Treat dict d like module.__test__. | 5f8b0b1fd4801b4808223077643c24d41130b81e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/5f8b0b1fd4801b4808223077643c24d41130b81e/doctest.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
972,
3813,
972,
12,
2890,
16,
302,
16,
508,
4672,
3536,
72,
16,
508,
317,
26564,
2065,
302,
3007,
1605,
16186,
3813,
25648,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
972,
3813,
972,
12,
2890,
16,
302,
16,
508,
4672,
3536,
72,
16,
508,
317,
26564,
2065,
302,
3007,
1605,
16186,
3813,
25648,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
<unittest.TestResult run=3 errors=0 failures=3> | <unittest.TestResult run=3 errors=0 failures=2> | def test_DocFileSuite(): """We can test tests found in text files using a DocFileSuite. We create a suite by providing the names of one or more text files that include examples: >>> import unittest >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt') >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=3> The test files are looked for in the directory containing the calling module. A package keyword argument can be provided to specify a different relative location. >>> import unittest >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt', ... package='test') >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=3> '/' should be used as a path separator. It will be converted to a native separator at run time: >>> suite = doctest.DocFileSuite('../test/test_doctest.txt') >>> suite.run(unittest.TestResult()) <unittest.TestResult run=1 errors=0 failures=1> If DocFileSuite is used from an interactive session, then files are resolved relative to the directory of sys.argv[0]: >>> import new, os.path, test.test_doctest >>> save_argv = sys.argv >>> sys.argv = [test.test_doctest.__file__] >>> suite = doctest.DocFileSuite('test_doctest.txt', ... package=new.module('__main__')) >>> sys.argv = save_argv By setting `module_relative=False`, os-specific paths may be used (including absolute paths and paths relative to the working directory): >>> # Get the absolute path of the test package. >>> test_doctest_path = os.path.abspath(test.test_doctest.__file__) >>> test_pkg_path = os.path.split(test_doctest_path)[0] >>> # Use it to find the absolute path of test_doctest.txt. >>> test_file = os.path.join(test_pkg_path, 'test_doctest.txt') >>> suite = doctest.DocFileSuite(test_file, module_relative=False) >>> suite.run(unittest.TestResult()) <unittest.TestResult run=1 errors=0 failures=1> It is an error to specify `package` when `module_relative=False`: >>> suite = doctest.DocFileSuite(test_file, module_relative=False, ... package='test') Traceback (most recent call last): ValueError: Package may only be specified for module-relative paths. You can specify initial global variables: >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt', ... globs={'favorite_color': 'blue'}) >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=2> In this case, we supplied a missing favorite color. You can provide doctest options: >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt', ... optionflags=doctest.DONT_ACCEPT_BLANKLINE, ... globs={'favorite_color': 'blue'}) >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=3> And, you can provide setUp and tearDown functions: You can supply setUp and teatDoen functions: >>> def setUp(t): ... import test.test_doctest ... test.test_doctest.sillySetup = True >>> def tearDown(t): ... import test.test_doctest ... del test.test_doctest.sillySetup Here, we installed a silly variable that the test expects: >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt', ... setUp=setUp, tearDown=tearDown) >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=2> But the tearDown restores sanity: >>> import test.test_doctest >>> test.test_doctest.sillySetup Traceback (most recent call last): ... AttributeError: 'module' object has no attribute 'sillySetup' The setUp and tearDown funtions are passed test objects. Here, we'll use a setUp function to set the favorite color in test_doctest.txt: >>> def setUp(test): ... test.globs['favorite_color'] = 'blue' >>> suite = doctest.DocFileSuite('test_doctest.txt', setUp=setUp) >>> suite.run(unittest.TestResult()) <unittest.TestResult run=1 errors=0 failures=0> Here, we didn't need to use a tearDown function because we modified the test globals. The test globals are automatically cleared for us after a test. Tests in a file run using `DocFileSuite` can also access the `__file__` global, which is set to the name of the file containing the tests: >>> suite = doctest.DocFileSuite('test_doctest3.txt') >>> suite.run(unittest.TestResult()) <unittest.TestResult run=1 errors=0 failures=0> If the tests contain non-ASCII characters, we have to specify which encoding the file is encoded with. We do so by using the `encoding` parameter: >>> suite = doctest.DocFileSuite('test_doctest.txt', ... 'test_doctest2.txt', ... 'test_doctest4.txt', ... encoding='utf-8') >>> suite.run(unittest.TestResult()) <unittest.TestResult run=3 errors=0 failures=2> """ | b1a9f2772ba21f268a8164908714e5a106be50d1 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8546/b1a9f2772ba21f268a8164908714e5a106be50d1/test_doctest.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
1759,
812,
13587,
13332,
3536,
3218,
848,
1842,
7434,
1392,
316,
977,
1390,
1450,
279,
3521,
812,
13587,
18,
225,
1660,
752,
279,
11371,
635,
17721,
326,
1257,
434,
1245,
578,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
1759,
812,
13587,
13332,
3536,
3218,
848,
1842,
7434,
1392,
316,
977,
1390,
1450,
279,
3521,
812,
13587,
18,
225,
1660,
752,
279,
11371,
635,
17721,
326,
1257,
434,
1245,
578,
... |
path = os.path.join(self.configpath, LAYER_DIRS[layer]) | path = os.path.join(self.mapServ_inst.configpath, LAYER_DIRS[layer]) | def coord_to_path(self, tile_coord, layer): self.lock.acquire() path = os.path.join(self.configpath, LAYER_DIRS[layer]) path = fileUtils.check_dir(path) path = fileUtils.check_dir(path, '%d' % tile_coord[2]) path = fileUtils.check_dir(path, "%d" % (tile_coord[0] / 1024)) path = fileUtils.check_dir(path, "%d" % (tile_coord[0] % 1024)) path = fileUtils.check_dir(path, "%d" % (tile_coord[1] / 1024)) self.lock.release() return os.path.join(path, "%d.png" % (tile_coord[1] % 1024)) | e47ed51cf02f70923536b0b3bfe9e89d6c1e5fbf /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5373/e47ed51cf02f70923536b0b3bfe9e89d6c1e5fbf/tilesRepoFS.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2745,
67,
869,
67,
803,
12,
2890,
16,
4769,
67,
5732,
16,
3018,
4672,
365,
18,
739,
18,
1077,
1039,
1435,
589,
273,
1140,
18,
803,
18,
5701,
12,
2890,
18,
1458,
827,
90,
67,
8591,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2745,
67,
869,
67,
803,
12,
2890,
16,
4769,
67,
5732,
16,
3018,
4672,
365,
18,
739,
18,
1077,
1039,
1435,
589,
273,
1140,
18,
803,
18,
5701,
12,
2890,
18,
1458,
827,
90,
67,
8591,
... |
return self.tk.call(self._w, 'info', 'right', index) | return self.tk.call(self._w, 'info', 'right', index) | def info_right(self, index): | 0049f0036898e656085d5c214ba07642a9c25a6e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/0049f0036898e656085d5c214ba07642a9c25a6e/Tix.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1123,
67,
4083,
12,
2890,
16,
770,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1123,
67,
4083,
12,
2890,
16,
770,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
r""" A simple periodic sequence. INPUT: n -- non negative integer OUTPUT: integer -- function value EXAMPLES: sage: a = sloane.A000035;a | def __init__(self): r""" | def _eval(self, n): return Integer(2*n) | 69a199217d612e1c590af73e16003812c85b93ec /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/69a199217d612e1c590af73e16003812c85b93ec/sloane_functions.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8622,
12,
2890,
16,
290,
4672,
327,
2144,
12,
22,
14,
82,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8622,
12,
2890,
16,
290,
4672,
327,
2144,
12,
22,
14,
82,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
outputs[-1:] = \ | outputs.extend ( | def copy_tree (src, dst, preserve_mode=1, preserve_times=1, preserve_symlinks=0, update=0, verbose=0, dry_run=0): """Copy an entire directory tree 'src' to a new location 'dst'. Both 'src' and 'dst' must be directory names. If 'src' is not a directory, raise DistutilsFileError. If 'dst' does not exist, it is created with 'mkpath()'. The end result of the copy is that every file in 'src' is copied to 'dst', and directories under 'src' are recursively copied to 'dst'. Return the list of files copied (under their output names) -- note that if 'update' is true, this might be less than the list of files considered. Return value is not affected by 'dry_run'. 'preserve_mode' and 'preserve_times' are the same as for 'copy_file'; note that they only apply to regular files, not to directories. If 'preserve_symlinks' is true, symlinks will be copied as symlinks (on platforms that support them!); otherwise (the default), the destination of the symlink will be copied. 'update' and 'verbose' are the same as for 'copy_file'.""" if not dry_run and not os.path.isdir (src): raise DistutilsFileError, \ "cannot copy tree %s: not a directory" % src try: names = os.listdir (src) except os.error, (errno, errstr): if dry_run: names = [] else: raise DistutilsFileError, \ "error listing files in %s: %s" % (src, errstr) if not dry_run: mkpath (dst, verbose=verbose) outputs = [] for n in names: src_name = os.path.join (src, n) dst_name = os.path.join (dst, n) if preserve_symlinks and os.path.islink (src_name): link_dest = os.readlink (src_name) if verbose: print "linking %s -> %s" % (dst_name, link_dest) if not dry_run: os.symlink (link_dest, dst_name) outputs.append (dst_name) elif os.path.isdir (src_name): outputs[-1:] = \ copy_tree (src_name, dst_name, preserve_mode, preserve_times, preserve_symlinks, update, verbose, dry_run) else: if (copy_file (src_name, dst_name, preserve_mode, preserve_times, update, verbose, dry_run)): outputs.append (dst_name) return outputs | f49afccb1be8414afed28dee04df060fdc754715 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/f49afccb1be8414afed28dee04df060fdc754715/util.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1610,
67,
3413,
261,
4816,
16,
3046,
16,
9420,
67,
3188,
33,
21,
16,
9420,
67,
8293,
33,
21,
16,
9420,
67,
21278,
87,
33,
20,
16,
1089,
33,
20,
16,
3988,
33,
20,
16,
10299,
67,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1610,
67,
3413,
261,
4816,
16,
3046,
16,
9420,
67,
3188,
33,
21,
16,
9420,
67,
8293,
33,
21,
16,
9420,
67,
21278,
87,
33,
20,
16,
1089,
33,
20,
16,
3988,
33,
20,
16,
10299,
67,
2... |
process_list = get_cmd_output('ps -e -o comm,user=').strip().split('\n') | process_list = commands.getoutput('ps -e -o comm,user=').strip().split('\n') | def testing_status() : server_name = 'rcssserver' server_user = '' process_list = get_cmd_output('ps -e -o comm,user=').strip().split('\n') process_list.pop(0) team_count_map = {} cmd_count_map = {} matched_cmds = {} for process in process_list : info = process.split() (cmd, user) = (info[0], info[1]) cmd_count_map[cmd] = 1 + cmd_count_map.get(cmd, 0) for pattern in team_name_map.keys() : if pattern.match(cmd) : matched_cmds[cmd] = 1 team_name = team_name_map[pattern] team_count_map[team_name] = 1 + team_count_map.get(team_name, 0) break if not server_user and cmd == server_name : server_user = user message = ' #rcssserver: ' if server_user : server_count = cmd_count_map[server_name] if len(team_count_map) <= 1 : count_cmd_map = {} sucess = False for cmd in cmd_count_map.keys() : count_cmd_map.setdefault(cmd_count_map[cmd], []).append(cmd) count_list = count_cmd_map.keys() count_list.sort() count_list.reverse() for count in count_list : for cmd in count_cmd_map[count] : if not matched_cmds.has_key(cmd) : team_count_map['[' + cmd + ']'] = cmd_count_map[cmd] if len(team_count_map) >= 2 : sucess = True break if sucess : break message += '%d, %s, (' % (server_count, server_user) for team in sorted(team_count_map.keys()) : message += '%s x %d, ' % (team, team_count_map[team]) message = message.rstrip(', ') + ')' else : message += '0' return message | 004ed3b935e239e61feea17b4ad35f4c4ed06166 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/228/004ed3b935e239e61feea17b4ad35f4c4ed06166/client.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7769,
67,
2327,
1435,
294,
1438,
67,
529,
273,
296,
1310,
1049,
3567,
11,
1438,
67,
1355,
273,
875,
225,
1207,
67,
1098,
273,
4364,
18,
588,
2844,
2668,
1121,
300,
73,
300,
83,
1543,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7769,
67,
2327,
1435,
294,
1438,
67,
529,
273,
296,
1310,
1049,
3567,
11,
1438,
67,
1355,
273,
875,
225,
1207,
67,
1098,
273,
4364,
18,
588,
2844,
2668,
1121,
300,
73,
300,
83,
1543,
... |
elif text.find(' | elif text.find(' | #enddef - closes a scope | 0ae3bc473d70b671840d2febe0a653a31a4589b5 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9355/0ae3bc473d70b671840d2febe0a653a31a4589b5/wmliterator.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
468,
409,
536,
300,
13445,
279,
2146,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
468,
409,
536,
300,
13445,
279,
2146,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
self.enableCmdBtns() self.enableHistBtns() | self.enableCmdButtons() self.enableHistButtons() | def __init__(self, master, actor, | d1909c3750f356969ac52bbe45612907850f17e7 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6354/d1909c3750f356969ac52bbe45612907850f17e7/GuideWdg.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
4171,
16,
8327,
16,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
4171,
16,
8327,
16,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
connection = MTurkConnection(operation.action) | connection = MTurkConnection(action) | def operation_execute(key): operation = datastore.get(key) if not operation.completed and not operation.error: connection = MTurkConnection(operation.action) try: operation.execute(connection) self.completed = datetime.now() except (BotoClientError, BotoServerError), response: self.error = response_error(response) self.put() | b4cc6abefee1ce6aef654df5be65743d01c92c4c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14392/b4cc6abefee1ce6aef654df5be65743d01c92c4c/main.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1674,
67,
8837,
12,
856,
4672,
1674,
273,
9290,
18,
588,
12,
856,
13,
225,
309,
486,
1674,
18,
13615,
471,
486,
1674,
18,
1636,
30,
1459,
273,
22068,
20628,
1952,
12,
1128,
13,
225,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1674,
67,
8837,
12,
856,
4672,
1674,
273,
9290,
18,
588,
12,
856,
13,
225,
309,
486,
1674,
18,
13615,
471,
486,
1674,
18,
1636,
30,
1459,
273,
22068,
20628,
1952,
12,
1128,
13,
225,
... |
loader = importer.find_module(modname) if hasattr(loader,'get_source'): | try: loader = importer.find_module(modname) except SyntaxError: continue if hasattr(loader, 'get_source'): | def run(self, callback, key=None, completer=None, onerror=None): if key: key = key.lower() self.quit = False seen = {} | 126c879b406f24363bdfccffa16e62acc695aea1 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8546/126c879b406f24363bdfccffa16e62acc695aea1/pydoc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
16,
1348,
16,
498,
33,
7036,
16,
31848,
33,
7036,
16,
20991,
33,
7036,
4672,
309,
498,
30,
498,
273,
498,
18,
8167,
1435,
365,
18,
27176,
273,
1083,
5881,
273,
2618,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
16,
1348,
16,
498,
33,
7036,
16,
31848,
33,
7036,
16,
20991,
33,
7036,
4672,
309,
498,
30,
498,
273,
498,
18,
8167,
1435,
365,
18,
27176,
273,
1083,
5881,
273,
2618,
... |
raise ValueError, 'You must provide a source address for this\ | raise ValueError, 'You must provide a source address for this \ | def __call__(self): recipients = [str(mail.strip()) for mail in \ self.element.recipients.split(',')] mailhost = getToolByName(aq_inner(self.context), "MailHost") if not mailhost: raise ComponentLookupError, 'You must have a Mailhost utility to \ | d21e9ee976da48d9609790851a11196e38318d7a /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12184/d21e9ee976da48d9609790851a11196e38318d7a/mail.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
4672,
12045,
273,
306,
701,
12,
4408,
18,
6406,
10756,
364,
4791,
316,
521,
365,
18,
2956,
18,
27925,
18,
4939,
12,
2187,
6134,
65,
4791,
2564,
273,
336,
636... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
4672,
12045,
273,
306,
701,
12,
4408,
18,
6406,
10756,
364,
4791,
316,
521,
365,
18,
2956,
18,
27925,
18,
4939,
12,
2187,
6134,
65,
4791,
2564,
273,
336,
636... |
and job.runningJob['status'] != 'SD': : | and job.runningJob['status'] != 'SD': | def query(self, obj, objType='node') : """ query status and eventually other scheduler related information """ | c39edd21061a0424fe50fad47d6984792fe4bfcf /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8886/c39edd21061a0424fe50fad47d6984792fe4bfcf/Scheduler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
843,
12,
2890,
16,
1081,
16,
30078,
2218,
2159,
6134,
294,
3536,
843,
1267,
471,
18011,
1308,
8129,
3746,
1779,
3536,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
843,
12,
2890,
16,
1081,
16,
30078,
2218,
2159,
6134,
294,
3536,
843,
1267,
471,
18011,
1308,
8129,
3746,
1779,
3536,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
print 'called Break' | def trapit(what): print 'TRAPPED', what what.trap(juice.UnhandledRemoteJuiceError) | e56dcf564dc442d95c85eb44f592b6672bad21ad /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8962/e56dcf564dc442d95c85eb44f592b6672bad21ad/test_q2q.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
23034,
305,
12,
23770,
4672,
1172,
296,
4349,
2203,
12120,
2187,
4121,
4121,
18,
27170,
12,
78,
89,
1812,
18,
26952,
5169,
46,
89,
1812,
668,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
23034,
305,
12,
23770,
4672,
1172,
296,
4349,
2203,
12120,
2187,
4121,
4121,
18,
27170,
12,
78,
89,
1812,
18,
26952,
5169,
46,
89,
1812,
668,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
... | |
self.isBusy = False | def onCompleted(self, links): self.sayTo(self.employer, "Build complete.") self.isBusy = False for k in links.keys(): self.sayTo(self.employer, k + ": " + links[k]) | f0130d66ba28b8a109656ac4672968689759b937 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10496/f0130d66ba28b8a109656ac4672968689759b937/buildbot.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
9556,
12,
2890,
16,
4716,
4672,
365,
18,
87,
528,
774,
12,
2890,
18,
351,
2817,
264,
16,
315,
3116,
3912,
1199,
13,
364,
417,
316,
4716,
18,
2452,
13332,
365,
18,
87,
528,
774,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
9556,
12,
2890,
16,
4716,
4672,
365,
18,
87,
528,
774,
12,
2890,
18,
351,
2817,
264,
16,
315,
3116,
3912,
1199,
13,
364,
417,
316,
4716,
18,
2452,
13332,
365,
18,
87,
528,
774,
... | |
pass | raise NotImplementedError | def index_shuffle(self, ind): """Change the indices. | 34282af38d23e10b103d48f60c68963923bb99f3 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5735/34282af38d23e10b103d48f60c68963923bb99f3/constraints.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
770,
67,
31980,
12,
2890,
16,
1547,
4672,
3536,
3043,
326,
4295,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
770,
67,
31980,
12,
2890,
16,
1547,
4672,
3536,
3043,
326,
4295,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
path = path.replace('/', os.sep) | path = self._absnorm(path) | def wait_until_removed(self, path, timeout='1 minute'): """Waits until the given file or directory is removed. | 4e92df2860c270b97129fb5577b1caec2faab8a2 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/6988/4e92df2860c270b97129fb5577b1caec2faab8a2/OperatingSystem.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2529,
67,
12198,
67,
14923,
12,
2890,
16,
589,
16,
2021,
2218,
21,
8044,
11,
4672,
3536,
26153,
3180,
326,
864,
585,
578,
1867,
353,
3723,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2529,
67,
12198,
67,
14923,
12,
2890,
16,
589,
16,
2021,
2218,
21,
8044,
11,
4672,
3536,
26153,
3180,
326,
864,
585,
578,
1867,
353,
3723,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
out += ch | out += chr(ord(ch)) | def PyUnicode_EncodeUTF7(s, size, encodeSetO, encodeWhiteSpace, errors): | 0e540b97b49394c1d38eed66d9b3eec5969d6c9a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6934/0e540b97b49394c1d38eed66d9b3eec5969d6c9a/unicodecodec.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4707,
16532,
67,
5509,
5159,
27,
12,
87,
16,
963,
16,
2017,
694,
51,
16,
2017,
23108,
16,
1334,
4672,
225,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4707,
16532,
67,
5509,
5159,
27,
12,
87,
16,
963,
16,
2017,
694,
51,
16,
2017,
23108,
16,
1334,
4672,
225,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
semi = string.rfind(type, ';') | semi = type.rfind(';') | def open_data(self, url, data=None): """Use "data" URL.""" # ignore POSTed data # # syntax of data URLs: # dataurl := "data:" [ mediatype ] [ ";base64" ] "," data # mediatype := [ type "/" subtype ] *( ";" parameter ) # data := *urlchar # parameter := attribute "=" value import StringIO, mimetools, time try: [type, data] = string.split(url, ',', 1) except ValueError: raise IOError, ('data error', 'bad data URL') if not type: type = 'text/plain;charset=US-ASCII' semi = string.rfind(type, ';') if semi >= 0 and '=' not in type[semi:]: encoding = type[semi+1:] type = type[:semi] else: encoding = '' msg = [] msg.append('Date: %s'%time.strftime('%a, %d %b %Y %T GMT', time.gmtime(time.time()))) msg.append('Content-type: %s' % type) if encoding == 'base64': import base64 data = base64.decodestring(data) else: data = unquote(data) msg.append('Content-length: %d' % len(data)) msg.append('') msg.append(data) msg = string.join(msg, '\n') f = StringIO.StringIO(msg) headers = mimetools.Message(f, 0) f.fileno = None # needed for addinfourl return addinfourl(f, headers, url) | 2a3596c7d04c9635576a712f85db07ce4de4bff2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/2a3596c7d04c9635576a712f85db07ce4de4bff2/urllib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1696,
67,
892,
12,
2890,
16,
880,
16,
501,
33,
7036,
4672,
3536,
3727,
315,
892,
6,
1976,
12123,
468,
2305,
5485,
329,
501,
468,
468,
6279,
434,
501,
10414,
30,
468,
501,
718,
282,
5... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1696,
67,
892,
12,
2890,
16,
880,
16,
501,
33,
7036,
4672,
3536,
3727,
315,
892,
6,
1976,
12123,
468,
2305,
5485,
329,
501,
468,
468,
6279,
434,
501,
10414,
30,
468,
501,
718,
282,
5... |
return False, "Cannot find block device %s" % (disk,) | _Fail("Cannot find block device %s", disk) | def BlockdevGrow(disk, amount): """Grow a stack of block devices. This function is called recursively, with the childrens being the first ones to resize. @type disk: L{objects.Disk} @param disk: the disk to be grown @rtype: (status, result) @return: a tuple with the status of the operation (True/False), and the errors message if status is False """ r_dev = _RecursiveFindBD(disk) if r_dev is None: return False, "Cannot find block device %s" % (disk,) try: r_dev.Grow(amount) except errors.BlockDeviceError, err: _Fail("Failed to grow block device: %s", err, exc=True) return True, None | afdc3985644be8d2b91d9f0e37dfda391a30d036 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7542/afdc3985644be8d2b91d9f0e37dfda391a30d036/backend.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3914,
5206,
30948,
12,
10863,
16,
3844,
4672,
3536,
30948,
279,
2110,
434,
1203,
7166,
18,
225,
1220,
445,
353,
2566,
8536,
16,
598,
326,
2325,
87,
3832,
326,
1122,
5945,
358,
7041,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3914,
5206,
30948,
12,
10863,
16,
3844,
4672,
3536,
30948,
279,
2110,
434,
1203,
7166,
18,
225,
1220,
445,
353,
2566,
8536,
16,
598,
326,
2325,
87,
3832,
326,
1122,
5945,
358,
7041,
18,
... |
def __init__(self, toplot, xvals=None, yvals=None, | def __init__(self, data, xvals=None, yvals=None, | def __init__(self, toplot, xvals=None, yvals=None, binary=1, inline=_unset, **keyw): """GridData constructor. | 57b7d171447e091edf5aff623b42289284f0cd78 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9569/57b7d171447e091edf5aff623b42289284f0cd78/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
501,
16,
619,
4524,
33,
7036,
16,
677,
4524,
33,
7036,
16,
3112,
33,
21,
16,
6370,
33,
67,
18579,
16,
2826,
856,
91,
4672,
3536,
6313,
751,
3885,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
501,
16,
619,
4524,
33,
7036,
16,
677,
4524,
33,
7036,
16,
3112,
33,
21,
16,
6370,
33,
67,
18579,
16,
2826,
856,
91,
4672,
3536,
6313,
751,
3885,
18,
... |
iid = model.InvoiceItem(description='Penguin Dinner ticket (included in registraion)', | iid = model.InvoiceItem(description='Penguin Dinner ticket (included in registration)', | def _build_invoice(self): person = c.registration.person r = c.registration | 4564a7b8c923eccde8041311a7f7d56b61a99834 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12856/4564a7b8c923eccde8041311a7f7d56b61a99834/registration.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
3510,
67,
16119,
12,
2890,
4672,
6175,
273,
276,
18,
14170,
18,
12479,
436,
273,
276,
18,
14170,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
3510,
67,
16119,
12,
2890,
4672,
6175,
273,
276,
18,
14170,
18,
12479,
436,
273,
276,
18,
14170,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
for base in (_HKEY_CLASSES_ROOT, _HKEY_LOCAL_MACHINE, _HKEY_CURRENT_USER, _HKEY_USERS): | for base in (HKEY_CLASSES_ROOT, HKEY_LOCAL_MACHINE, HKEY_CURRENT_USER, HKEY_USERS): | def get_msvc_paths (path, version='6.0', platform='x86'): """Get a list of devstudio directories (include, lib or path). Return a list of strings; will be empty list if unable to access the registry or appropriate registry keys not found.""" if not _can_read_reg: return [] L = [] if path=='lib': path= 'Library' path = string.upper(path + ' Dirs') K = ('Software\\Microsoft\\Devstudio\\%s\\' + 'Build System\\Components\\Platforms\\Win32 (%s)\\Directories') % \ (version,platform) for base in (_HKEY_CLASSES_ROOT, _HKEY_LOCAL_MACHINE, _HKEY_CURRENT_USER, _HKEY_USERS): try: k = _RegOpenKeyEx(base,K) i = 0 while 1: try: (p,v,t) = _RegEnumValue(k,i) if string.upper(p) == path: V = string.split(v,';') for v in V: if v == '' or v in L: continue L.append(v) break i = i + 1 except _RegError: break except _RegError: pass return L | f524abb944b7edf65acd200dfe2f4076696b42e3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/f524abb944b7edf65acd200dfe2f4076696b42e3/msvccompiler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
959,
4227,
67,
4481,
261,
803,
16,
1177,
2218,
26,
18,
20,
2187,
4072,
2218,
92,
5292,
11,
4672,
3536,
967,
279,
666,
434,
4461,
334,
4484,
6402,
261,
6702,
16,
2561,
578,
5... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
959,
4227,
67,
4481,
261,
803,
16,
1177,
2218,
26,
18,
20,
2187,
4072,
2218,
92,
5292,
11,
4672,
3536,
967,
279,
666,
434,
4461,
334,
4484,
6402,
261,
6702,
16,
2561,
578,
5... |
line = line[m.end() - 1:] | line = line[m.end(1):] elif empty: line = line[m.end(1):] | def subconvert(line, ofp, table, discards, autoclosing, knownempty, endchar=None): stack = [] while line: if line[0] == endchar and not stack: return line[1:] m = _comment_rx.match(line) if m: text = m.group(1) if text: ofp.write("(COMMENT\n") ofp.write("-%s\n" % encode(text)) ofp.write(")COMMENT\n") ofp.write("-\\n\n") else: ofp.write("-\\n\n") line = line[m.end():] continue m = _begin_env_rx.match(line) if m: # re-write to use the macro handler line = r"\%s%s" % (m.group(1), line[m.end():]) continue m =_end_env_rx.match(line) if m: # end of environment envname = m.group(1) if envname == "document": # special magic for n in stack[1:]: if n not in autoclosing: raise LaTeXFormatError("open element on stack: " + `n`) # should be more careful, but this is easier to code: stack = [] ofp.write(")document\n") elif envname == stack[-1]: ofp.write(")%s\n" % envname) del stack[-1] else: | fd9c2ca53cfbd9d2df1c1c71382345041d3849ae /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/fd9c2ca53cfbd9d2df1c1c71382345041d3849ae/latex2esis.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
720,
6283,
12,
1369,
16,
434,
84,
16,
1014,
16,
10388,
87,
16,
23173,
10910,
16,
4846,
5531,
16,
679,
3001,
33,
7036,
4672,
2110,
273,
5378,
1323,
980,
30,
309,
980,
63,
20,
65,
422,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
720,
6283,
12,
1369,
16,
434,
84,
16,
1014,
16,
10388,
87,
16,
23173,
10910,
16,
4846,
5531,
16,
679,
3001,
33,
7036,
4672,
2110,
273,
5378,
1323,
980,
30,
309,
980,
63,
20,
65,
422,... |
if not type(obj) in [str, unicode]: | if not _is_in(type(obj), [str, unicode]): | def _require_string(obj): if not type(obj) in [str, unicode]: raise NamespaceRequirementError | 3c551ef3827ddde6b2f4571bc3357a59857633a1 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7995/3c551ef3827ddde6b2f4571bc3357a59857633a1/namespace.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
6528,
67,
1080,
12,
2603,
4672,
309,
486,
389,
291,
67,
267,
12,
723,
12,
2603,
3631,
306,
701,
16,
5252,
65,
4672,
1002,
6005,
18599,
668,
565,
2,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
6528,
67,
1080,
12,
2603,
4672,
309,
486,
389,
291,
67,
267,
12,
723,
12,
2603,
3631,
306,
701,
16,
5252,
65,
4672,
1002,
6005,
18599,
668,
565,
2,
-100,
-100,
-100,
-100,
-100,
... |
row.coinc_event_id = coinc_event_id | cid = lsctables.CoincTable.get_next_id() row.coinc_event_id = cid | def populate_inspiral_tables(MBTA_frame, UID, set_keys = MBTA_set_keys, \ process_id = 'process:process_id:0', \ event_id_dict = insp_event_id_dict, \ coinc_event_id_base=InspiralCoincIdBase): """ create xml file and populate the SnglInspiral and CoincInspiral tables from a coinc .gwf file from MBTA xmldoc: xml file to append the tables to MBTA_frame: frame file to get info about triggers from set_keys: columns in the SnglInspiral Table to set process_id: process_id event_id_dict: {ifo:event_id} dictionary to assign event_id's coinc_event_id: coinc_event_id detectors: detectors participating in the coinc returns xmldoc and contents of the comment field """ #initialize xml document xmldoc = ligolw.Document() xmldoc.appendChild(ligolw.LIGO_LW()) #dictionaries to store about individual triggers end_time_s = {} end_time_ns = {} snr = {} mass1 = {} mass2 = {} Deff = {} mchirp = {} eta = {} #extract the information from the frame file events = Fr.frgetevent(MBTA_frame) #get the ifos from the event name for event in events: if 'MbtaHLV' in event['name']: detectors = H1L1V1_detlist elif 'MbtaHL' in event['name']: detectors = H1L1_detlist elif 'MbtaHV' in event['name']: detectors = H1V1_detlist elif 'MbtaH' in event['name']: detectors = H1_detlist elif 'MbtaLV' in event['name']: detectors = L1V1_detlist elif 'MbtaL' in event['name']: detectors = L1_detlist elif 'MbtaV' in event['name']: detectors = V1_detlist else: raise ValueError, "Invalid FrEvent name" log_data = event['comment'] + '\n' far = [line.split(':')[1].split()[0] for line in log_data.splitlines() if \ 'False Alarm Rate' in line][0] for ifo in detectors: end_time_s[ifo], end_time_ns[ifo] = str(event[ifo+':end_time']).split('.') snr[ifo] = float(event[ifo+':SNR']) mass1[ifo] = float(event[ifo+':mass1']) mass2[ifo] = float(event[ifo+':mass2']) mchirp[ifo], eta[ifo] = compute_mchirp_eta(mass1[ifo],mass2[ifo]) Deff[ifo] = float(event[ifo+':eff_distance']) #fill the SnglInspiralTable sin_table = lsctables.New(lsctables.SnglInspiralTable) xmldoc.childNodes[0].appendChild(sin_table) for ifo in detectors: row = sin_table.RowType() row.ifo = ifo row.search = 'MBTA' row.end_time = int(end_time_s[ifo]) row.end_time_ns = int(end_time_ns[ifo]) row.mass1 = mass1[ifo] row.mass2 = mass2[ifo] row.mchirp = mchirp[ifo] row.mtotal = mass1[ifo] + mass2[ifo] row.eta = eta[ifo] row.snr = snr[ifo] row.eff_distance = Deff[ifo] row.event_id = event_id_dict[ifo] row.process_id = process_id row.channel = '' #zero out the rest of the columns #should work in chi2 and chi2cut for key in sin_table.validcolumns.keys(): if key not in set_keys: setattr(row,key,None) sin_table.append(row) #CoincInspiralTable #using the conventions found in: #https://www.lsc-group.phys.uwm.edu/ligovirgo/cbcnote/S6Plan/ #090505160219S6PlanningNotebookCoinc_and_Experiment_Tables_ihope_implementation? #highlight=%28coinc%29|%28table%29 if len(detectors) < 2: return xmldoc, log_data, detectors coinc_event_id = coinc_event_id_base + str(UID) cin_table = lsctables.New(lsctables.CoincInspiralTable) xmldoc.childNodes[0].appendChild(cin_table) row = cin_table.RowType() row.set_ifos(detectors) row.coinc_event_id = coinc_event_id row.end_time = int(end_time_s['H1']) row.end_time_ns = int(end_time_ns['H1']) row.mass = (sum(mass1.values()) + sum(mass2.values()))/3 row.mchirp = sum(mchirp.values())/3 #the snr here is really the snr NOT effective snr row.snr = pow(sum([x*x for x in snr.values()]),0.5) #far is triggers/day row.false_alarm_rate = float(far) row.combined_far = 0 cin_table.append(row) return xmldoc, log_data, detectors | af257218e83ac728902869fa75f174902babf41f /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5758/af257218e83ac728902869fa75f174902babf41f/utils.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6490,
67,
267,
1752,
481,
287,
67,
9373,
12,
7969,
9833,
67,
3789,
16,
10034,
16,
444,
67,
2452,
273,
17897,
9833,
67,
542,
67,
2452,
16,
521,
1207,
67,
350,
273,
296,
2567,
30,
2567... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6490,
67,
267,
1752,
481,
287,
67,
9373,
12,
7969,
9833,
67,
3789,
16,
10034,
16,
444,
67,
2452,
273,
17897,
9833,
67,
542,
67,
2452,
16,
521,
1207,
67,
350,
273,
296,
2567,
30,
2567... |
obj = self.parse_header(data) | obj = self._parse_realmedia_file(data) | def parse_objects(self, stream): | 47877dfb08c6f43c07b0e458250fb48397a0d7e6 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/5551/47877dfb08c6f43c07b0e458250fb48397a0d7e6/realmedia.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
6911,
12,
2890,
16,
1407,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
6911,
12,
2890,
16,
1407,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
serializer.start_keyword(self._keyword) serializer.end_keyword(self._keyword) | if self._keyword: serializer.start_keyword(self._keyword) serializer.end_keyword(self._keyword) | def serialize(self, serializer): serializer.start_keyword(self._keyword) serializer.end_keyword(self._keyword) | 2aaf76c987baf2de8e2b740de02e1f6dd44f4b75 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6988/2aaf76c987baf2de8e2b740de02e1f6dd44f4b75/fixture.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4472,
12,
2890,
16,
6340,
4672,
6340,
18,
1937,
67,
11041,
12,
2890,
6315,
11041,
13,
6340,
18,
409,
67,
11041,
12,
2890,
6315,
11041,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4472,
12,
2890,
16,
6340,
4672,
6340,
18,
1937,
67,
11041,
12,
2890,
6315,
11041,
13,
6340,
18,
409,
67,
11041,
12,
2890,
6315,
11041,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
types_callProxyMethod = [StringType,ListType,DictionaryType] | types_callProxyMethod = [StringType,TupleType,DictionaryType] | def initializeLcgFileCatalogProxyHandler(serviceInfo): global lfcCombinedClient lfcCombinedClient = LcgFileCatalogCombinedClient() return S_OK() | 1b93f09d0e7ef9a769bbec95dee9da3c61cca28b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12864/1b93f09d0e7ef9a769bbec95dee9da3c61cca28b/LcgFileCatalogProxyHandler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4046,
48,
26275,
812,
9769,
3886,
1503,
12,
3278,
966,
4672,
2552,
328,
7142,
27994,
1227,
328,
7142,
27994,
1227,
273,
511,
26275,
812,
9769,
27994,
1227,
1435,
327,
348,
67,
3141,
1435,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4046,
48,
26275,
812,
9769,
3886,
1503,
12,
3278,
966,
4672,
2552,
328,
7142,
27994,
1227,
328,
7142,
27994,
1227,
273,
511,
26275,
812,
9769,
27994,
1227,
1435,
327,
348,
67,
3141,
1435,
... |
H = G._list_of_elements_in_H() | H = [ int(x) for x in G._list_of_elements_in_H() ] | def _coset_reduction_data_first_coord(G): """ Compute data used for determining the canonical coset representative of an element of SL_2(Z) modulo G. This function specfically returns data needed for the first part of the reduction step (the first coordinate). | eb301dff111c888467560bbc7545e7b8e22d5329 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/eb301dff111c888467560bbc7545e7b8e22d5329/congroup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
14445,
278,
67,
1118,
4062,
67,
892,
67,
3645,
67,
5732,
12,
43,
4672,
3536,
8155,
501,
1399,
364,
23789,
326,
7378,
4987,
278,
23174,
434,
392,
930,
434,
348,
48,
67,
22,
12,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
14445,
278,
67,
1118,
4062,
67,
892,
67,
3645,
67,
5732,
12,
43,
4672,
3536,
8155,
501,
1399,
364,
23789,
326,
7378,
4987,
278,
23174,
434,
392,
930,
434,
348,
48,
67,
22,
12,
6... |
def finalize_product(self): "if requested, make bonds and/or join units into one part" desc = " (N = %d)" % self.ncopies self.final_msg_accum = "%s making %s%s: " % (self.msg_modename.split()[0], self.product_type, desc) msg0 = "leaving mode, finalizing product..." self.status_msg(self.final_msg_accum + msg0) print "fyi: extrude params not mentioned in statusbar: offset = %r, tol = %r" % (self.offset, self.bond_tolerance) if self.whendone_make_bonds: | def finalize_product(self, cancelling = 0): "if requested, make bonds and/or join units into one part; cancelling = 1 means just do cleanup, use diff msgs" if not cancelling: desc = " (N = %d)" % self.ncopies self.final_msg_accum = "%s making %s%s: " % (self.msg_modename.split()[0], self.product_type, desc) msg0 = "leaving mode, finalizing product..." self.status_msg(self.final_msg_accum + msg0) print "fyi: extrude params not mentioned in statusbar: offset = %r, tol = %r" % (self.offset, self.bond_tolerance) else: msg = "%s cancelled (alpha warning: might not fully restore initial state)" % (self.msg_modename.split()[0],) self.status_msg( msg) if self.whendone_make_bonds and not cancelling: | def StateDone(self): ## self.update_from_controls() #k 041017 night - will this help or hurt? since hard to know, not adding it now. # restore normal appearance for mol in self.molcopies: try: del mol._colorfunc mol.changeapp(0) except: pass self.finalize_product() # ... and emit status message about it return None | 93def9cf394f749ee4f6f85c71a4645a60945878 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/93def9cf394f749ee4f6f85c71a4645a60945878/extrudeMode.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3287,
7387,
12,
2890,
4672,
7541,
365,
18,
2725,
67,
2080,
67,
24350,
1435,
468,
79,
374,
9803,
1611,
27,
290,
750,
300,
903,
333,
2809,
578,
366,
295,
88,
35,
3241,
7877,
358,
5055,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3287,
7387,
12,
2890,
4672,
7541,
365,
18,
2725,
67,
2080,
67,
24350,
1435,
468,
79,
374,
9803,
1611,
27,
290,
750,
300,
903,
333,
2809,
578,
366,
295,
88,
35,
3241,
7877,
358,
5055,
... |
print "*** FINISHED ***" | print "***_FINISHED_***" | def updateMonitoringInformation( jobs ): | f28943d46a1275727d8c070b84ce3598227b3c8b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1488/f28943d46a1275727d8c070b84ce3598227b3c8b/Remote.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
18410,
5369,
12,
6550,
262,
30,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
18410,
5369,
12,
6550,
262,
30,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
assert descr in node.curfields | fieldnode = node.curfields[descr] resbox = op.result opt.assertval(resbox, fieldnode.source) | def getfield_gc(self, opt, op): node = opt.spec.getnode(op.args[0]) descr = op.descr assert isinstance(descr, AbstractValue) if node.virtual: assert descr in node.curfields return None return op | 2a4f036953ba7f84f56cdc173003fb1c30722732 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6934/2a4f036953ba7f84f56cdc173003fb1c30722732/optimize3.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
1518,
67,
13241,
12,
2890,
16,
2153,
16,
1061,
4672,
756,
273,
2153,
18,
2793,
18,
588,
2159,
12,
556,
18,
1968,
63,
20,
5717,
18426,
273,
1061,
18,
28313,
1815,
1549,
12,
28313,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
1518,
67,
13241,
12,
2890,
16,
2153,
16,
1061,
4672,
756,
273,
2153,
18,
2793,
18,
588,
2159,
12,
556,
18,
1968,
63,
20,
5717,
18426,
273,
1061,
18,
28313,
1815,
1549,
12,
28313,
... |
self._stop = False | self._stop = self._closed = False | def __init__(self, name, env=None, prefix='zodb_', config=None): """Create a new storage. | ea82bc70eae6ac2afa78935c67947cd34432ac5f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9658/ea82bc70eae6ac2afa78935c67947cd34432ac5f/BerkeleyBase.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
508,
16,
1550,
33,
7036,
16,
1633,
2218,
94,
10605,
67,
2187,
642,
33,
7036,
4672,
3536,
1684,
279,
394,
2502,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
508,
16,
1550,
33,
7036,
16,
1633,
2218,
94,
10605,
67,
2187,
642,
33,
7036,
4672,
3536,
1684,
279,
394,
2502,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,... |
"%s cannot be child of %s" % (repr(node), repr(self) ) | "%s cannot be child of %s" % (repr(node), repr(self)) | def appendChild(self, node): if node.nodeType not in self.childNodeTypes: raise HierarchyRequestErr, \ "%s cannot be child of %s" % (repr(node), repr(self) ) if node.parentNode is not None: node.parentNode.removeChild(node) if self.childNodes: last = self.lastChild node.previousSibling = last last.nextSibling = node else: node.previousSibling = None node.nextSibling = None self.childNodes.append(node) if self._makeParentNodes: node.parentNode = self return node | 3d9b92654f18f440a68a8e80ed41265487140c54 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/3d9b92654f18f440a68a8e80ed41265487140c54/minidom.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5833,
12,
2890,
16,
756,
4672,
309,
756,
18,
2159,
559,
486,
316,
365,
18,
3624,
907,
2016,
30,
1002,
670,
6000,
691,
2524,
16,
521,
2213,
87,
2780,
506,
1151,
434,
738,
87,
6,
738,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5833,
12,
2890,
16,
756,
4672,
309,
756,
18,
2159,
559,
486,
316,
365,
18,
3624,
907,
2016,
30,
1002,
670,
6000,
691,
2524,
16,
521,
2213,
87,
2780,
506,
1151,
434,
738,
87,
6,
738,
... |
'</text:list-item>' '</text:list>') | '<text:list-item>' '<text:list>' '<text:list-item>' '<text:p>eggs</text:p>' '</text:list-item>' '</text:list>' '</text:list-item>' '</text:list>' '</text:list-item>' '</text:list>') | def test_append_sub_item(self): # XXX Fail until a true odf_list element is implemented breakfast = odf_create_list() spam = odf_create_list_item(u'spam') ham = odf_create_list_item(u'ham') eggs = odf_create_list_item(u'eggs') | c85637406aa6ba3e15ed3c5f3443dfe819814624 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10612/c85637406aa6ba3e15ed3c5f3443dfe819814624/test_content.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
6923,
67,
1717,
67,
1726,
12,
2890,
4672,
468,
11329,
8911,
3180,
279,
638,
320,
2180,
67,
1098,
930,
353,
8249,
898,
8076,
273,
320,
2180,
67,
2640,
67,
1098,
1435,
24824,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
6923,
67,
1717,
67,
1726,
12,
2890,
4672,
468,
11329,
8911,
3180,
279,
638,
320,
2180,
67,
1098,
930,
353,
8249,
898,
8076,
273,
320,
2180,
67,
2640,
67,
1098,
1435,
24824,
2... |
rcr t1, t1, imm | rcri t1, t1, imm | def macroop RCR_M_I | 66911a1fab9900768db67610346585245a484ef4 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/6787/66911a1fab9900768db67610346585245a484ef4/rotate.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11522,
556,
534,
5093,
67,
49,
67,
45,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11522,
556,
534,
5093,
67,
49,
67,
45,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
self.liststore = liststore = gtk.ListStore(str, str) | self.liststore = liststore = gtk.ListStore(str) | def __init__(self, *arg): gtk.Entry.__init__(self, *arg) self.set_editable(False) self.hwindow = gtk.Window() self.hwindow.set_position(gtk.WIN_POS_MOUSE) self.hwindow.set_decorated(False) vbox = gtk.VBox() self.liststore = liststore = gtk.ListStore(str, str) self.treeview = treeview = gtk.TreeView(liststore) self.hwindow.set_default_size(200, 300) | 52a18dc9abc5f0d68a536f1ffff67bac53cc3297 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2456/52a18dc9abc5f0d68a536f1ffff67bac53cc3297/config.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
380,
3175,
4672,
22718,
18,
1622,
16186,
2738,
972,
12,
2890,
16,
380,
3175,
13,
365,
18,
542,
67,
19653,
12,
8381,
13,
225,
365,
18,
76,
5668,
273,
22... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
380,
3175,
4672,
22718,
18,
1622,
16186,
2738,
972,
12,
2890,
16,
380,
3175,
13,
365,
18,
542,
67,
19653,
12,
8381,
13,
225,
365,
18,
76,
5668,
273,
22... |
""" | """% d | def c_code(self, node, name, (img2d, filtersflipped), (z, ), sub): if node.inputs[0].type.dtype != node.inputs[1].type.dtype: raise NotImplementedError() assert node.inputs[0].type.dtype == node.inputs[1].type.dtype d=locals() d.update(sub) | 6ef4fe09e7b91d6db497a10ee67d728eda46182a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12438/6ef4fe09e7b91d6db497a10ee67d728eda46182a/conv.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
276,
67,
710,
12,
2890,
16,
756,
16,
508,
16,
261,
6081,
22,
72,
16,
3415,
12357,
1845,
3631,
261,
94,
16,
262,
16,
720,
4672,
309,
756,
18,
10029,
63,
20,
8009,
723,
18,
8972,
480... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
276,
67,
710,
12,
2890,
16,
756,
16,
508,
16,
261,
6081,
22,
72,
16,
3415,
12357,
1845,
3631,
261,
94,
16,
262,
16,
720,
4672,
309,
756,
18,
10029,
63,
20,
8009,
723,
18,
8972,
480... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.