rem stringlengths 2 226k | add stringlengths 0 227k | context stringlengths 8 228k | meta stringlengths 156 215 | input_ids list | attention_mask list | labels list |
|---|---|---|---|---|---|---|
line_pat = re.compile("^(?P<group>[^ \t]+)[ \t]+(.*)$") | line_pat = re.compile(b'^(?P<group>[^ \t]+)[ \t]+(.*)$') | def descriptions(self, group_pattern): """Get descriptions for a range of groups.""" line_pat = re.compile("^(?P<group>[^ \t]+)[ \t]+(.*)$") # Try the more std (acc. to RFC2980) LIST NEWSGROUPS first resp, raw_lines = self.longcmd('LIST NEWSGROUPS ' + group_pattern) if resp[:3] != "215": # Now the deprecated XGTITLE. This either raises an error # or succeeds with the same output structure as LIST # NEWSGROUPS. resp, raw_lines = self.longcmd('XGTITLE ' + group_pattern) lines = [] for raw_line in raw_lines: match = line_pat.search(raw_line.strip()) if match: lines.append(match.group(1, 2)) return resp, lines | 9bd2973908ef080e31fd8a2b44c929188a135cb0 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8125/9bd2973908ef080e31fd8a2b44c929188a135cb0/nntplib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
15550,
12,
2890,
16,
1041,
67,
4951,
4672,
3536,
967,
15550,
364,
279,
1048,
434,
3252,
12123,
980,
67,
16330,
273,
283,
18,
11100,
12,
70,
19856,
3680,
52,
32,
1655,
34,
5969,
521,
88... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
15550,
12,
2890,
16,
1041,
67,
4951,
4672,
3536,
967,
15550,
364,
279,
1048,
434,
3252,
12123,
980,
67,
16330,
273,
283,
18,
11100,
12,
70,
19856,
3680,
52,
32,
1655,
34,
5969,
521,
88... |
string = string.replace(define, value) strings[locale][id] = string | cur_string = cur_string.replace(define, value) strings[locale][string_id] = cur_string | def main(argv): if len(argv) < 4: print __doc__ % argv[0] sys.exit(1) locales = [] translation_dir = argv.pop(-1) source_file = argv.pop(-1) target_file = argv.pop(-1) defines = getDefines(argv) # If this is an official build, warnings are treated as errors. treat_warnings_as_errors = defines.get('OFFICIAL_BUILD') != None # Languages are specified via a define. if defines.has_key('I18N_LANGUAGES'): raw_locales = re.sub(r'[()]', '', defines['I18N_LANGUAGES']) locales = re.split(r',', raw_locales) filename = os.path.basename(source_file) # The file specified as the source is considered the most up to date set of # strings. The translated files are compared to this to determine if the # localized strings match. source_strings = getStrings(source_file) strings = {} for locale in locales: localized_strings = {} try: localized_strings = getStrings(os.path.join(translation_dir, locale, filename)) except: print "Warning: %s missing for locale %s" % (filename, locale) if treat_warnings_as_errors: sys.exit(2) # This block simply checks if the localized strings are out of date. if locale == "en-US": if len(localized_strings) > len(source_strings): print "Warning: Strings are out of date, build is not localized." if treat_warnings_as_errors: sys.exit(2) else: for string_id, string in source_strings.items(): # If the english string is missing or different from the source, the # string is out of date. if ((not localized_strings.has_key(string_id)) or localized_strings[string_id] != source_strings[string_id]): print "Warning: Strings are out of date, build is not localized." if treat_warnings_as_errors: sys.exit(2) break strings[locale] = {} for id in source_strings.keys(): # If there is no localized string, substitute the string from the source. string = localized_strings.get(id, source_strings[id]) # Replace any macros as specified on the commandline for define, value in defines.items(): string = string.replace(define, value) strings[locale][id] = string # Extract the file extension from the target filename. match = re.search(r'\.(.*?)$', target_file) if not match: print 'Target %s does not have a file extension' % target_file sys.exit(1) file_type = match.group(1) parse_funcs = {'js': createJavaScriptFromStrings, 'rc': createRCFromStrings} if parse_funcs.has_key(file_type): parse_funcs[file_type](target_file, strings) else: print "Unknown output file type: %s\n" % (file_type) sys.exit(3) | 00ea35b9791951af01325916fb87ff7ff1bbf5a5 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1918/00ea35b9791951af01325916fb87ff7ff1bbf5a5/parse_stab.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
19485,
4672,
309,
562,
12,
19485,
13,
411,
1059,
30,
1172,
1001,
2434,
972,
738,
5261,
63,
20,
65,
2589,
18,
8593,
12,
21,
13,
225,
6922,
273,
5378,
225,
4794,
67,
1214,
27... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
19485,
4672,
309,
562,
12,
19485,
13,
411,
1059,
30,
1172,
1001,
2434,
972,
738,
5261,
63,
20,
65,
2589,
18,
8593,
12,
21,
13,
225,
6922,
273,
5378,
225,
4794,
67,
1214,
27... |
return '<a href="https://bugs.webkit.org/attachment.cgi?id=' + value + '&action=prettypatch">' + value + '</a>' | return '<a href="https://bugs.webkit.org/attachment.cgi?id=%s&action=prettypatch">%s</a>' % (value, value) | def webkit_attachment_id(value): return '<a href="https://bugs.webkit.org/attachment.cgi?id=' + value + '&action=prettypatch">' + value + '</a>' | 6d4913aeadcbd99535c47d023588cc7510e64e49 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9392/6d4913aeadcbd99535c47d023588cc7510e64e49/webkit_extras.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
28491,
67,
11461,
67,
350,
12,
1132,
4672,
327,
2368,
69,
3897,
1546,
4528,
2207,
19381,
18,
29319,
18,
3341,
19,
11461,
18,
19062,
35,
350,
5095,
87,
10,
1128,
33,
1484,
748,
879,
505... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
28491,
67,
11461,
67,
350,
12,
1132,
4672,
327,
2368,
69,
3897,
1546,
4528,
2207,
19381,
18,
29319,
18,
3341,
19,
11461,
18,
19062,
35,
350,
5095,
87,
10,
1128,
33,
1484,
748,
879,
505... |
pept = get_pept_linked(N, CA, PRO) | pept = get_pept_linked(N, CA, PRO) | def calc_blocks_normal(self, N, CA, PRO, Carbon, Oxigen, Nitrogen): """MBH scheme with linked blocks where each side chain is considered as a block attached to the CA atom.""" | 0545e65e19aa908b8fca4a796cf437e3e835b905 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11051/0545e65e19aa908b8fca4a796cf437e3e835b905/nmatools.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7029,
67,
7996,
67,
6130,
12,
2890,
16,
423,
16,
6425,
16,
4629,
16,
13353,
16,
531,
92,
30577,
16,
423,
8912,
4507,
4672,
3536,
7969,
44,
4355,
598,
8459,
4398,
1625,
1517,
4889,
2687... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7029,
67,
7996,
67,
6130,
12,
2890,
16,
423,
16,
6425,
16,
4629,
16,
13353,
16,
531,
92,
30577,
16,
423,
8912,
4507,
4672,
3536,
7969,
44,
4355,
598,
8459,
4398,
1625,
1517,
4889,
2687... |
def quotient(self, sub, check=True): | def quotient(self, sub, check=True, positive_point=None, positive_dual_point=None): | def quotient(self, sub, check=True): """ Return the quotient of ``self`` by the given sublattice ``sub``. | 2ecb64877587a338c723df487639345b09bed118 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9417/2ecb64877587a338c723df487639345b09bed118/toric_lattice.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
26708,
12,
2890,
16,
720,
16,
866,
33,
5510,
16,
6895,
67,
1153,
33,
7036,
16,
6895,
67,
72,
1462,
67,
1153,
33,
7036,
4672,
3536,
2000,
326,
26708,
434,
12176,
2890,
10335,
635,
326,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
26708,
12,
2890,
16,
720,
16,
866,
33,
5510,
16,
6895,
67,
1153,
33,
7036,
16,
6895,
67,
72,
1462,
67,
1153,
33,
7036,
4672,
3536,
2000,
326,
26708,
434,
12176,
2890,
10335,
635,
326,
... |
PARAMETER["false_northing",0], PARAMETER["pseudo_standard_parallel_1",-6.5]]""" | PARAMETER["false_northing",0]]""" | def isis2_1(): tst = gdaltest.GDALTest( 'ISIS2', 'arvidson_original_truncated.cub', 1, 382 ) expected_prj = """PROJCS["SIMPLE_CYLINDRICAL VENUS", GEOGCS["GCS_VENUS", DATUM["D_VENUS", SPHEROID["VENUS",6051000,0]], PRIMEM["Reference_Meridian",0], UNIT["degree",0.0174532925199433]], PROJECTION["Equirectangular"], PARAMETER["latitude_of_origin",-6.5], PARAMETER["central_meridian",0], PARAMETER["false_easting",0], PARAMETER["false_northing",0], PARAMETER["pseudo_standard_parallel_1",-6.5]]""" expected_gt = (10157400.403618813, 1200.0000476837158, 0.0, -585000.02324581146, 0.0, -1200.0000476837158) return tst.testOpen( check_prj = expected_prj, check_gt = expected_gt ) | b089ef5fa35dd86c33ecb76f2b720151beaf4133 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/10290/b089ef5fa35dd86c33ecb76f2b720151beaf4133/isis2.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
353,
291,
22,
67,
21,
13332,
225,
268,
334,
273,
15551,
2390,
395,
18,
27338,
1013,
4709,
12,
296,
45,
15664,
22,
2187,
296,
297,
1246,
816,
67,
8830,
67,
23558,
690,
18,
71,
373,
21... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
353,
291,
22,
67,
21,
13332,
225,
268,
334,
273,
15551,
2390,
395,
18,
27338,
1013,
4709,
12,
296,
45,
15664,
22,
2187,
296,
297,
1246,
816,
67,
8830,
67,
23558,
690,
18,
71,
373,
21... |
Makefile.write("cm3: $(Objects)\n\t$(Link) -o cm3 *.o\n") Make.write(Link + " -o cm3 *.o\n") MakeVerbose.write("echo " + Link + " -o cm3 *.o\n") MakeVerbose.write(Link + " -o cm3 *.o\n") Make.close() Makefile.close() MakeVerbose.close() | Makefile.write("cm3: $(Objects)\n\t") for a in [Make, Makefile]: a.write("$(Link) -o cm3 *.o\n") for a in [ "m3-libs/libm3/src/os/POSIX/OSConfigPosix.m3", "m3-libs/libm3/src/random/m3makefile", "m3-libs/m3core/src/C/m3makefile", "m3-libs/m3core/src/Csupport/m3makefile", "m3-libs/m3core/src/float/m3makefile", "m3-libs/m3core/src/runtime/m3makefile", "m3-libs/m3core/src/runtime/common/Compiler.tmpl", "m3-libs/m3core/src/thread/m3makefile", "m3-libs/m3core/src/time/POSIX/m3makefile", "m3-libs/m3core/src/unix/m3makefile", "m3-sys/cminstall/src/config-no-install/Unix.common", "m3-sys/m3cc/src/m3makefile", "m3-sys/m3middle/src/Target.i3", "m3-sys/m3middle/src/Target.m3", "scripts/python/pylib.py", "m3-libs/m3core/src/C/" + Target + "/Csetjmp.i3", "m3-libs/m3core/src/C/" + Target + "/Csignal.i3", "m3-libs/m3core/src/C/" + Target + "/Cstdio.i3", "m3-libs/m3core/src/C/" + Target + "/Cstring.i3", "m3-libs/m3core/src/C/" + Target + "/m3makefile"]: source = os.path.join(Root, a) if FileExists(source): name = GetLastPathElement(a) reldir = RemoveLastPathElement(a) destdir = os.path.join(BootDir, reldir) dest = os.path.join(destdir, name) try: os.makedirs(destdir) except: pass CopyFile(source, dest) for b in [UpdateSource, Make]: b.write("mkdir -p /dev2/cm3/" + reldir + "\n") b.write("cp " + a + " /dev2/cm3/" + a + "\n") for a in [UpdateSource, Make, Makefile]: a.close() | def Boot(): global BuildLocal BuildLocal += " -boot -keep" Version = "1" Compile = ("gcc -gstabs+ " + ({ "SPARC32_LINUX" : "-m32 -munaligned-doubles", "SPARC64_LINUX" : "-m64 -munaligned-doubles", "LINUXLIBC6" : "-m32 -fPIC -mno-align-double", "AMD64_LINUX" : "-m64 -fPIC -mno-align-double" }.get(Target) or "")) Link = (Compile + " " + ({ "PPC32_OPENBSD" : "-lm -lpthread", "SPARC64_OPENBSD" : "-lm -lpthread" }.get(Target) or "")) Assemble = ("as " + ({ "SPARC32_LINUX" : "-32", "SPARC64_LINUX" : "-64", "LINUXLIBC6" : "--32", "AMD64_LINUX" : "--64" }.get(Target) or "")) BootDir = "/cm3-boot-POSIX-" + Target + "-" + Version P = [ "import-libs", "m3core", "libm3", "sysutils", "m3middle", "m3quake", "m3objfile", "m3linker", "m3back", "m3front", "cm3" ] if Target == "NT386": P += ["mklib"] #DoPackage(["", "realclean"] + P) or sys.exit(1) DoPackage(["", "buildlocal"] + P) or sys.exit(1) if os.path.isdir(BootDir): shutil.rmtree(BootDir) os.mkdir(BootDir) # # This would probably be a good use of XSL (xml style sheets) # Make = open(os.path.join(BootDir, "make.sh"), "wb") MakeVerbose = open(os.path.join(BootDir, "makeverbose.sh"), "wb") Makefile = open(os.path.join(BootDir, "Makefile"), "wb") Makefile.write("all: cm3\nAssemble=" + Assemble + "\nCompile=" + Compile + "\nLink=" + Link + "\n") for q in P: dir = GetPackagePath(q) for a in os.listdir(os.path.join(Root, dir, Config)): if (a.endswith(".ms") or a.endswith(".is") or a.endswith(".c")): CopyFile(os.path.join(Root, dir, Config, a), BootDir) Makefile.write("Objects += " + a + ".o\n" + a + ".o: " + a + "\n\t") if a.endswith(".c"): Make.write(Compile + " -c " + a + "\n") MakeVerbose.write("echo " + Compile + " -c " + a + "\n") MakeVerbose.write(Compile + " -C " + a + "\n") Makefile.write("$(Compile) -c " + a + "\n") else: Make.write(Assemble + " " + a + " -o " + a + ".o\n") MakeVerbose.write("echo " + Assemble + " " + a + " -o " + a + ".o\n") MakeVerbose.write(Assemble + " " + a + " -o " + a + ".o\n") Makefile.write("$(Assemble) " + a + " -o " + a + ".o\n") if a.endswith(".h"): CopyFile(os.path.join(Root, dir, Config, a), BootDir) Makefile.write("cm3: $(Objects)\n\t$(Link) -o cm3 *.o\n") Make.write(Link + " -o cm3 *.o\n") MakeVerbose.write("echo " + Link + " -o cm3 *.o\n") MakeVerbose.write(Link + " -o cm3 *.o\n") Make.close() Makefile.close() MakeVerbose.close() os.chdir("/") _MakeArchive(BootDir[1:]) | 5ede0f80a9877b27e6fe73475bd5c595c157b542 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9328/5ede0f80a9877b27e6fe73475bd5c595c157b542/pylib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
26254,
13332,
225,
2552,
3998,
2042,
3998,
2042,
1011,
315,
300,
7137,
300,
10102,
6,
225,
4049,
273,
315,
21,
6,
225,
16143,
273,
7566,
75,
952,
300,
75,
334,
5113,
15,
315,
397,
1079... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
26254,
13332,
225,
2552,
3998,
2042,
3998,
2042,
1011,
315,
300,
7137,
300,
10102,
6,
225,
4049,
273,
315,
21,
6,
225,
16143,
273,
7566,
75,
952,
300,
75,
334,
5113,
15,
315,
397,
1079... |
(metric_size_lat, metric_size_long) = \ spheric_sizes(cleaned_data.get("lat_upper_left"), cleaned_data.get("lon_upper_left"), cleaned_data.get("lat_bottom_right"), cleaned_data.get("lon_bottom_right")) | lat_upper_left = cleaned_data.get("lat_upper_left") lon_upper_left = cleaned_data.get("lon_upper_left") lat_bottom_right = cleaned_data.get("lat_bottom_right") lon_bottom_right = cleaned_data.get("lon_bottom_right") boundingbox = OCMBoundingBox(lat_upper_left, lon_upper_left, lat_bottom_right, lon_bottom_right) (metric_size_lat, metric_size_long) = boundingbox.spheric_sizes() | def clean(self): cleaned_data = self.cleaned_data | 263315bc39302ced739157f21a344946c6532a05 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/14904/263315bc39302ced739157f21a344946c6532a05/views.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2721,
12,
2890,
4672,
9426,
67,
892,
273,
365,
18,
6200,
329,
67,
892,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2721,
12,
2890,
4672,
9426,
67,
892,
273,
365,
18,
6200,
329,
67,
892,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
break | end = self.chunkOffset else: end = m.end() | def charsUntil(self, characters, opposite = False): """ Returns a string of characters from the stream up to but not including any character in 'characters' or EOF. 'characters' must be a container that supports the 'in' method and iteration over its characters. """ | 3c5b658fb41a6a20d376345279662bb20fca9d69 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9368/3c5b658fb41a6a20d376345279662bb20fca9d69/inputstream.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5230,
9716,
12,
2890,
16,
3949,
16,
20759,
273,
1083,
4672,
3536,
2860,
279,
533,
434,
3949,
628,
326,
1407,
731,
358,
1496,
486,
6508,
1281,
3351,
316,
296,
20558,
11,
578,
6431,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5230,
9716,
12,
2890,
16,
3949,
16,
20759,
273,
1083,
4672,
3536,
2860,
279,
533,
434,
3949,
628,
326,
1407,
731,
358,
1496,
486,
6508,
1281,
3351,
316,
296,
20558,
11,
578,
6431,
18,
... |
self.req.connect('task-added',self.update_tag_task_added) self.req.connect('task-deleted',self.update_tag_task_deleted) | self.req.connect('task-added',self.on_task_added) self.req.connect('task-deleted',self.on_task_deleted) self.req.connect('task-modified',self.on_task_modified) | def __init__(self, requester): gtk.GenericTreeModel.__init__(self) self.req = requester self.tree = self.req.get_tag_tree() self.workview = False | 2f000587cf1f61e4600cf231c42fe3ceba3cd9bb /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8234/2f000587cf1f61e4600cf231c42fe3ceba3cd9bb/tagtree.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
19961,
4672,
22718,
18,
7014,
2471,
1488,
16186,
2738,
972,
12,
2890,
13,
365,
18,
3658,
225,
273,
19961,
365,
18,
3413,
273,
365,
18,
3658,
18,
588,
67,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
19961,
4672,
22718,
18,
7014,
2471,
1488,
16186,
2738,
972,
12,
2890,
13,
365,
18,
3658,
225,
273,
19961,
365,
18,
3413,
273,
365,
18,
3658,
18,
588,
67,... |
for p in zdc: | for p in cache: | def uncache_zipdir(path): """Ensure that the zip directory cache doesn't have stale info for path""" from zipimport import _zip_directory_cache as zdc if path in zdc: del zdc[path] else: path = normalize_path(path) for p in zdc: if normalize_path(p)==path: del zdc[p] return | 053705f5f95ae3b44a685b1d4d9551e58df39883 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8186/053705f5f95ae3b44a685b1d4d9551e58df39883/easy_install.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6301,
807,
67,
4450,
1214,
12,
803,
4672,
3536,
12512,
716,
326,
3144,
1867,
1247,
3302,
1404,
1240,
14067,
1123,
364,
589,
8395,
628,
3144,
5666,
1930,
389,
4450,
67,
5149,
67,
2493,
48... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6301,
807,
67,
4450,
1214,
12,
803,
4672,
3536,
12512,
716,
326,
3144,
1867,
1247,
3302,
1404,
1240,
14067,
1123,
364,
589,
8395,
628,
3144,
5666,
1930,
389,
4450,
67,
5149,
67,
2493,
48... |
" if (!${name}.IsValid()) ${failBlock}\n", | " if (!${name}.IsValid())\n" " return JS_FALSE;\n", | def replacement(match): return vals[match.group(1)] | 85ee3e89273f0ac0b9295128759b8ada04a837bb /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/11102/85ee3e89273f0ac0b9295128759b8ada04a837bb/qsgen.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6060,
12,
1916,
4672,
327,
5773,
63,
1916,
18,
1655,
12,
21,
25887,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6060,
12,
1916,
4672,
327,
5773,
63,
1916,
18,
1655,
12,
21,
25887,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
try: os.rename(self.baseFilename, dfn) except (KeyboardInterrupt, SystemExit): raise except: self.handleError(record) | os.rename(self.baseFilename, dfn) | def doRollover(self): """ do a rollover; in this case, a date/time stamp is appended to the filename when the rollover happens. However, you want the file to be named for the start of the interval, not the current time. If there is a backup count, then we have to get a list of matching filenames, sort them and remove the one with the oldest suffix. """ self.stream.close() # get the time that this sequence started at and make it a TimeTuple t = self.rolloverAt - self.interval timeTuple = time.localtime(t) dfn = self.baseFilename + "." + time.strftime(self.suffix, timeTuple) if os.path.exists(dfn): os.remove(dfn) try: os.rename(self.baseFilename, dfn) except (KeyboardInterrupt, SystemExit): raise except: self.handleError(record) if self.backupCount > 0: # find the oldest log file and delete it s = glob.glob(self.baseFilename + ".20*") if len(s) > self.backupCount: s.sort() os.remove(s[0]) #print "%s -> %s" % (self.baseFilename, dfn) if self.encoding: self.stream = codecs.open(self.baseFilename, 'w', self.encoding) else: self.stream = open(self.baseFilename, 'w') self.rolloverAt = self.rolloverAt + self.interval | 6dd59f1632fd174ffdd255acca4b596a69e1a5b4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/6dd59f1632fd174ffdd255acca4b596a69e1a5b4/handlers.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
4984,
21896,
12,
2890,
4672,
3536,
741,
279,
10643,
21896,
31,
316,
333,
648,
16,
279,
1509,
19,
957,
14429,
353,
12317,
358,
326,
1544,
1347,
326,
10643,
21896,
10555,
18,
225,
107... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
4984,
21896,
12,
2890,
4672,
3536,
741,
279,
10643,
21896,
31,
316,
333,
648,
16,
279,
1509,
19,
957,
14429,
353,
12317,
358,
326,
1544,
1347,
326,
10643,
21896,
10555,
18,
225,
107... |
r"(https?://\S+)/show_bug.cgi\?id=(?P<bug>\w+)" | r"(?P<url>https?://\S+/)show_bug.cgi\?id=(?P<bug>\w+)" | def snarfBugUrl(self, irc, msg, match): r"(https?://\S+)/show_bug.cgi\?id=(?P<bug>\w+)" channel = msg.args[0] if (not self.registryValue('bugSnarfer', channel)): return | 964156d6c97e6d66fec03547486486fff0cf4fea /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3483/964156d6c97e6d66fec03547486486fff0cf4fea/plugin.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4556,
297,
74,
19865,
1489,
12,
2890,
16,
277,
1310,
16,
1234,
16,
845,
4672,
436,
6,
3680,
52,
32,
718,
34,
4528,
35,
2207,
64,
55,
31298,
13,
4500,
67,
925,
18,
19062,
64,
35,
35... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4556,
297,
74,
19865,
1489,
12,
2890,
16,
277,
1310,
16,
1234,
16,
845,
4672,
436,
6,
3680,
52,
32,
718,
34,
4528,
35,
2207,
64,
55,
31298,
13,
4500,
67,
925,
18,
19062,
64,
35,
35... |
(AUTOTEST_SVN, autodir) | (AUTOTEST_SVN, autodir)) | def install(self, host = None): """ Install autotest. If get() was not called previously, an attempt will be made to install from the autotest svn repository. Args: host: a Host instance on which autotest will be installed Raises: AutoservError: if a tarball was not specified and the target host does not have svn installed in its path TODO(poirier): check dependencies autotest needs: bzcat liboptdev (oprofile) binutils-dev (oprofile) make psutils (netperf) """ if not host: host = self.host if not self.got: self.get() host.ensure_up() host.setup() print "Installing autotest on %s" % host.hostname | 814312f5c9754734ae035a370240a33d06d034c6 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12268/814312f5c9754734ae035a370240a33d06d034c6/autotest.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3799,
12,
2890,
16,
1479,
273,
599,
4672,
3536,
10284,
2059,
352,
395,
18,
225,
971,
336,
1435,
1703,
486,
2566,
7243,
16,
392,
4395,
903,
506,
7165,
358,
3799,
628,
326,
2059,
352,
39... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3799,
12,
2890,
16,
1479,
273,
599,
4672,
3536,
10284,
2059,
352,
395,
18,
225,
971,
336,
1435,
1703,
486,
2566,
7243,
16,
392,
4395,
903,
506,
7165,
358,
3799,
628,
326,
2059,
352,
39... |
dat = re.split(' ',line) dat[1] = dat[1].replace("\n",'') if len(dat[0]) != 32: | md5 = line[:32] mpath = line[34:] mpath = mpath.replace("\n",'') if len(md5) != 32: | def manifest(self,path): from os.path import join, getsize cfiles = 0 if os.path.isdir(path): mname = '%s.manifest' % time.strftime('%d-%B-%Y',time.gmtime()) of = file(mname, 'w+') for root, dirs, files in os.walk(path): self.progress('Please wait, getting initial file count (%s) ...' % (cfiles)) #get count first loop for fname in files: cfiles += 1 print q = 'There are currently %s files in %s, do you want to proceed with the manifest?:' % (cfiles,path) a = raw_input(q) while a not in ('y','n'): q = 'Invalid response (y/n):' a = raw_input(q) if a == 'n': sys.exit(0) else: print tfiles = cfiles cfiles = 0 ltime = 0 lfiles = 0 filessec = 0 for root, dirs, files in os.walk(path): for fname in files: ctime = time.time() if (ctime - ltime) >= 1: if ltime > 0: filessec = round((cfiles - lfiles) / (ctime - ltime),2) ltime = ctime lfiles = cfiles fpath = join(root,fname) hash = self._checksum(fpath)['md5'] of.write("%s %s\n" % (hash, fpath)) cfiles += 1 fper = round(((1.00*cfiles)/(1.00*tfiles))*100.00,2) self.progress('Added %s/%s Files to manifest (%s%%) (%s/s)' % (cfiles,tfiles,fper,filessec)) print elif os.path.isfile(path): mcount = 0 #get manifest linecount for line in open(path,'r'): mcount += 1 self.progress('Please wait getting manifest count (%s)' % (mcount)) print print 'Manifest count complete' if mcount == 0: self.error('Counted 0 lines in manifest ... aborting') vcount = 0 #verify manifest for line in open(path,'r'): vcount += 1 vper = round(((1.00*vcount)/(1.00*mcount))*100.00,2) self.progress('Please wait verifying manifest (%s%%)' % (vper)) dat = re.split(' ',line) dat[1] = dat[1].replace("\n",'') if len(dat[0]) != 32: self.error('Manifest Verification error line %s md5 is invalid' % vcount,False) elif not os.path.isfile(dat[1]): self.error('Manifest Verification error line %s path is invalid (file may be missing)' % vcount,False) print print 'Manifest verification complete' vcount = 0 fcount = 0 pcount = 0 failed = [] filessec = 0 lfiles = 0 ltime = 0 lstr = 0 for line in open(path,'r'): ctime = time.time() if (ctime - ltime) >= 1: if ltime > 0: filessec = round((vcount - lfiles) / (ctime - ltime),2) ltime = ctime lfiles = vcount dat = re.split(' ',line) dat[1] = dat[1].replace("\n",'') if self._checksum(dat[1])['md5'] == dat[0]: pcount += 1 else: fcount += 1 failed.append(dat[1]) vcount += 1 vper = round(((1.00*vcount)/(1.00*mcount))*100.00,2) fper = round(((1.00*fcount)/(1.00*mcount))*100.00,2) pper = round(((1.00*pcount)/(1.00*mcount))*100.00,2) bar = self.progressbar(vper, 50) #[==================================================] Pass (00%) Fail(00%) str = 'Verification in progress: %s - %s%% Pass(%s%%) Fail(%s%%) %s/s' % (bar,vper,pper,fper,filessec) self.progress(str) print if fcount > 0: print '--- START FAILED LIST ---' for f in failed: print f print '--- END FAIL LIST ---' | 39593460b2989cc69762f491b9c2bac7796d818b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/14038/39593460b2989cc69762f491b9c2bac7796d818b/sysadmin.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5643,
12,
2890,
16,
803,
4672,
628,
1140,
18,
803,
1930,
1233,
16,
30208,
276,
2354,
273,
374,
225,
309,
1140,
18,
803,
18,
291,
1214,
12,
803,
4672,
312,
529,
273,
1995,
87,
18,
143... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5643,
12,
2890,
16,
803,
4672,
628,
1140,
18,
803,
1930,
1233,
16,
30208,
276,
2354,
273,
374,
225,
309,
1140,
18,
803,
18,
291,
1214,
12,
803,
4672,
312,
529,
273,
1995,
87,
18,
143... |
logging.info('grand total: %s added, %s updated.', stats['added'], stats['updated']) | logging.info('grand total: %s added, %s updated in %s models.', stats['added'], stats['updated'], stats['models']) | def safe_fetch_thread(*args): try: fetch_thread(*args) except: traceback.print_exc() | c988a7136c588ec46c99f57110ee9bd002f11810 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7737/c988a7136c588ec46c99f57110ee9bd002f11810/dsync.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4183,
67,
5754,
67,
5930,
30857,
1968,
4672,
775,
30,
2158,
67,
5930,
30857,
1968,
13,
1335,
30,
10820,
18,
1188,
67,
10075,
1435,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4183,
67,
5754,
67,
5930,
30857,
1968,
4672,
775,
30,
2158,
67,
5930,
30857,
1968,
13,
1335,
30,
10820,
18,
1188,
67,
10075,
1435,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
as with the startElement event.""" | as with the startElementNS event.""" | def endElement(self, name ): """Signals the end of an element. | 52c4e834e77743da5017f08efdb43e047504d2bc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/52c4e834e77743da5017f08efdb43e047504d2bc/handler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
14840,
12,
2890,
16,
508,
262,
30,
3536,
23346,
326,
679,
434,
392,
930,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
14840,
12,
2890,
16,
508,
262,
30,
3536,
23346,
326,
679,
434,
392,
930,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
if self.TLS.starttls<>'proceed': self.disconnected() | if self.TLS.starttls<>'proceed': self.disconnected() | def connect(self,proxy=None): if proxy: transports.HTTPPROXYsocket(proxy).PlugIn(self) else: transports.TCPsocket().PlugIn(self) dispatcher.Dispatcher().PlugIn(m) self.send_header() transports.TLS().PlugIn(self) while self.Process() and not self.Dispatcher.Stream._document_attrs: pass if self.Dispatcher.Stream._document_attrs.has_key('version') and self.Dispatcher.Stream._document_attrs['version']=='1.0': while self.Process() and not self.TLS.starttls: pass if self.TLS.starttls<>'proceed': self.disconnected() | af745a2ea2f5d240be56cd08d13ad776084219b6 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7194/af745a2ea2f5d240be56cd08d13ad776084219b6/client.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3077,
12,
2890,
16,
5656,
33,
7036,
4672,
309,
2889,
30,
28567,
18,
3693,
16085,
7814,
12,
5656,
2934,
1749,
637,
382,
12,
2890,
13,
469,
30,
28567,
18,
13891,
7814,
7675,
1749,
637,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3077,
12,
2890,
16,
5656,
33,
7036,
4672,
309,
2889,
30,
28567,
18,
3693,
16085,
7814,
12,
5656,
2934,
1749,
637,
382,
12,
2890,
13,
469,
30,
28567,
18,
13891,
7814,
7675,
1749,
637,
3... |
cache.writemulti(cacheId, deps) | classInfo[cacheId] = deps self._writeClassCache(classInfo) | def transitiveDepsAreFresh(depsStruct, cacheModTime): if cacheModTime is None: # TODO: this can currently only occur with a Cache.memcache result return False for dep in depsStruct["load"]: if dep.requestor != self.id: # this was included through a recursive traversal if dep.name in self._classesObj: classObj = self._classesObj[dep.name] if cacheModTime < classObj.m_time(): console.debug("Invalidating dep cache for %s, as %s is newer" % (self.id, classObj.id)) return False return True | 354224088ca9534ff655eb7deb172e17d995823e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5718/354224088ca9534ff655eb7deb172e17d995823e/Class.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
30442,
14430,
4704,
42,
1955,
12,
14877,
3823,
16,
1247,
1739,
950,
4672,
309,
1247,
1739,
950,
353,
599,
30,
225,
468,
2660,
30,
333,
848,
4551,
1338,
3334,
598,
279,
4379,
18,
3917,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
30442,
14430,
4704,
42,
1955,
12,
14877,
3823,
16,
1247,
1739,
950,
4672,
309,
1247,
1739,
950,
353,
599,
30,
225,
468,
2660,
30,
333,
848,
4551,
1338,
3334,
598,
279,
4379,
18,
3917,
... |
field = utils.indent("copied_fields['%s'],\n" % name, indent_level) | field = " copied_fields['%s'],\n\n" % name | def getFieldSpecFromAttribute(self, attr, classelement, indent_level=0): """Gets the schema field code.""" | ddf22e83eb1c201b87f839faf3f90bdd0a1c19aa /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/11941/ddf22e83eb1c201b87f839faf3f90bdd0a1c19aa/ArchetypesGenerator.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5031,
1990,
1265,
1499,
12,
2890,
16,
1604,
16,
667,
2956,
16,
3504,
67,
2815,
33,
20,
4672,
3536,
3002,
326,
1963,
652,
981,
12123,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5031,
1990,
1265,
1499,
12,
2890,
16,
1604,
16,
667,
2956,
16,
3504,
67,
2815,
33,
20,
4672,
3536,
3002,
326,
1963,
652,
981,
12123,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
sorted(list(requiredSet-providedSet))) | sorted(list(requiredSet-providedSet)), pathString) | def _createELFDepSet(self, m, elfinfo, recipe=None, basedir=None, soname=None, soflags=None, libPathMap={}, getRPATH=None): """ Add dependencies from ELF information. | 00a2ea0404d57dd20caeb2a27e6d0e8893314c8e /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8747/00a2ea0404d57dd20caeb2a27e6d0e8893314c8e/packagepolicy.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2640,
2247,
42,
16316,
694,
12,
2890,
16,
312,
16,
415,
926,
617,
16,
16100,
33,
7036,
16,
15573,
33,
7036,
16,
18882,
339,
33,
7036,
16,
272,
792,
80,
1341,
33,
7036,
16,
2561,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2640,
2247,
42,
16316,
694,
12,
2890,
16,
312,
16,
415,
926,
617,
16,
16100,
33,
7036,
16,
15573,
33,
7036,
16,
18882,
339,
33,
7036,
16,
272,
792,
80,
1341,
33,
7036,
16,
2561,... |
logging.error("Error getting sensor thresholds: " + str(rv)) | logging.error("Error getting sensor event enables: " + str(rv)) | def sensor_cb(self, sensor): if (self.setting): self.err = sensor.set_thresholds(self.threshold_str, self) else: rv = sensor.get_thresholds(self) if (rv != 0): logging.error("Error getting sensor thresholds: " + str(rv)) self.Destroy() return self.setting = True | 0e8f7115de513b9ddef5db6b3913a6d5b8f57658 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3867/0e8f7115de513b9ddef5db6b3913a6d5b8f57658/_sensor.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10583,
67,
7358,
12,
2890,
16,
10583,
4672,
309,
261,
2890,
18,
8920,
4672,
365,
18,
370,
273,
10583,
18,
542,
67,
8699,
87,
12,
2890,
18,
8699,
67,
701,
16,
365,
13,
469,
30,
5633,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10583,
67,
7358,
12,
2890,
16,
10583,
4672,
309,
261,
2890,
18,
8920,
4672,
365,
18,
370,
273,
10583,
18,
542,
67,
8699,
87,
12,
2890,
18,
8699,
67,
701,
16,
365,
13,
469,
30,
5633,
... |
if request.purchase_line.purchase.state == 'done': | elif request.purchase_line.purchase.state == 'done': | def get_state(self, cursor, user, ids, name, args, context=None): res = {}.fromkeys(ids, 'draft') for request in self.browse(cursor, user, ids, context=context): if request.purchase_line: if request.purchase_line.purchase.state == 'cancel': res[request.id] = 'cancel' if request.purchase_line.purchase.state == 'done': res[request.id] = 'done' else: res[request.id] = 'running' return res | 01723a909ce8bc238e7346d90061a2f3a543290a /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9299/01723a909ce8bc238e7346d90061a2f3a543290a/purchase_request.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
2019,
12,
2890,
16,
3347,
16,
729,
16,
3258,
16,
508,
16,
833,
16,
819,
33,
7036,
4672,
400,
273,
20151,
2080,
2452,
12,
2232,
16,
296,
17153,
6134,
364,
590,
316,
365,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
2019,
12,
2890,
16,
3347,
16,
729,
16,
3258,
16,
508,
16,
833,
16,
819,
33,
7036,
4672,
400,
273,
20151,
2080,
2452,
12,
2232,
16,
296,
17153,
6134,
364,
590,
316,
365,
18,
... |
setattr(values, dest, 1) | setattr(values, dest, True) | def take_action (self, action, dest, opt, value, values, parser): if action == "store": setattr(values, dest, value) elif action == "store_const": setattr(values, dest, self.const) elif action == "store_true": setattr(values, dest, 1) elif action == "store_false": setattr(values, dest, 0) elif action == "append": values.ensure_value(dest, []).append(value) elif action == "count": setattr(values, dest, values.ensure_value(dest, 0) + 1) elif action == "callback": args = self.callback_args or () kwargs = self.callback_kwargs or {} self.callback(self, opt, value, parser, *args, **kwargs) elif action == "help": parser.print_help() sys.exit(0) elif action == "version": parser.print_version() sys.exit(0) else: raise RuntimeError, "unknown action %r" % self.action | 05fbfbcaa14cc9093790f5eb3ae75558f3fdb250 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/05fbfbcaa14cc9093790f5eb3ae75558f3fdb250/optparse.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4862,
67,
1128,
261,
2890,
16,
1301,
16,
1570,
16,
2153,
16,
460,
16,
924,
16,
2082,
4672,
309,
1301,
422,
315,
2233,
6877,
9241,
12,
2372,
16,
1570,
16,
460,
13,
1327,
1301,
422,
31... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4862,
67,
1128,
261,
2890,
16,
1301,
16,
1570,
16,
2153,
16,
460,
16,
924,
16,
2082,
4672,
309,
1301,
422,
315,
2233,
6877,
9241,
12,
2372,
16,
1570,
16,
460,
13,
1327,
1301,
422,
31... |
return jobDB.getJobsAttributes( [jobId], self.summary ) | return jobDB.getAllJobAttributes( jobId ) | def export_getJobSummary(self, jobId): return jobDB.getJobsAttributes( [jobId], self.summary ) | 606e8d81395beaf79a4263a73f5da606576cb215 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12864/606e8d81395beaf79a4263a73f5da606576cb215/JobMonitoringHandler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3359,
67,
588,
2278,
4733,
12,
2890,
16,
13845,
4672,
327,
1719,
2290,
18,
588,
7276,
2498,
12,
306,
4688,
548,
6487,
365,
18,
7687,
262,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3359,
67,
588,
2278,
4733,
12,
2890,
16,
13845,
4672,
327,
1719,
2290,
18,
588,
7276,
2498,
12,
306,
4688,
548,
6487,
365,
18,
7687,
262,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
if isinstance(self.dbpath, unicode): | if isinstance(self.dbpath, unicode) and not iswindows: | def __init__(self, library_path, row_factory=False): self.field_metadata = FieldMetadata() if not os.path.exists(library_path): os.makedirs(library_path) self.listeners = set([]) self.library_path = os.path.abspath(library_path) self.row_factory = row_factory self.dbpath = os.path.join(library_path, 'metadata.db') self.dbpath = os.environ.get('CALIBRE_OVERRIDE_DATABASE_PATH', self.dbpath) if isinstance(self.dbpath, unicode): self.dbpath = self.dbpath.encode(filesystem_encoding) | dbcda9a0d8fe5d10931187e1a2c9a4f246ce85d7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9125/dbcda9a0d8fe5d10931187e1a2c9a4f246ce85d7/database2.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
5313,
67,
803,
16,
1027,
67,
6848,
33,
8381,
4672,
365,
18,
1518,
67,
4165,
273,
2286,
2277,
1435,
309,
486,
1140,
18,
803,
18,
1808,
12,
12083,
67,
80... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
5313,
67,
803,
16,
1027,
67,
6848,
33,
8381,
4672,
365,
18,
1518,
67,
4165,
273,
2286,
2277,
1435,
309,
486,
1140,
18,
803,
18,
1808,
12,
12083,
67,
80... |
return False, u'Socket Error: %s' % arg[1] | return False, u'Socket Error: %s' % repr(arg[1]) | def check(self, useHEAD = True): """ Returns True and the server status message if the page is alive. Otherwise returns false """ try: wasRedirected = self.resolveRedirect(useHEAD = useHEAD) except UnicodeError, arg: return False, u'Encoding Error: %s (%s)' % (arg.__class__.__name__, unicode(arg)) except httplib.error, arg: return False, u'HTTP Error: %s (%s)' % (arg.__class__.__name__, arg.line) except socket.error, arg: # TODO: decode arg[1]. On Linux, it's encoded in UTF-8. # How is it encoded in Windows? Or can we somehow just # get the English message? return False, u'Socket Error: %s' % arg[1] #except UnicodeEncodeError, arg: # return False, u'Non-ASCII Characters in URL: %s' % arg if wasRedirected: if self.url in self.redirectChain: if useHEAD: # Some servers don't seem to handle HEAD requests properly, # which leads to a cyclic list of redirects. # We simply start from the beginning, but this time, # we don't use HEAD, but GET requests. redirChecker = LinkChecker(self.redirectChain[0], serverEncoding = self.serverEncoding) return redirChecker.check(useHEAD = False) else: urlList = ['[%s]' % url for url in self.redirectChain + [self.url]] return False, u'HTTP Redirect Loop: %s' % ' -> '.join(urlList) elif len(self.redirectChain) >= 19: if useHEAD: # Some servers don't seem to handle HEAD requests properly, # which leads to a long (or infinite) list of redirects. # We simply start from the beginning, but this time, # we don't use HEAD, but GET requests. redirChecker = LinkChecker(self.redirectChain[0], serverEncoding = self.serverEncoding) return redirChecker.check(useHEAD = False) else: urlList = ['[%s]' % url for url in self.redirectChain + [self.url]] return False, u'Long Chain of Redirects: %s' % ' -> '.join(urlList) else: redirChecker = LinkChecker(self.url, self.redirectChain, self.serverEncoding) return redirChecker.check(useHEAD = useHEAD) else: try: conn = self.getConnection() except httplib.error, arg: return False, u'HTTP Error: %s (%s)' % (arg.__class__.__name__, arg.line) try: conn.request('GET', '%s%s' % (self.path, self.query), None, self.header) except socket.error, arg: return False, u'Socket Error: %s' % arg[1] #except UnicodeEncodeError, arg: # return False, u'Non-ASCII Characters in URL: %s' % arg try: response = conn.getresponse() except Exception, arg: return False, u'Error: %s' % arg # read the server's encoding, in case we need it later self.readEncodingFromResponse(response) # site down if the server status is between 400 and 499 siteDown = response.status in range(400, 500) return not siteDown, '%s %s' % (response.status, response.reason) | 67b543391a2cac4fd1a112b73d4b61220744895c /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/4404/67b543391a2cac4fd1a112b73d4b61220744895c/weblinkchecker.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
12,
2890,
16,
999,
12458,
273,
1053,
4672,
3536,
2860,
1053,
471,
326,
1438,
1267,
883,
309,
326,
1363,
353,
13714,
18,
5272,
1135,
629,
3536,
775,
30,
1703,
5961,
329,
273,
365,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
12,
2890,
16,
999,
12458,
273,
1053,
4672,
3536,
2860,
1053,
471,
326,
1438,
1267,
883,
309,
326,
1363,
353,
13714,
18,
5272,
1135,
629,
3536,
775,
30,
1703,
5961,
329,
273,
365,
... |
self.assertTrue('usage' in self.start_python('-h')) | self.assertTrue(b'usage' in self.start_python('-h')) | def test_usage(self): self.assertTrue('usage' in self.start_python('-h')) | 8fc4a4ab7dec085fade8bf8feaf4f3e94a8c26a8 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12029/8fc4a4ab7dec085fade8bf8feaf4f3e94a8c26a8/test_cmd_line.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
9167,
12,
2890,
4672,
365,
18,
11231,
5510,
2668,
9167,
11,
316,
365,
18,
1937,
67,
8103,
2668,
17,
76,
26112,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
9167,
12,
2890,
4672,
365,
18,
11231,
5510,
2668,
9167,
11,
316,
365,
18,
1937,
67,
8103,
2668,
17,
76,
26112,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
return util.checksignature(self.fn)(ui, *args, **opts) | if self.definition.startswith('!'): return self.fn(ui, *args, **opts) else: return util.checksignature(self.fn)(ui, *args, **opts) | def __call__(self, ui, *args, **opts): if self.shadows: ui.debug("alias '%s' shadows command\n" % self.name) | f50510bafb81cedbf955eefadbb1c0272afd52e9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11312/f50510bafb81cedbf955eefadbb1c0272afd52e9/dispatch.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
5915,
16,
380,
1968,
16,
2826,
4952,
4672,
309,
365,
18,
19119,
87,
30,
5915,
18,
4148,
2932,
4930,
1995,
87,
11,
10510,
87,
1296,
64,
82,
6,
738,
365,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
5915,
16,
380,
1968,
16,
2826,
4952,
4672,
309,
365,
18,
19119,
87,
30,
5915,
18,
4148,
2932,
4930,
1995,
87,
11,
10510,
87,
1296,
64,
82,
6,
738,
365,... |
result.append((sur.id, sur.title)) | res = True break | def _get_survey(self, cr, uid, context=None): """ Set the value In survey_id field. | dbe6aa5f692e4d95f56d3c9a402e8d3765d43ede /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/dbe6aa5f692e4d95f56d3c9a402e8d3765d43ede/survey_selection.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
21448,
12,
2890,
16,
4422,
16,
4555,
16,
819,
33,
7036,
4672,
3536,
1000,
326,
460,
657,
14902,
67,
350,
652,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
21448,
12,
2890,
16,
4422,
16,
4555,
16,
819,
33,
7036,
4672,
3536,
1000,
326,
460,
657,
14902,
67,
350,
652,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
self._conn = sqlite.connect(self.filename, autocommit=autoCommit) | self._conn = sqlite.connect(self.filename, **opts) | def __init__(self, filename, autoCommit=1, **kw): global sqlite if sqlite is None: import sqlite self.module = sqlite self.filename = filename # full path to sqlite-db-file if not autoCommit and not kw.has_key('pool'): # Pooling doesn't work with transactions... kw['pool'] = 0 # use only one connection for sqlite - supports multiple # cursors per connection self._conn = sqlite.connect(self.filename, autocommit=autoCommit) DBAPI.__init__(self, **kw) | 6d51286c7e1c86ee59fe2de2a53f19cb09f59ce3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8798/6d51286c7e1c86ee59fe2de2a53f19cb09f59ce3/sqliteconnection.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1544,
16,
3656,
5580,
33,
21,
16,
2826,
9987,
4672,
2552,
16184,
309,
16184,
353,
599,
30,
1930,
16184,
365,
18,
2978,
273,
16184,
365,
18,
3459,
273,
15... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1544,
16,
3656,
5580,
33,
21,
16,
2826,
9987,
4672,
2552,
16184,
309,
16184,
353,
599,
30,
1930,
16184,
365,
18,
2978,
273,
16184,
365,
18,
3459,
273,
15... |
return | return created_dirs | def mkpath (name, mode=0777, verbose=0, dry_run=0): """Create a directory and any missing ancestor directories. If the directory already exists, return silently. Raise DistutilsFileError if unable to create some directory along the way (eg. some sub-path exists, but is a file rather than a directory). If 'verbose' is true, print a one-line summary of each mkdir to stdout.""" global PATH_CREATED # XXX what's the better way to handle verbosity? print as we create # each directory in the path (the current behaviour), or only announce # the creation of the whole path? (quite easy to do the latter since # we're not using a recursive algorithm) name = os.path.normpath (name) if os.path.isdir (name) or name == '': return if PATH_CREATED.get (name): return (head, tail) = os.path.split (name) tails = [tail] # stack of lone dirs to create while head and tail and not os.path.isdir (head): #print "splitting '%s': " % head, (head, tail) = os.path.split (head) #print "to ('%s','%s')" % (head, tail) tails.insert (0, tail) # push next higher dir onto stack #print "stack of tails:", tails # now 'head' contains the deepest directory that already exists # (that is, the child of 'head' in 'name' is the highest directory # that does *not* exist) for d in tails: #print "head = %s, d = %s: " % (head, d), head = os.path.join (head, d) if PATH_CREATED.get (head): continue if verbose: print "creating", head if not dry_run: try: os.mkdir (head) except os.error, (errno, errstr): raise DistutilsFileError, \ "could not create '%s': %s" % (head, errstr) PATH_CREATED[head] = 1 | 2fe59fd6fd75b9ab3a88b32cd6af501955fac4e0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/2fe59fd6fd75b9ab3a88b32cd6af501955fac4e0/util.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5028,
803,
261,
529,
16,
1965,
33,
20,
14509,
16,
3988,
33,
20,
16,
10299,
67,
2681,
33,
20,
4672,
3536,
1684,
279,
1867,
471,
1281,
3315,
9731,
6402,
18,
225,
971,
326,
1867,
1818,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5028,
803,
261,
529,
16,
1965,
33,
20,
14509,
16,
3988,
33,
20,
16,
10299,
67,
2681,
33,
20,
4672,
3536,
1684,
279,
1867,
471,
1281,
3315,
9731,
6402,
18,
225,
971,
326,
1867,
1818,
... |
"""FileEntry - Entry field with button that invokes a FileSelectDialog | """FileEntry - Entry field with button that invokes a FileSelectDialog. The user can type in the filename manually. Alternatively, the user can press the button widget that sits next to the entry, which will bring up a file selection dialog. | def popdown(self): self.tk.call(self._w, 'popdown') | 01c920b9b56baabd2e1dd945affd1c17c1669b20 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/01c920b9b56baabd2e1dd945affd1c17c1669b20/Tix.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1843,
2378,
12,
2890,
4672,
365,
18,
16099,
18,
1991,
12,
2890,
6315,
91,
16,
296,
5120,
2378,
6134,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1843,
2378,
12,
2890,
4672,
365,
18,
16099,
18,
1991,
12,
2890,
6315,
91,
16,
296,
5120,
2378,
6134,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
print meta.rjust(20),':',value | print meta.rjust(20),':',value if setFlag and expandFlag: result = self.fc.getMetadataSet(value,expandFlag) if not result['OK']: print ("Error: %s" % result['Message']) return for m,v in result['Value'].items(): print " "*10,m.rjust(20),':',v | def do_get(self,args): """ Get metadata definitions and values | 6604aa2817ca2c6fd92be946a57731b479c950fa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/6604aa2817ca2c6fd92be946a57731b479c950fa/FileCatalogClientCLI.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
588,
12,
2890,
16,
1968,
4672,
3536,
968,
1982,
6377,
471,
924,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
588,
12,
2890,
16,
1968,
4672,
3536,
968,
1982,
6377,
471,
924,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
while 1: row = cursor.fetchone() if not row: break if len(info) == 0: | for idx, row in enum(cursor): if idx == 0: | def get_info(self, req, start, stop, maxrows, filters=AVAILABLE_FILTERS): perm_map = {'ticket': perm.TICKET_VIEW, 'changeset': perm.CHANGESET_VIEW, 'wiki': perm.WIKI_VIEW, 'milestone': perm.MILESTONE_VIEW} filters = list(filters) # copy list so we can make modifications for k,v in perm_map.items(): if k in filters and not self.perm.has_permission(v): filters.remove(k) if not filters: return [] | 937ad3787bccc993339741322e1694074671580c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2831/937ad3787bccc993339741322e1694074671580c/Timeline.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
1376,
12,
2890,
16,
1111,
16,
787,
16,
2132,
16,
943,
3870,
16,
3415,
33,
23222,
67,
11126,
55,
4672,
4641,
67,
1458,
273,
13666,
16282,
4278,
4641,
18,
56,
16656,
1584,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
1376,
12,
2890,
16,
1111,
16,
787,
16,
2132,
16,
943,
3870,
16,
3415,
33,
23222,
67,
11126,
55,
4672,
4641,
67,
1458,
273,
13666,
16282,
4278,
4641,
18,
56,
16656,
1584,
67,
... |
self.__conductor = 1 return self.__conductor | return 1 | def conductor(self): """ Computes and returns the conductor of this character. EXAMPLES:: sage: G.<a,b> = DirichletGroup(20) sage: a.conductor() 4 sage: b.conductor() 5 sage: (a*b).conductor() 20 """ try: return self.__conductor except AttributeError: pass if self.modulus() == 1 or self.is_trivial(): self.__conductor = 1 return self.__conductor F = arith.factor(self.modulus()) if len(F) > 1: self.__conductor = misc.mul([d.conductor() for d in self.decomposition()]) return self.__conductor p = F[0][0] # When p is odd, and x =/= 1, the conductor is the smallest p**r such that # Order(x) divides EulerPhi(p**r) = p**(r-1)*(p-1). # For a given r, whether or not the above divisiblity holds # depends only on the factor of p**(r-1) on the right hand side. # Since p-1 is coprime to p, this smallest r such that the # divisibility holds equals Valuation(Order(x),p)+1. self.__conductor = p**(arith.valuation(self.order(),p) + 1) if p == 2 and F[0][1] > 2 and self.values_on_gens()[1].multiplicative_order() != 1: self.__conductor *= 2; return self.__conductor | 59b9d951da50fd609601a8510ce35029ae856971 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/59b9d951da50fd609601a8510ce35029ae856971/dirichlet.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
356,
1828,
280,
12,
2890,
4672,
3536,
14169,
281,
471,
1135,
326,
356,
1828,
280,
434,
333,
3351,
18,
225,
5675,
8900,
11386,
2866,
225,
272,
410,
30,
611,
22782,
69,
16,
70,
34,
273,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
356,
1828,
280,
12,
2890,
4672,
3536,
14169,
281,
471,
1135,
326,
356,
1828,
280,
434,
333,
3351,
18,
225,
5675,
8900,
11386,
2866,
225,
272,
410,
30,
611,
22782,
69,
16,
70,
34,
273,
... |
self.assertRaises(ValueError, Boolean.encode, 'true') self.assertRaises(ValueError, Boolean.encode, 1) | self.assertRaises(TypeError, Boolean.encode, 'true') self.assertRaises(TypeError, Boolean.encode, 1) | def test_bad_encode(self): self.assertRaises(ValueError, Boolean.encode, 'true') self.assertRaises(ValueError, Boolean.encode, 1) | 36fa948462bb46ac3722360a5f2326878a3feb62 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10612/36fa948462bb46ac3722360a5f2326878a3feb62/test_utils.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
8759,
67,
3015,
12,
2890,
4672,
365,
18,
11231,
12649,
6141,
12,
23610,
16,
3411,
18,
3015,
16,
296,
3767,
6134,
365,
18,
11231,
12649,
6141,
12,
23610,
16,
3411,
18,
3015,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
8759,
67,
3015,
12,
2890,
4672,
365,
18,
11231,
12649,
6141,
12,
23610,
16,
3411,
18,
3015,
16,
296,
3767,
6134,
365,
18,
11231,
12649,
6141,
12,
23610,
16,
3411,
18,
3015,
1... |
while count < len(lst): | while count < len: | def waitall(lst, trap_errors=False): queue = coros.queue() results = [None] * len(lst) for (index, linkable) in enumerate(lst): linkable.link(decorate_send(queue, index)) count = 0 while count < len(lst): try: index, value = queue.wait() except Exception: if not trap_errors: raise else: results[index] = value count += 1 return results | 8714c4689c29f7e07806fd7b37495d8cc7dac46a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10078/8714c4689c29f7e07806fd7b37495d8cc7dac46a/proc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2529,
454,
12,
16923,
16,
23034,
67,
4324,
33,
8381,
4672,
2389,
273,
1858,
538,
18,
4000,
1435,
1686,
273,
306,
7036,
65,
380,
562,
12,
16923,
13,
364,
261,
1615,
16,
1692,
429,
13,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2529,
454,
12,
16923,
16,
23034,
67,
4324,
33,
8381,
4672,
2389,
273,
1858,
538,
18,
4000,
1435,
1686,
273,
306,
7036,
65,
380,
562,
12,
16923,
13,
364,
261,
1615,
16,
1692,
429,
13,
... |
manifest = [(htmlfile, 'text/x-oeb1-document'), | manifest = [(htmlfile, 'application/xhtml+xml'), | def create_opf(self, htmlfile, guide=None, root=None): mi = getattr(self.book_header.exth, 'mi', self.embedded_mi) if mi is None: mi = MetaInformation(self.book_header.title, [_('Unknown')]) opf = OPFCreator(os.path.dirname(htmlfile), mi) if hasattr(self.book_header.exth, 'cover_offset'): opf.cover = 'images/%05d.jpg' % (self.book_header.exth.cover_offset + 1) elif mi.cover is not None: opf.cover = mi.cover else: opf.cover = 'images/%05d.jpg' % 1 if not os.path.exists(os.path.join(os.path.dirname(htmlfile), * opf.cover.split('/'))): opf.cover = None | 0d1df2ad7b80b019ee684c43c5d26c514df81f12 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9125/0d1df2ad7b80b019ee684c43c5d26c514df81f12/reader.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
556,
74,
12,
2890,
16,
1729,
768,
16,
7343,
33,
7036,
16,
1365,
33,
7036,
4672,
12837,
273,
3869,
12,
2890,
18,
3618,
67,
3374,
18,
408,
76,
16,
296,
9197,
2187,
365,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
556,
74,
12,
2890,
16,
1729,
768,
16,
7343,
33,
7036,
16,
1365,
33,
7036,
4672,
12837,
273,
3869,
12,
2890,
18,
3618,
67,
3374,
18,
408,
76,
16,
296,
9197,
2187,
365,
18,
... |
gobutton = gtk.Button(label=_("Run!") | gobutton = gtk.Button(label=_("Run!")) | def __init__(self, handle): """Set up the Pippy activity.""" Activity.__init__(self, handle) self._logger = logging.getLogger('pippy-activity') | b3d6deea4a5123bcf5b21fd772ef3eee70bf3673 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7610/b3d6deea4a5123bcf5b21fd772ef3eee70bf3673/activity.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1640,
4672,
3536,
694,
731,
326,
29305,
2074,
5728,
12123,
9621,
16186,
2738,
972,
12,
2890,
16,
1640,
13,
365,
6315,
4901,
273,
2907,
18,
588,
3328,
2668,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1640,
4672,
3536,
694,
731,
326,
29305,
2074,
5728,
12123,
9621,
16186,
2738,
972,
12,
2890,
16,
1640,
13,
365,
6315,
4901,
273,
2907,
18,
588,
3328,
2668,... |
The p-adic valuation of the norm of self on the r-overconvergent region. | The `p`-adic valuation of the norm of self on the `r`-overconvergent region. | def r_ord(self, r): r""" The p-adic valuation of the norm of self on the r-overconvergent region. | 51aabb4f4c1d25de708da4cd9e8295aeb8a186e7 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/51aabb4f4c1d25de708da4cd9e8295aeb8a186e7/genus0.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
436,
67,
517,
12,
2890,
16,
436,
4672,
436,
8395,
1021,
293,
17,
20333,
21748,
367,
434,
326,
4651,
434,
365,
603,
326,
436,
17,
1643,
591,
502,
75,
319,
3020,
18,
2,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
436,
67,
517,
12,
2890,
16,
436,
4672,
436,
8395,
1021,
293,
17,
20333,
21748,
367,
434,
326,
4651,
434,
365,
603,
326,
436,
17,
1643,
591,
502,
75,
319,
3020,
18,
2,
-100,
-100,
-10... |
return reduce(operator.mul, range(1, n+1), 1) | return prod(range(1, n+1)) def permutations(iterable, r=None): pool = tuple(iterable) n = len(pool) r = n if r is None else r for indices in product(range(n), repeat=r): if len(set(indices)) == r: yield tuple(pool[i] for i in indices) | def fact(n): 'Factorial' return reduce(operator.mul, range(1, n+1), 1) | 607f70f7e750824edcf97db01d1950e1dd890f14 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8125/607f70f7e750824edcf97db01d1950e1dd890f14/test_itertools.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5410,
12,
82,
4672,
296,
6837,
649,
11,
327,
10791,
12,
3676,
12,
21,
16,
290,
15,
21,
3719,
225,
1652,
26468,
12,
15364,
16,
436,
33,
7036,
4672,
225,
2845,
273,
3193,
12,
15364,
13... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5410,
12,
82,
4672,
296,
6837,
649,
11,
327,
10791,
12,
3676,
12,
21,
16,
290,
15,
21,
3719,
225,
1652,
26468,
12,
15364,
16,
436,
33,
7036,
4672,
225,
2845,
273,
3193,
12,
15364,
13... |
'Read the files and paste their contents to %s.\n' | 'Read the files and paste their contents to LodgeIt pastebin.\n' | def main(): """Main script entry point.""" usage = ('Usage: %%prog [options] [FILE ...]\n\n' 'Read the files and paste their contents to %s.\n' 'If no file is given, read from standard input.\n' 'If multiple files are given, they are put into a single paste.' % SERVICE_URL) parser = OptionParser(usage=usage) settings = load_default_settings() parser.add_option('-v', '--version', action='store_true', help='Print script version') parser.add_option('-L', '--languages', action='store_true', default=False, help='Retrieve a list of supported languages') parser.add_option('-l', '--language', default=settings['language'], help='Used syntax highlighter for the file') parser.add_option('-e', '--encoding', default=settings['encoding'], help='Specify the encoding of a file (default is ' 'utf-8 or guessing if available)') parser.add_option('-b', '--open-browser', dest='open_browser', action='store_true', default=settings['open_browser'], help='Open the paste in a web browser') parser.add_option('-p', '--private', action='store_true', default=False, help='Paste as private') parser.add_option('--no-clipboard', dest='clipboard', action='store_false', default=settings['clipboard'], help="Don't copy the url into the clipboard") parser.add_option('--download', metavar='UID', help='Download a given paste') opts, args = parser.parse_args() # special modes of operation: # - paste script version if opts.version: print '%s: version %s' % (SCRIPT_NAME, VERSION) sys.exit() # - print list of languages elif opts.languages: print_languages() sys.exit() # - download Paste elif opts.download: download_paste(opts.download) sys.exit() # check language if given if opts.language and not language_exists(opts.language): fail('Language %s is not supported.' % opts.language, 3) # load file(s) try: data, language, filename, mimetype = compile_paste(args, opts.language) except Exception, err: fail('Error while reading the file(s): %s' % err, 2) if not data: fail('Aborted, no content to paste.', 4) # create paste code = make_utf8(data, opts.encoding) pid = create_paste(code, language, filename, mimetype, opts.private) url = '%sshow/%s/' % (SERVICE_URL, pid) print url if opts.open_browser: open_webbrowser(url) if opts.clipboard: copy_url(url) | b75206fd8de1f7f74e3150b949de916ec5b50ca3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11759/b75206fd8de1f7f74e3150b949de916ec5b50ca3/lodgeit.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
3536,
6376,
2728,
1241,
1634,
12123,
225,
4084,
273,
7707,
5357,
30,
26510,
14654,
306,
2116,
65,
306,
3776,
1372,
13944,
82,
64,
82,
11,
296,
1994,
326,
1390,
471,
19401,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
3536,
6376,
2728,
1241,
1634,
12123,
225,
4084,
273,
7707,
5357,
30,
26510,
14654,
306,
2116,
65,
306,
3776,
1372,
13944,
82,
64,
82,
11,
296,
1994,
326,
1390,
471,
19401,
3... |
EXAMPLES: sage: A2 = WeylCharacterRing(['A',2]) sage: chi = A2(1,0,0) sage: [chi^k for k in range(6)] [A2(0,0,0), A2(1,0,0), A2(1,1,0) + A2(2,0,0), A2(1,1,1) + 2*A2(2,1,0) + A2(3,0,0), 3*A2(2,1,1) + 2*A2(2,2,0) + 3*A2(3,1,0) + A2(4,0,0), 5*A2(2,2,1) + 6*A2(3,1,1) + 5*A2(3,2,0) + 4*A2(4,1,0) + A2(5,0,0)] | Returns self^n. | def __pow__(self, n): """ EXAMPLES: sage: A2 = WeylCharacterRing(['A',2]) sage: chi = A2(1,0,0) sage: [chi^k for k in range(6)] [A2(0,0,0), A2(1,0,0), A2(1,1,0) + A2(2,0,0), A2(1,1,1) + 2*A2(2,1,0) + A2(3,0,0), 3*A2(2,1,1) + 2*A2(2,2,0) + 3*A2(3,1,0) + A2(4,0,0), 5*A2(2,2,1) + 6*A2(3,1,1) + 5*A2(3,2,0) + 4*A2(4,1,0) + A2(5,0,0)] | da4d6149538661041003f964c0d06f212d52f0f5 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/da4d6149538661041003f964c0d06f212d52f0f5/weyl_characters.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
23509,
972,
12,
2890,
16,
290,
4672,
3536,
5675,
8900,
11386,
30,
272,
410,
30,
432,
22,
273,
678,
402,
80,
7069,
10369,
12,
3292,
37,
2187,
22,
5717,
272,
410,
30,
17198,
273,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
23509,
972,
12,
2890,
16,
290,
4672,
3536,
5675,
8900,
11386,
30,
272,
410,
30,
432,
22,
273,
678,
402,
80,
7069,
10369,
12,
3292,
37,
2187,
22,
5717,
272,
410,
30,
17198,
273,
... |
if self.align == 'c': gc.translate_ctm((gc.width()-scale_width)/2, (gc.height()+scale_height)/2) elif self.align == 'ur': gc.translate_ctm((gc.width()+scale_width)/2, gc.height() - scale_height/2) | if self.align == 'ur': gc.translate_ctm(origin_x + (plot_width-scale_width), origin_y + plot_height) | def overlay(self, other_component, gc, view_bounds=None, mode="normal"): """ Draws this component overlaid on another component. | 40adcfe1f6491cc7c671ce13983018b8831c3524 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/13167/40adcfe1f6491cc7c671ce13983018b8831c3524/status_overlay.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9218,
12,
2890,
16,
1308,
67,
4652,
16,
8859,
16,
1476,
67,
10576,
33,
7036,
16,
1965,
1546,
6130,
6,
4672,
3536,
10184,
87,
333,
1794,
1879,
11821,
350,
603,
4042,
1794,
18,
2,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9218,
12,
2890,
16,
1308,
67,
4652,
16,
8859,
16,
1476,
67,
10576,
33,
7036,
16,
1965,
1546,
6130,
6,
4672,
3536,
10184,
87,
333,
1794,
1879,
11821,
350,
603,
4042,
1794,
18,
2,
-100,
... |
if len( args ) != 2: | if len( args ) != 3: | def itemtimer( char, args ): if len( fishingItems ) < 1: socket.sysmessage( 'This script has not been configured correctly.' ) return 0 socket = char.socket socket.deltag( 'is_fishing' ) if len( args ) != 2: return # Is there any fish left there? (8x8 grid) resource = findResourceGem( args[0] ) if resource: amount = max( 0, int( resource.gettag( 'resourcecount' ) ) ) # We need this twice else: amount = random.randint( FISHING_FISH[0], FISHING_FISH[1] ) if resource and amount <= 0: socket.clilocmessage( 0x7ad80, "", 0x3b2, 3, char ) # The fish don't seem to be biting here. return # Fail if not char.checkskill( FISHING, 0, 1000 ): socket.clilocmessage( 0x7AD83, "", 0x3b2, 3, char ) return spawnmonster = 0 # Should i spawn a monster ? monster = '' # NPC id of what i should spawn near the player ( itemid, itemname ) = getFish( char.skill[ FISHING ], args[1] ) # Only add if there is something to add if itemid: if not resource: # Create a resource gem resource = wolfpack.additem( "1ea7" ) resource.name = 'Resource Item: fish' resource.settag( 'resourcecount', int( amount - 1 ) ) resource.settag( 'resource', 'fish' ) resource.visible = 0 # GM Visible only pos = args[ 0 ] resource.moveto( wolfpack.coord( int( floor( pos.x / 8 ) ) * 8, int( floor( pos.y / 8 ) ) * 8, int( pos.z - 5 ), pos.map ) ) resource.decay = 0 decaytime = random.randint( FISHING_REFILLTIME[0], FISHING_REFILLTIME[1] ) resource.addtimer( decaytime, 'skills.fishing.resourceDecayTimer', [], True ) resource.update() # Send to GMs else: # Use the old one resource.settag( 'resourcecount', resource.gettag( 'resourcecount' ) - 1 ) item = wolfpack.additem( itemid ) # Was there an error? if not item: socket.sysmessage( "Please report to a gamemaster that the item '%s' couldn't be found." % itemid ) # Otherwise try to stack it elif not wolfpack.utilities.tobackpack( item, char ): item.update() # Success! if not spawnmonster: socket.clilocmessage( 0xf61fc, "", 0x3b2, 3, None, str(itemname) ) # You pull out an item : else: socket.clilocmessage( 0xf61fd, "", 0x3b2, 3, None, str(itemname) ) # You pull out an item along with a monster : | 67990f35e8c14c12b60ab13d8d6aa2a84af6c4cd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2534/67990f35e8c14c12b60ab13d8d6aa2a84af6c4cd/fishing.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
761,
12542,
12,
1149,
16,
833,
262,
30,
309,
562,
12,
284,
1468,
310,
3126,
262,
411,
404,
30,
2987,
18,
9499,
2150,
12,
296,
2503,
2728,
711,
486,
2118,
4351,
8783,
1093,
262,
327,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
761,
12542,
12,
1149,
16,
833,
262,
30,
309,
562,
12,
284,
1468,
310,
3126,
262,
411,
404,
30,
2987,
18,
9499,
2150,
12,
296,
2503,
2728,
711,
486,
2118,
4351,
8783,
1093,
262,
327,
... |
if single_article_per_section: single_list = [] for author in self.authors: single_list.append(author[0]) if len(single_list) > 1: author_list = '%s -\n%s' % (single_list[0], single_list[-1]) | master_author_list = [] current_letter = self.authors[0][1][0] current_author_list = [] for author in self.authors: if author[1][0] != current_letter: author_list = " • ".join(current_author_list) if len(current_author_list) == self.descriptionClip: author_list += " …" author_list = self.formatNCXText(author_list) if self.verbose: print " adding '%s' to master_author_list" % current_letter master_author_list.append((author_list, current_letter)) current_letter = author[1][0] current_author_list = [author[0]] | def generateNCXByAuthor(self, tocTitle, single_article_per_section=True): | 073710f405f291eedc7992694a78db2be7983d10 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9125/073710f405f291eedc7992694a78db2be7983d10/catalog.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2103,
10346,
60,
858,
3594,
12,
2890,
16,
17919,
4247,
16,
2202,
67,
11480,
67,
457,
67,
3464,
33,
5510,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2103,
10346,
60,
858,
3594,
12,
2890,
16,
17919,
4247,
16,
2202,
67,
11480,
67,
457,
67,
3464,
33,
5510,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
return GenericBrowser(browser.split()) | browser = shlex.split(browser) if browser[-1] == '&': return BackgroundBrowser(browser[:-1]) else: return GenericBrowser(browser) | def get(using=None): """Return a browser launcher instance appropriate for the environment.""" if using is not None: alternatives = [using] else: alternatives = _tryorder for browser in alternatives: if '%s' in browser: # User gave us a command line, split it into name and args return GenericBrowser(browser.split()) else: # User gave us a browser name or path. try: command = _browsers[browser.lower()] except KeyError: command = _synthesize(browser) if command[1] is not None: return command[1] elif command[0] is not None: return command[0]() raise Error("could not locate runnable browser") | 972ca507aa9fe0aedc0d614b8837d11bf41f0f24 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8546/972ca507aa9fe0aedc0d614b8837d11bf41f0f24/webbrowser.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
12,
9940,
33,
7036,
4672,
3536,
990,
279,
4748,
26618,
791,
5505,
364,
326,
3330,
12123,
309,
1450,
353,
486,
599,
30,
20699,
273,
306,
9940,
65,
469,
30,
20699,
273,
389,
698,
10... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
12,
9940,
33,
7036,
4672,
3536,
990,
279,
4748,
26618,
791,
5505,
364,
326,
3330,
12123,
309,
1450,
353,
486,
599,
30,
20699,
273,
306,
9940,
65,
469,
30,
20699,
273,
389,
698,
10... |
file.Write(" Validate%s(this, arg_count%s);\n" % | file.Write(" Validate%s(this, immediate_data_size%s);\n" % | def WriteHandlerImplementation (self, func, file): """Overrriden from TypeHandler.""" file.Write(" uint32 client_id = c.client_id;\n") file.Write(" parse_error::ParseError result =\n") file.Write(" Validate%s(this, arg_count%s);\n" % (func.name, func.MakeOriginalArgString("", True))) file.Write(" if (result != parse_error::kParseNoError) {\n") file.Write(" return result;\n") file.Write(" }\n") file.Write(" %sHelper(%s);\n" % (func.name, func.MakeCmdArgString(""))) | 6b5cc7aeca6761c88c0f98b0e54990becbe32177 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9392/6b5cc7aeca6761c88c0f98b0e54990becbe32177/build_gles2_cmd_buffer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2598,
1503,
13621,
261,
2890,
16,
1326,
16,
585,
4672,
3536,
22042,
1691,
275,
628,
1412,
1503,
12123,
585,
18,
3067,
2932,
225,
2254,
1578,
1004,
67,
350,
273,
276,
18,
2625,
67,
350,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2598,
1503,
13621,
261,
2890,
16,
1326,
16,
585,
4672,
3536,
22042,
1691,
275,
628,
1412,
1503,
12123,
585,
18,
3067,
2932,
225,
2254,
1578,
1004,
67,
350,
273,
276,
18,
2625,
67,
350,
... |
self.CondPrint('All entries correct.') | self.CondPrint('verbose', 'All entries correct.') | def CondDisplayState(self, state, phase): self.CondPrint(state, 'Phase: %s' % phase) self.CondPrint(state, 'Correct entries:\t%d' % self.states.values().count(True)) self.CondPrint(state, 'Incorrect entries:\t%d' % self.states.values().count(False)) self.CondPrint(state, 'Total managed entries:\t%d' % len(self.states.values())) self.CondPrint(state, 'Unmanaged entries:\t%d' % len(self.pkgwork['remove'])) self.CondPrint(state, '') | 677ba9b6057e10f36a77177e426d4660bcb93359 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11867/677ba9b6057e10f36a77177e426d4660bcb93359/Toolset.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
735,
72,
4236,
1119,
12,
2890,
16,
919,
16,
6855,
4672,
365,
18,
12441,
5108,
12,
2019,
16,
296,
11406,
30,
738,
87,
11,
738,
6855,
13,
365,
18,
12441,
5108,
12,
2019,
16,
296,
16147... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
735,
72,
4236,
1119,
12,
2890,
16,
919,
16,
6855,
4672,
365,
18,
12441,
5108,
12,
2019,
16,
296,
11406,
30,
738,
87,
11,
738,
6855,
13,
365,
18,
12441,
5108,
12,
2019,
16,
296,
16147... |
memo = set() def flatten(dists): for dist in dists: if dist in memo: continue memo.add(dist) try: predecessors = working_set.resolve([dist.as_requirement()]) for predecessor in flatten(predecessors): yield predecessor yield dist except pkg_resources.DistributionNotFound, e: env.log.error('Skipping "%s" ("%s" not found)', dist, e) except pkg_resources.VersionConflict, e: env.log.error('Skipping "%s" (version conflict: "%s")', dist, e) | def _log_error(item, e): if isinstance(e, DistributionNotFound): env.log.error('Skipping "%s": ("%s" not found)', item, e) elif isinstance(e, VersionConflict): env.log.error('Skipping "%s": (version conflict "%s")', item, e) elif isinstance(e, UnknownExtra): env.log.error('Skipping "%s": (unknown extra "%s")', item, e) elif isinstance(e, ImportError): env.log.error('Skipping "%s": (can\'t import "%s")', item, e) else: env.log.error('Skipping "%s": (error "%s")', item, e) | def _load_eggs(env, search_path, auto_enable=None): working_set = pkg_resources.working_set for path in search_path: working_set.add_entry(path) | 4f97e682f87d34e59509f9877bd5ee56864dc858 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9317/4f97e682f87d34e59509f9877bd5ee56864dc858/loader.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
945,
67,
1332,
564,
12,
3074,
16,
1623,
67,
803,
16,
3656,
67,
7589,
33,
7036,
4672,
5960,
67,
542,
273,
3475,
67,
4683,
18,
20478,
67,
542,
364,
589,
316,
1623,
67,
803,
30,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
945,
67,
1332,
564,
12,
3074,
16,
1623,
67,
803,
16,
3656,
67,
7589,
33,
7036,
4672,
5960,
67,
542,
273,
3475,
67,
4683,
18,
20478,
67,
542,
364,
589,
316,
1623,
67,
803,
30,
... |
cursor.execute('SELECT values '\ 'FROM ir_model_data '\ 'WHERE model = %s '\ 'AND db_id in (' + ','.join('%s' for x in ids)+ ') ', [self._name]+ids) | cursor.execute('SELECT values ' \ 'FROM ir_model_data ' \ 'WHERE model = %s ' \ 'AND db_id in (' + ','.join('%s' for x in ids)+ ') ', [self._name] + ids) | def check_xml_record(self, cursor, user, ids, values, context=None): """ Check if a list of records and their corresponding fields are originating from xml data. This is used by write and delete functions: if the return value is True the records can be written/deleted, False otherwise. The default behaviour is to forbid all modification on records/fields originating from xml. Values is the dictionary of written values. If values is equal to None, no field by field check is performed, False is return has soon has one of the record comes from the xml. """ # Allow root user to update/delete if user == 0: return True cursor.execute('SELECT values '\ 'FROM ir_model_data '\ 'WHERE model = %s '\ 'AND db_id in (' + ','.join('%s' for x in ids)+ ') ', [self._name]+ids) if cursor.rowcount == 0: return True if values == None: return False for line in cursor.fetchall(): xml_values = eval(line[0]) for key, val in values.iteritems(): if key in xml_values and val != xml_values[key]: return False return True | 69bf9b728d03430a2d0ad6a2fdee54294ff1ef10 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9266/69bf9b728d03430a2d0ad6a2fdee54294ff1ef10/orm.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
2902,
67,
3366,
12,
2890,
16,
3347,
16,
729,
16,
3258,
16,
924,
16,
819,
33,
7036,
4672,
3536,
2073,
309,
279,
666,
434,
3853,
471,
3675,
4656,
1466,
854,
4026,
1776,
628,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
2902,
67,
3366,
12,
2890,
16,
3347,
16,
729,
16,
3258,
16,
924,
16,
819,
33,
7036,
4672,
3536,
2073,
309,
279,
666,
434,
3853,
471,
3675,
4656,
1466,
854,
4026,
1776,
628,
2... |
print 'Successfully logged in as', creds | def cbLoginSuccess(input): """User authenticated successfully. | f52c91b64c170484942fcadf870dc4af79989662 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6980/f52c91b64c170484942fcadf870dc4af79989662/websession.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2875,
5358,
4510,
12,
2630,
4672,
3536,
1299,
9370,
4985,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2875,
5358,
4510,
12,
2630,
4672,
3536,
1299,
9370,
4985,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... | |
self.valueOf_ = get_all_text_(node) if node.text is not None: obj_ = self.mixedclass_(MixedContainer.CategoryText, MixedContainer.TypeNone, '', node.text) self.content_.append(obj_) | def build(self, node): self.buildAttributes(node, node.attrib, []) self.valueOf_ = get_all_text_(node) if node.text is not None: obj_ = self.mixedclass_(MixedContainer.CategoryText, MixedContainer.TypeNone, '', node.text) self.content_.append(obj_) for child in node: nodeName_ = Tag_pattern_.match(child.tag).groups()[-1] self.buildChildren(child, nodeName_) | 9c12e50d449fa27d6f8f3415ece228ae97bb0266 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14016/9c12e50d449fa27d6f8f3415ece228ae97bb0266/_nexml.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1361,
12,
2890,
16,
756,
4672,
365,
18,
3510,
2498,
12,
2159,
16,
756,
18,
14588,
16,
5378,
13,
364,
1151,
316,
756,
30,
7553,
67,
273,
4034,
67,
4951,
27799,
1916,
12,
3624,
18,
269... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1361,
12,
2890,
16,
756,
4672,
365,
18,
3510,
2498,
12,
2159,
16,
756,
18,
14588,
16,
5378,
13,
364,
1151,
316,
756,
30,
7553,
67,
273,
4034,
67,
4951,
27799,
1916,
12,
3624,
18,
269... | |
self.add_library ( "python" + sys.version[0] + sys.version[2] ) | def __init__ (self, verbose=0, dry_run=0, force=0): | c9db91300cada056e565e829810030a89aaff9b4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/c9db91300cada056e565e829810030a89aaff9b4/msvccompiler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
261,
2890,
16,
3988,
33,
20,
16,
10299,
67,
2681,
33,
20,
16,
2944,
33,
20,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
261,
2890,
16,
3988,
33,
20,
16,
10299,
67,
2681,
33,
20,
16,
2944,
33,
20,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... | |
sb.appendData(m1_1) | sb.append(m1_1) | def test_getOneMessage_fragmented(self): """ Test the handling of fragmented frame data. """ m1_1 = 'CONNECT\nsession:207567f3-cce7-4a0a-930b-' m1_2 = '46fc394dd53d\n\n0123456789\x00\nSUBSCRIBE\nack:a' m1_3 = 'uto\ndestination:/queue/test\n\n\x00SE' m1_4 = 'ND\ndestination:/queue/test\n\n0123456789\x00' sb = StompFrameBuffer() sb.appendData(m1_1) assert sb.getOneMessage() is None sb.appendData(m1_2) f1 = sb.getOneMessage() assert f1.cmd == 'CONNECT' assert f1.body == '0123456789' assert sb.getOneMessage() is None sb.appendData(m1_3) f2 = sb.getOneMessage() assert f2.cmd == 'SUBSCRIBE' assert sb.getOneMessage() is None sb.appendData(m1_4) f3 = sb.getOneMessage() assert f3.cmd == 'SEND' assert f3.destination == '/queue/test' assert f3.body == '0123456789' | d449f708e10a4ccd6e53d9237356e9b70565b11d /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10258/d449f708e10a4ccd6e53d9237356e9b70565b11d/test_buffer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
588,
3335,
1079,
67,
11956,
329,
12,
2890,
4672,
3536,
7766,
326,
5057,
434,
5481,
329,
2623,
501,
18,
3536,
225,
312,
21,
67,
21,
225,
273,
296,
11032,
64,
2387,
893,
30,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
588,
3335,
1079,
67,
11956,
329,
12,
2890,
4672,
3536,
7766,
326,
5057,
434,
5481,
329,
2623,
501,
18,
3536,
225,
312,
21,
67,
21,
225,
273,
296,
11032,
64,
2387,
893,
30,
... |
self.scroll_to_cell (path=path, use_align=FALSE) | self.scroll_to_cell (path=path, use_align=False) | def selectFirstWritableApp (self): app = self.model.findFirstWritableApp() if app: path = self.model.getPathFromNode (app) self.get_selection().select_path (path) self.scroll_to_cell (path=path, use_align=FALSE) | ab8126074d7036807d1d842171687cc01617a181 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4580/ab8126074d7036807d1d842171687cc01617a181/driconf.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2027,
3759,
12061,
3371,
261,
2890,
4672,
595,
273,
365,
18,
2284,
18,
4720,
3759,
12061,
3371,
1435,
309,
595,
30,
589,
273,
365,
18,
2284,
18,
588,
743,
28570,
261,
2910,
13,
365,
18... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2027,
3759,
12061,
3371,
261,
2890,
4672,
595,
273,
365,
18,
2284,
18,
4720,
3759,
12061,
3371,
1435,
309,
595,
30,
589,
273,
365,
18,
2284,
18,
588,
743,
28570,
261,
2910,
13,
365,
18... |
try: dict(senseless={}) except TypeError: pass else: raise TestFailed("no TypeError from dict(senseless={})") | def dict_constructor(): if verbose: print "Testing dict constructor ..." d = dict() vereq(d, {}) d = dict({}) vereq(d, {}) d = dict(items={}) vereq(d, {}) d = dict({1: 2, 'a': 'b'}) vereq(d, {1: 2, 'a': 'b'}) vereq(d, dict(d.items())) vereq(d, dict(items=d.iteritems())) for badarg in 0, 0L, 0j, "0", [0], (0,): try: dict(badarg) except TypeError: pass except ValueError: if badarg == "0": # It's a sequence, and its elements are also sequences (gotta # love strings <wink>), but they aren't of length 2, so this # one seemed better as a ValueError than a TypeError. pass else: raise TestFailed("no TypeError from dict(%r)" % badarg) else: raise TestFailed("no TypeError from dict(%r)" % badarg) try: dict(senseless={}) except TypeError: pass else: raise TestFailed("no TypeError from dict(senseless={})") try: dict({}, {}) except TypeError: pass else: raise TestFailed("no TypeError from dict({}, {})") class Mapping: # Lacks a .keys() method; will be added later. dict = {1:2, 3:4, 'a':1j} try: dict(Mapping()) except TypeError: pass else: raise TestFailed("no TypeError from dict(incomplete mapping)") Mapping.keys = lambda self: self.dict.keys() Mapping.__getitem__ = lambda self, i: self.dict[i] d = dict(items=Mapping()) vereq(d, Mapping.dict) # Init from sequence of iterable objects, each producing a 2-sequence. class AddressBookEntry: def __init__(self, first, last): self.first = first self.last = last def __iter__(self): return iter([self.first, self.last]) d = dict([AddressBookEntry('Tim', 'Warsaw'), AddressBookEntry('Barry', 'Peters'), AddressBookEntry('Tim', 'Peters'), AddressBookEntry('Barry', 'Warsaw')]) vereq(d, {'Barry': 'Warsaw', 'Tim': 'Peters'}) d = dict(zip(range(4), range(1, 5))) vereq(d, dict([(i, i+1) for i in range(4)])) # Bad sequence lengths. for bad in [('tooshort',)], [('too', 'long', 'by 1')]: try: dict(bad) except ValueError: pass else: raise TestFailed("no ValueError from dict(%r)" % bad) | d65fb152f1cb970df809fba08a4132871bd4aace /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/d65fb152f1cb970df809fba08a4132871bd4aace/test_descr.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2065,
67,
12316,
13332,
309,
3988,
30,
1172,
315,
22218,
2065,
3885,
18483,
302,
273,
2065,
1435,
331,
822,
85,
12,
72,
16,
2618,
13,
302,
273,
2065,
23506,
13,
331,
822,
85,
12,
72,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2065,
67,
12316,
13332,
309,
3988,
30,
1172,
315,
22218,
2065,
3885,
18483,
302,
273,
2065,
1435,
331,
822,
85,
12,
72,
16,
2618,
13,
302,
273,
2065,
23506,
13,
331,
822,
85,
12,
72,
... | |
self.anytingElse() | self.anythingElse() | def startTagOther(self, name, attributes): self.anytingElse() self.parser.processStartTag(name, attributes) | b0d0689f6cbb1bd6f3d02d0facb874bc0c678a52 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10463/b0d0689f6cbb1bd6f3d02d0facb874bc0c678a52/parser.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
787,
1805,
8290,
12,
2890,
16,
508,
16,
1677,
4672,
365,
18,
2273,
4274,
12427,
1435,
365,
18,
4288,
18,
2567,
30512,
12,
529,
16,
1677,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
787,
1805,
8290,
12,
2890,
16,
508,
16,
1677,
4672,
365,
18,
2273,
4274,
12427,
1435,
365,
18,
4288,
18,
2567,
30512,
12,
529,
16,
1677,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
def abspath(p): return normpath(join(os.getcwd(), p)) | abspath = os.expand | def expandvars(p): """ Expand environment variables using OS_GSTrans. """ l= 512 b= swi.block(l) return b.tostring(0, swi.swi('OS_GSTrans', 'sbi;..i', p, b, l)) | 052c4fc00b0d0fbc170d27dedb696dfc21178ed1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/052c4fc00b0d0fbc170d27dedb696dfc21178ed1/riscospath.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4542,
4699,
12,
84,
4672,
3536,
16429,
3330,
3152,
1450,
5932,
67,
43,
882,
28001,
18,
3536,
328,
33,
13908,
324,
33,
1352,
77,
18,
2629,
12,
80,
13,
327,
324,
18,
14627,
371,
12,
20... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4542,
4699,
12,
84,
4672,
3536,
16429,
3330,
3152,
1450,
5932,
67,
43,
882,
28001,
18,
3536,
328,
33,
13908,
324,
33,
1352,
77,
18,
2629,
12,
80,
13,
327,
324,
18,
14627,
371,
12,
20... |
universal_newlines=universal_newlines) | universal_newlines=True) | def RunShell(command, silent_ok=False, universal_newlines=False): logging.info("Running %s", command) p = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, shell=use_shell, universal_newlines=universal_newlines) data = p.stdout.read() p.wait() p.stdout.close() if p.returncode: ErrorExit("Got error status from %s" % command) if not silent_ok and not data: ErrorExit("No output from %s" % command) return data | 51830c0e218e798dbe421fbf5161f68626b9b3b5 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/476/51830c0e218e798dbe421fbf5161f68626b9b3b5/upload.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1939,
13220,
12,
3076,
16,
10404,
67,
601,
33,
8381,
16,
26182,
67,
31276,
33,
8381,
4672,
2907,
18,
1376,
2932,
7051,
738,
87,
3113,
1296,
13,
293,
273,
6652,
18,
52,
3190,
12,
3076,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1939,
13220,
12,
3076,
16,
10404,
67,
601,
33,
8381,
16,
26182,
67,
31276,
33,
8381,
4672,
2907,
18,
1376,
2932,
7051,
738,
87,
3113,
1296,
13,
293,
273,
6652,
18,
52,
3190,
12,
3076,
... |
old_path = sys.path[:] ei_cmd = self.get_finalized_command("egg_info") path_item = normalize_path(ei_cmd.egg_base) metadata = PathMetadata( path_item, normalize_path(ei_cmd.egg_info) ) dist = Distribution(path_item, metadata, project_name=ei_cmd.egg_name) working_set.add(dist) require(str(dist.as_requirement())) | def run_tests(self): import unittest old_path = sys.path[:] ei_cmd = self.get_finalized_command("egg_info") path_item = normalize_path(ei_cmd.egg_base) metadata = PathMetadata( path_item, normalize_path(ei_cmd.egg_info) ) dist = Distribution(path_item, metadata, project_name=ei_cmd.egg_name) working_set.add(dist) require(str(dist.as_requirement())) loader_ep = EntryPoint.parse("x="+self.test_loader) loader_class = loader_ep.load(require=False) unittest.main( None, None, [unittest.__file__]+self.test_args, testLoader = loader_class() ) | b662ae02faab8b53f40c6d81cf478b537ec808f7 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8186/b662ae02faab8b53f40c6d81cf478b537ec808f7/test.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
67,
16341,
12,
2890,
4672,
1930,
2836,
3813,
4088,
67,
881,
273,
3841,
2148,
18,
2670,
2932,
92,
1546,
15,
2890,
18,
3813,
67,
6714,
13,
4088,
67,
1106,
273,
4088,
67,
881,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
67,
16341,
12,
2890,
4672,
1930,
2836,
3813,
4088,
67,
881,
273,
3841,
2148,
18,
2670,
2932,
92,
1546,
15,
2890,
18,
3813,
67,
6714,
13,
4088,
67,
1106,
273,
4088,
67,
881,
18,
... | |
self.skip("{0} does not exist".format(zero)) | self.skipTest("{0} does not exist".format(zero)) | def test_unbounded_file(self): # Issue #1174606: reading from an unbounded stream such as /dev/zero. zero = "/dev/zero" if not os.path.exists(zero): self.skip("{0} does not exist".format(zero)) if sys.maxsize > 0x7FFFFFFF: self.skip("test can only run in a 32-bit address space") if support.real_max_memuse < support._2G: self.skip("test requires at least 2GB of memory") with open(zero, "rb", buffering=0) as f: self.assertRaises(OverflowError, f.read) with open(zero, "rb") as f: self.assertRaises(OverflowError, f.read) with open(zero, "r") as f: self.assertRaises(OverflowError, f.read) | 7796577d72c9740d3bf3fb2422d81c4142634322 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3187/7796577d72c9740d3bf3fb2422d81c4142634322/test_io.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
318,
26220,
67,
768,
12,
2890,
4672,
468,
11820,
468,
2499,
5608,
4848,
26,
30,
6453,
628,
392,
640,
26220,
1407,
4123,
487,
342,
5206,
19,
7124,
18,
3634,
273,
2206,
5206,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
318,
26220,
67,
768,
12,
2890,
4672,
468,
11820,
468,
2499,
5608,
4848,
26,
30,
6453,
628,
392,
640,
26220,
1407,
4123,
487,
342,
5206,
19,
7124,
18,
3634,
273,
2206,
5206,
1... |
return self.recognize(buf) | return self.recognize(buf, attrs) | def finish (self, data, **attrs): """feed data to recognizer""" if not attrs.has_key('mimerecognizer_buf'): return data buf = attrs['mimerecognizer_buf'] buf.write(data) return self.recognize(buf) | 528e381abada85d187d95d575ece7406b045d305 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3948/528e381abada85d187d95d575ece7406b045d305/MimeRecognizer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4076,
261,
2890,
16,
501,
16,
2826,
7039,
4672,
3536,
7848,
501,
358,
14603,
1824,
8395,
309,
486,
3422,
18,
5332,
67,
856,
2668,
81,
4417,
557,
4198,
1824,
67,
4385,
11,
4672,
327,
50... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4076,
261,
2890,
16,
501,
16,
2826,
7039,
4672,
3536,
7848,
501,
358,
14603,
1824,
8395,
309,
486,
3422,
18,
5332,
67,
856,
2668,
81,
4417,
557,
4198,
1824,
67,
4385,
11,
4672,
327,
50... |
if os.path.isfile(mapfile): | if os.access(mapfile, os.R_OK): | def discoverAutoMountMedias(filename="/etc/auto.master"): result = [] if os.path.isfile(filename): for line in open(filename): line = line.strip() if not line or line[0] == "#": continue tokens = line.split() if len(tokens) < 2: continue # +auto.master syntax, not yet supported prefix, mapfile = tokens[:2] if os.path.isfile(mapfile): firstline = False for line in open(mapfile): if firstline and line.startswith("#!"): firstline = False break line = line.strip() if not line or line[0] == "#": continue tokens = line.split() if len(tokens) == 2: key, location = tokens type = None elif len(tokens) == 3: key, type, location = tokens else: continue if (type and "-fstype=iso9660" in type or location in (":/dev/cdrom", ":/dev/dvd")): mountpoint = os.path.join(prefix, key) device = location[1:] result.append(AutoMountMedia(mountpoint, device, removable=True)) return result | 0eec6fa9e4992f7301367800d0eadbc4307a37fa /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8317/0eec6fa9e4992f7301367800d0eadbc4307a37fa/media.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10372,
4965,
8725,
13265,
1795,
12,
3459,
1546,
19,
14175,
19,
6079,
18,
7525,
6,
4672,
563,
273,
5378,
309,
1140,
18,
803,
18,
291,
768,
12,
3459,
4672,
364,
980,
316,
1696,
12,
3459,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10372,
4965,
8725,
13265,
1795,
12,
3459,
1546,
19,
14175,
19,
6079,
18,
7525,
6,
4672,
563,
273,
5378,
309,
1140,
18,
803,
18,
291,
768,
12,
3459,
4672,
364,
980,
316,
1696,
12,
3459,... |
forbidden_break = False | forbidden_break_level -= 1 | def _split_message(message, srx_handler=None): # Concatenation! concat_text = [] for type, value, line in message: if type == TEXT: concat_text.append(value[0]) concat_text = u''.join(concat_text) # Get the rules if srx_handler is None: srx_handler = default_srx_handler # XXX we must handle the language here! rules = srx_handler.get_compiled_rules('en') # Get the breaks breaks = set() no_breaks = set() for break_value, regexp in rules: for match in regexp.finditer(concat_text): pos = match.end() if break_value and pos not in no_breaks: breaks.add(pos) if not break_value and pos not in breaks: no_breaks.add(pos) breaks = list(breaks) breaks.sort() # And now cut the message forbidden_break = False current_message = Message() current_length = 0 for type, value, line in message: if type == TEXT: text, context = value if forbidden_break: current_message.append_text(text, line, context) current_length += len(text) else: line_offset = 0 for absolute_pos in breaks: pos = absolute_pos - current_length if 0 <= pos and pos < len(text): before = text[:pos] text = text[pos:] # Add before to the current message if before: current_message.append_text(before, line + line_offset, context) current_length += len(before) line_offset += before.count('\n') # Send the message if it is not empty if current_message: yield current_message current_message = Message() if text: current_length += len(text) current_message.append_text(text, line + line_offset, context) elif type == START_FORMAT: forbidden_break = True current_message.append_start_format(value[0], value[1], line) elif type == END_FORMAT: forbidden_break = False current_message.append_end_format(value[0], value[1], line) # Send the last message if current_message: yield current_message | 77700ad8592a60a239690125a67e996f785afc23 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12681/77700ad8592a60a239690125a67e996f785afc23/segment.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4939,
67,
2150,
12,
2150,
16,
9133,
92,
67,
4176,
33,
7036,
4672,
468,
23873,
275,
367,
5,
3835,
67,
955,
273,
5378,
364,
618,
16,
460,
16,
980,
316,
883,
30,
309,
618,
422,
9... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4939,
67,
2150,
12,
2150,
16,
9133,
92,
67,
4176,
33,
7036,
4672,
468,
23873,
275,
367,
5,
3835,
67,
955,
273,
5378,
364,
618,
16,
460,
16,
980,
316,
883,
30,
309,
618,
422,
9... |
result = self.rescheduleJob(jobID) | result = self.rescheduleJob( jobID ) | def rescheduleJobs(self, jobIDs ): """ Reschedule all the jobs in the given list """ | 99c1bc850ba087890925b3180df206f65bb1d4b3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/99c1bc850ba087890925b3180df206f65bb1d4b3/JobDB.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
400,
2847,
7276,
12,
2890,
16,
1719,
5103,
262,
30,
3536,
1124,
2847,
777,
326,
6550,
316,
326,
864,
666,
3536,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
400,
2847,
7276,
12,
2890,
16,
1719,
5103,
262,
30,
3536,
1124,
2847,
777,
326,
6550,
316,
326,
864,
666,
3536,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
return bool(self._dict) | raise NotImplementedError("Please use a subclass that implements the __nonzero__ method") | def __nonzero__(self): #{{{ return bool(self._dict) | 5692ed97988b844b3fc73c0ff916ecbd137ea54b /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2635/5692ed97988b844b3fc73c0ff916ecbd137ea54b/mappingset.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
26449,
972,
12,
2890,
4672,
3735,
12187,
327,
1426,
12,
2890,
6315,
1576,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
26449,
972,
12,
2890,
4672,
3735,
12187,
327,
1426,
12,
2890,
6315,
1576,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
sage: E=EllipticCurve('14a1') | sage: E = EllipticCurve('14a1') | def has_split_multiplicative_reduction(self): r""" Return True if there is split multiplicative reduction. | d268d826e9ff15e8e9c26c0ba3dd27591f648207 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/d268d826e9ff15e8e9c26c0ba3dd27591f648207/ell_local_data.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
711,
67,
4939,
67,
7027,
28884,
67,
1118,
4062,
12,
2890,
4672,
436,
8395,
2000,
1053,
309,
1915,
353,
1416,
3309,
28884,
225,
20176,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
711,
67,
4939,
67,
7027,
28884,
67,
1118,
4062,
12,
2890,
4672,
436,
8395,
2000,
1053,
309,
1915,
353,
1416,
3309,
28884,
225,
20176,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
self.setup_config() for section, default_options in self.config.defaults().items(): for name, value in default_options.items(): self.config.set(section, name, value) | self.setup_config(load_defaults=True) | def _create_file(fname, data=None): fd = open(fname, 'w') if data: fd.write(data) fd.close() | 852f75ecea452223c0f0da49e9778bf587fad96a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9317/852f75ecea452223c0f0da49e9778bf587fad96a/env.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2640,
67,
768,
12,
12749,
16,
501,
33,
7036,
4672,
5194,
273,
1696,
12,
12749,
16,
296,
91,
6134,
309,
501,
30,
5194,
18,
2626,
12,
892,
13,
5194,
18,
4412,
1435,
2,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2640,
67,
768,
12,
12749,
16,
501,
33,
7036,
4672,
5194,
273,
1696,
12,
12749,
16,
296,
91,
6134,
309,
501,
30,
5194,
18,
2626,
12,
892,
13,
5194,
18,
4412,
1435,
2,
-100,
-100,... |
EXAMPLES: | EXAMPLES:: | def extend(self): r""" A deliberately broken extend method. | 03555b4381371de7688c2f08f833898ba9658b7b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/03555b4381371de7688c2f08f833898ba9658b7b/explain_pickle.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2133,
12,
2890,
4672,
436,
8395,
432,
1464,
14776,
5173,
12933,
2133,
707,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2133,
12,
2890,
4672,
436,
8395,
432,
1464,
14776,
5173,
12933,
2133,
707,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
class EditorView(gui.GladeView): | class EditorView(GladeView): | def _on_button_release(self, widget, event): """ GTK callback """ if event.button == 1 and self._pixmap != None: self.controller.stop_drawing(event) return True | f023dc351a621767d8086642a5d655543b8032f6 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7999/f023dc351a621767d8086642a5d655543b8032f6/gtk_gui.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
265,
67,
5391,
67,
9340,
12,
2890,
16,
3604,
16,
871,
4672,
3536,
19688,
47,
1348,
3536,
309,
871,
18,
5391,
422,
404,
471,
365,
6315,
14861,
1458,
480,
599,
30,
365,
18,
5723,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
265,
67,
5391,
67,
9340,
12,
2890,
16,
3604,
16,
871,
4672,
3536,
19688,
47,
1348,
3536,
309,
871,
18,
5391,
422,
404,
471,
365,
6315,
14861,
1458,
480,
599,
30,
365,
18,
5723,
... |
if object == None: print "UPDATE END", self | def _update_end(self, object = None): if object == None: print "UPDATE END", self self.thaw() return True | 138067d8187215de00b8dcd3a5c6f49c3279b10f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11399/138067d8187215de00b8dcd3a5c6f49c3279b10f/mplayercanvas.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2725,
67,
409,
12,
2890,
16,
733,
273,
599,
4672,
365,
18,
451,
2219,
1435,
327,
1053,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2725,
67,
409,
12,
2890,
16,
733,
273,
599,
4672,
365,
18,
451,
2219,
1435,
327,
1053,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... | |
event = Event(u'fakesource', u'message') event.message = u'%s%sfoo' % prefix | event = self.create_event(u'%s%sfoo' % prefix) | def test_sad_prefix_names(self): for prefix in self.sad_prefixes: event = Event(u'fakesource', u'message') event.message = u'%s%sfoo' % prefix self.processor.process(event) self.assert_addressed(event, False, u'%s%sfoo' % prefix) | 57bdc8f4c914d9e4e6ca6457bb5e08c462e27a3d /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12048/57bdc8f4c914d9e4e6ca6457bb5e08c462e27a3d/test_core.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
87,
361,
67,
3239,
67,
1973,
12,
2890,
4672,
364,
1633,
316,
365,
18,
87,
361,
67,
17777,
30,
871,
273,
365,
18,
2640,
67,
2575,
12,
89,
11,
9,
87,
9,
87,
11351,
11,
73... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
87,
361,
67,
3239,
67,
1973,
12,
2890,
4672,
364,
1633,
316,
365,
18,
87,
361,
67,
17777,
30,
871,
273,
365,
18,
2640,
67,
2575,
12,
89,
11,
9,
87,
9,
87,
11351,
11,
73... |
rpath.rename(temprp, finalrp) | rpath.rename(temprp[0], finalrp) | def recreate_meta(meta_manager): """Make regress_time mirror_metadata snapshot by patching We write to a tempfile first. Otherwise, in case of a crash, it would seem we would have an intact snapshot and partial diff, not the reverse. """ temprp = TempFile.new_in_dir(Globals.rbdir) writer = metadata.MetadataFile(temprp, 'w', check_path = 0) for rorp in meta_manager.get_meta_at_time(regress_time, None): writer.write_object(rorp) writer.close() finalrp = Globals.rbdir.append("mirror_metadata.%s.snapshot.gz" % Time.timetostring(regress_time)) assert not finalrp.lstat(), finalrp rpath.rename(temprp, finalrp) if Globals.fsync_directories: Globals.rbdir.fsync() | 6d2baeb3fdaa9646127554421913f348ccf35afd /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8033/6d2baeb3fdaa9646127554421913f348ccf35afd/regress.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
23052,
67,
3901,
12,
3901,
67,
4181,
4672,
3536,
6464,
960,
663,
67,
957,
15593,
67,
4165,
4439,
635,
4729,
310,
225,
1660,
1045,
358,
279,
13275,
1122,
18,
225,
5272,
16,
316,
648,
43... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
23052,
67,
3901,
12,
3901,
67,
4181,
4672,
3536,
6464,
960,
663,
67,
957,
15593,
67,
4165,
4439,
635,
4729,
310,
225,
1660,
1045,
358,
279,
13275,
1122,
18,
225,
5272,
16,
316,
648,
43... |
shutil.rmtree(env.OUTDIR_ROOT) | try: shutil.rmtree(env.OUTDIR_ROOT) except shutil.Error: pass | def clean(): | 1e2deee8dd29171cbae4069694d6dabe19590689 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/13644/1e2deee8dd29171cbae4069694d6dabe19590689/eagle3d.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2721,
13332,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2721,
13332,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
ON SE.RDB$INDEX_NAME=IX2.RDB$INDEX_NAME | ON SE.RDB$INDEX_NAME=IX2.RDB$INDEX_NAME AND SE.RDB$FIELD_POSITION=CSE.RDB$FIELD_POSITION | def reflecttable(self, connection, table): #TODO: map these better column_func = { 14 : lambda r: sqltypes.String(r['FLEN']), # TEXT 7 : lambda r: sqltypes.Integer(), # SHORT 8 : lambda r: sqltypes.Integer(), # LONG 9 : lambda r: sqltypes.Float(), # QUAD 10 : lambda r: sqltypes.Float(), # FLOAT 27 : lambda r: sqltypes.Float(), # DOUBLE 35 : lambda r: sqltypes.DateTime(), # TIMESTAMP 37 : lambda r: sqltypes.String(r['FLEN']), # VARYING 261: lambda r: sqltypes.TEXT(), # BLOB 40 : lambda r: sqltypes.Char(r['FLEN']), # CSTRING 12 : lambda r: sqltypes.Date(), # DATE 13 : lambda r: sqltypes.Time(), # TIME 16 : lambda r: sqltypes.Numeric(precision=r['FPREC'], length=r['FSCALE'] * -1) #INT64 } tblqry = """\ SELECT DISTINCT R.RDB$FIELD_NAME AS FNAME, R.RDB$NULL_FLAG AS NULL_FLAG, R.RDB$FIELD_POSITION, F.RDB$FIELD_TYPE AS FTYPE, F.RDB$FIELD_SUB_TYPE AS STYPE, F.RDB$FIELD_LENGTH AS FLEN, F.RDB$FIELD_PRECISION AS FPREC, F.RDB$FIELD_SCALE AS FSCALE FROM RDB$RELATION_FIELDS R JOIN RDB$FIELDS F ON R.RDB$FIELD_SOURCE=F.RDB$FIELD_NAME WHERE F.RDB$SYSTEM_FLAG=0 and R.RDB$RELATION_NAME=? ORDER BY R.RDB$FIELD_POSITION""" keyqry = """\ SELECT SE.RDB$FIELD_NAME SENAME FROM RDB$RELATION_CONSTRAINTS RC JOIN RDB$INDEX_SEGMENTS SE ON RC.RDB$INDEX_NAME=SE.RDB$INDEX_NAME WHERE RC.RDB$CONSTRAINT_TYPE=? AND RC.RDB$RELATION_NAME=?""" fkqry = """\ SELECT RC.RDB$CONSTRAINT_NAME CNAME, CSE.RDB$FIELD_NAME FNAME, IX2.RDB$RELATION_NAME RNAME, SE.RDB$FIELD_NAME SENAME FROM RDB$RELATION_CONSTRAINTS RC JOIN RDB$INDICES IX1 ON IX1.RDB$INDEX_NAME=RC.RDB$INDEX_NAME JOIN RDB$INDICES IX2 ON IX2.RDB$INDEX_NAME=IX1.RDB$FOREIGN_KEY JOIN RDB$INDEX_SEGMENTS CSE ON CSE.RDB$INDEX_NAME=IX1.RDB$INDEX_NAME JOIN RDB$INDEX_SEGMENTS SE ON SE.RDB$INDEX_NAME=IX2.RDB$INDEX_NAME WHERE RC.RDB$CONSTRAINT_TYPE=? AND RC.RDB$RELATION_NAME=? ORDER BY SE.RDB$INDEX_NAME, SE.RDB$FIELD_POSITION""" | 046703498161bc3cd3851e24e20ce87796440d31 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/1074/046703498161bc3cd3851e24e20ce87796440d31/firebird.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3037,
2121,
12,
2890,
16,
1459,
16,
1014,
4672,
468,
6241,
30,
852,
4259,
7844,
1057,
67,
644,
273,
288,
5045,
294,
3195,
436,
30,
21899,
18,
780,
12,
86,
3292,
42,
13017,
3546,
3631,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3037,
2121,
12,
2890,
16,
1459,
16,
1014,
4672,
468,
6241,
30,
852,
4259,
7844,
1057,
67,
644,
273,
288,
5045,
294,
3195,
436,
30,
21899,
18,
780,
12,
86,
3292,
42,
13017,
3546,
3631,
... |
"[Status: %s] " % (jdict.get('status', 'None'))] | "[State: %s] " % (jdict.get('state', 'None'))] | def log_job_end(jdict): ln = "" ln += "[%s]: " % (time.asctime(time.localtime(time.time()))) ln += "Finished Job %s" % (jdict["job_id"]) log_write(ln) ## write to a special log file if jdict.get("private_job", True): private_text = "private" else: private_text = "public" submit_time = jdict.get('submit_time', 0.0) run_time_begin = jdict.get('run_time_begin', 0.0) run_time_end = jdict.get('run_time_end', 0.0) processing_time = timediff(run_time_begin, run_time_end) l = ["[Submit time: %s]" % (timestring(submit_time)), "[Start time: %s] " % (timestring(run_time_begin)), "[End time: %s] " % (timestring(run_time_end)), "[Processing time: %s] " % (processing_time), "[IP : %s] " % (jdict.get("ip_addr", "000.000.000.000")), "[Email: %s] " % (jdict.get("email", "nobody@nowhere.com")), "[Privacy: %s] " % (private_text), "[Job ID: %s] " % (jdict.get("job_id", "EEK!!")), "[Structure ID: %s] " % (jdict.get("structure_id", "----")), "[Chain sizes: %s] " % (chain_size_string(jdict)), "[TLS Model: %s] " % (jdict.get('tls_model', 'None')), "[Weight: %s] " % (jdict.get('weight', 'None')), "[Atoms: %s] " % (jdict.get('include_atoms', 'None')), "[Status: %s] " % (jdict.get('status', 'None'))] try: open(conf.LOG_PATH, "a").write(" ".join(l) + "\n") except IOError: log_write("ERROR: cannot open logfile %s" % (conf.LOG_PATH)) | b48d806b00beca8c5e289445a77097960da89cac /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10674/b48d806b00beca8c5e289445a77097960da89cac/webtlsmdrund.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
613,
67,
4688,
67,
409,
12,
78,
1576,
4672,
7211,
225,
273,
1408,
7211,
1011,
30048,
87,
14542,
315,
738,
261,
957,
18,
345,
21261,
12,
957,
18,
3729,
957,
12,
957,
18,
957,
1435,
20... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
613,
67,
4688,
67,
409,
12,
78,
1576,
4672,
7211,
225,
273,
1408,
7211,
1011,
30048,
87,
14542,
315,
738,
261,
957,
18,
345,
21261,
12,
957,
18,
3729,
957,
12,
957,
18,
957,
1435,
20... |
return stp, f, g | def step(self, stp, f, g, c1, c2, xtol, isave, dsave): if self.task[:5] == 'START': | 399495d517e260560383dce42d11299b57963647 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1380/399495d517e260560383dce42d11299b57963647/linesearch.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2235,
12,
2890,
16,
384,
84,
16,
284,
16,
314,
16,
276,
21,
16,
276,
22,
16,
619,
3490,
16,
353,
836,
16,
3780,
836,
4672,
309,
365,
18,
4146,
10531,
25,
65,
422,
296,
7570,
4278,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2235,
12,
2890,
16,
384,
84,
16,
284,
16,
314,
16,
276,
21,
16,
276,
22,
16,
619,
3490,
16,
353,
836,
16,
3780,
836,
4672,
309,
365,
18,
4146,
10531,
25,
65,
422,
296,
7570,
4278,
... | |
small = 'a|an|and|as|at|but|by|en|for|if|in|of|on|or|the|to|v\.?|via|vs\.?' punct = "[!\" small_words = re.compile(r'^(%s)$' % small, re.I) inline_period = re.compile(r'[a-zA-Z][.][a-zA-Z]') uc_elsewhere = re.compile(r'%s*?[a-zA-Z]+[A-Z]+?' % punct) capfirst = re.compile(r"^%s*?([A-Za-z])" % punct) small_first = re.compile(r'^(%s*)(%s)\b' % (punct, small), re.I) small_last = re.compile(r'\b(%s)%s?$' % (small, punct), re.I) subphrase = re.compile(r'([:.;?!][ ])(%s)' % small) | def titlecase(text): """ This filter changes all words to Title Caps, and attempts to be clever about *un*capitalizing small words like a/an/the in the input. The list of "small words" which are not capped comes from the New York Times Manual of Style, plus 'vs' and 'v'. """ small = 'a|an|and|as|at|but|by|en|for|if|in|of|on|or|the|to|v\.?|via|vs\.?' punct = "[!\"#$%&'()*+,-./:;?@[\\\\\\]_`{|}~]" small_words = re.compile(r'^(%s)$' % small, re.I) inline_period = re.compile(r'[a-zA-Z][.][a-zA-Z]') uc_elsewhere = re.compile(r'%s*?[a-zA-Z]+[A-Z]+?' % punct) capfirst = re.compile(r"^%s*?([A-Za-z])" % punct) small_first = re.compile(r'^(%s*)(%s)\b' % (punct, small), re.I) small_last = re.compile(r'\b(%s)%s?$' % (small, punct), re.I) subphrase = re.compile(r'([:.;?!][ ])(%s)' % small) words = re.split('\s', text) line = [] for word in words: if inline_period.search(word) or uc_elsewhere.match(word): line.append(word) continue if small_words.match(word): line.append(word.lower()) continue line.append(capfirst.sub(lambda m: m.group(0).upper(), word)) line = " ".join(line) line = small_first.sub(lambda m: '%s%s' % (m.group(1), m.group(2).capitalize()), line) line = small_last.sub(lambda m: m.group(0).capitalize(), line) line = subphrase.sub(lambda m: '%s%s' % (m.group(1), m.group(2).capitalize()), line) return line | faef923bff691bbede9b642e912e5874bbe1fbb1 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8761/faef923bff691bbede9b642e912e5874bbe1fbb1/titlecase.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2077,
3593,
12,
955,
4672,
225,
3536,
1220,
1034,
3478,
777,
4511,
358,
10984,
385,
6679,
16,
471,
7531,
358,
506,
1619,
502,
2973,
380,
318,
14,
5909,
7053,
6894,
5264,
4511,
3007,
279,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2077,
3593,
12,
955,
4672,
225,
3536,
1220,
1034,
3478,
777,
4511,
358,
10984,
385,
6679,
16,
471,
7531,
358,
506,
1619,
502,
2973,
380,
318,
14,
5909,
7053,
6894,
5264,
4511,
3007,
279,... | |
show = eval(show) | show = eval(show, {}, {}) | def locations_kml(self, path='', show='', geo_query='', REQUEST=None): """ """ path = path or '/' show = eval(show) | 9aeafcb331cc3cc2e4727af0c5fc630e5a44af84 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/3287/9aeafcb331cc3cc2e4727af0c5fc630e5a44af84/GeoMapTool.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7838,
67,
79,
781,
12,
2890,
16,
589,
2218,
2187,
2405,
2218,
2187,
7856,
67,
2271,
2218,
2187,
12492,
33,
7036,
4672,
3536,
3536,
589,
273,
589,
578,
2023,
2405,
273,
5302,
12,
4500,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7838,
67,
79,
781,
12,
2890,
16,
589,
2218,
2187,
2405,
2218,
2187,
7856,
67,
2271,
2218,
2187,
12492,
33,
7036,
4672,
3536,
3536,
589,
273,
589,
578,
2023,
2405,
273,
5302,
12,
4500,
... |
item.expire() | item.stopUpload() | def handle_cancel_all_downloads(self, message): for item in views.pausedItems: item.expire() | ecfdda35df9231c3e76dfc1a25cad2038ea25c6f /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12354/ecfdda35df9231c3e76dfc1a25cad2038ea25c6f/messagehandler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
67,
10996,
67,
454,
67,
7813,
87,
12,
2890,
16,
883,
4672,
364,
761,
316,
7361,
18,
8774,
3668,
3126,
30,
761,
18,
14070,
1435,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
67,
10996,
67,
454,
67,
7813,
87,
12,
2890,
16,
883,
4672,
364,
761,
316,
7361,
18,
8774,
3668,
3126,
30,
761,
18,
14070,
1435,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
import re _queryprog = re.compile('^(.*)\?([^?]*)$') | import re _queryprog = re.compile('^(.*)\?([^?]*)$') | def splitquery(url): global _queryprog if _queryprog is None: import re _queryprog = re.compile('^(.*)\?([^?]*)$') match = _queryprog.match(url) if match: return match.group(1, 2) return url, None | 52c990fef6d843442d24f88c38914bc52c0210e0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/52c990fef6d843442d24f88c38914bc52c0210e0/urllib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1416,
2271,
12,
718,
4672,
2552,
389,
2271,
14654,
309,
389,
2271,
14654,
353,
599,
30,
1930,
283,
389,
2271,
14654,
273,
283,
18,
11100,
2668,
66,
27816,
35,
8178,
35,
15471,
8,
6134,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1416,
2271,
12,
718,
4672,
2552,
389,
2271,
14654,
309,
389,
2271,
14654,
353,
599,
30,
1930,
283,
389,
2271,
14654,
273,
283,
18,
11100,
2668,
66,
27816,
35,
8178,
35,
15471,
8,
6134,
... |
if input_type not in ['ui', 'layouts', 'modules', 'qt4', 'languages', 'external'] or output is None: | if input_type not in ['ui', 'layouts', 'modules', 'qt4', 'languages', 'external', 'formats'] or output is None: | def external_l10n(input_files, output, base): '''Generate pot file from lib/external_templates''' output = open(output, 'w') Template = re.compile(r'^Template\s+(.*)') GuiName = re.compile(r'\s*GuiName\s+(.*)') HelpTextStart = re.compile(r'\s*HelpText\s') HelpTextSection = re.compile(r'\s*(\S.*)\s*$') HelpTextEnd = re.compile(r'\s*HelpTextEnd\s') i = -1 for src in input_files: input = open(src) inHelp = False hadHelp = False prev_help_string = '' for lineno, line in enumerate(input.readlines()): if Template.match(line): (string,) = Template.match(line).groups() elif GuiName.match(line): (string,) = GuiName.match(line).groups() elif inHelp: if HelpTextEnd.match(line): if hadHelp: print >> output, '\nmsgstr ""\n' inHelp = False hadHelp = False prev_help_string = '' elif HelpTextSection.match(line): (help_string,) = HelpTextSection.match(line).groups() help_string = help_string.replace('"', '') if help_string != "" and prev_help_string == '': print >> output, '#: %s:%d\nmsgid ""\n"%s\\n"' % \ (relativePath(src, base), lineno+1, help_string) hadHelp = True elif help_string != "": print >> output, '"%s\\n"' % help_string prev_help_string = help_string elif HelpTextStart.match(line): inHelp = True prev_help_string = '' else: continue string = string.replace('"', '') if string != "" and not inHelp: print >> output, '#: %s:%d\nmsgid "%s"\nmsgstr ""\n' % \ (relativePath(src, base), lineno+1, string) input.close() output.close() | 21bd7671ec96b6c2f4934bead5c21fdf114e8861 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7514/21bd7671ec96b6c2f4934bead5c21fdf114e8861/lyx_pot.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3903,
67,
80,
2163,
82,
12,
2630,
67,
2354,
16,
876,
16,
1026,
4672,
9163,
4625,
5974,
585,
628,
2561,
19,
9375,
67,
8502,
26418,
876,
273,
1696,
12,
2844,
16,
296,
91,
6134,
5035,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3903,
67,
80,
2163,
82,
12,
2630,
67,
2354,
16,
876,
16,
1026,
4672,
9163,
4625,
5974,
585,
628,
2561,
19,
9375,
67,
8502,
26418,
876,
273,
1696,
12,
2844,
16,
296,
91,
6134,
5035,
2... |
try: self.connect(**kwargs).close() | try: c = self.connect(**kwargs) try: c.session() finally: c.close() | def ready(self, **kwargs): """Wait till broker is ready to serve clients""" # First make sure the broker is listening by checking the log. if not retry(self.log_ready): raise Exception( "Timed out waiting for broker %s%s"%(self.name, error_line(self.log,4))) # Make a connection, this will wait for extended cluster init to finish. try: self.connect(**kwargs).close() except: raise RethrownException( "Broker %s failed ready test%s"%(self.name,error_line(self.log,4))) | 9e15aa608f16ea9c30acd922a5b410cdf12e00a4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/197/9e15aa608f16ea9c30acd922a5b410cdf12e00a4/brokertest.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5695,
12,
2890,
16,
2826,
4333,
4672,
3536,
5480,
21364,
8625,
353,
5695,
358,
12175,
7712,
8395,
468,
5783,
1221,
3071,
326,
8625,
353,
13895,
635,
6728,
326,
613,
18,
309,
486,
3300,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5695,
12,
2890,
16,
2826,
4333,
4672,
3536,
5480,
21364,
8625,
353,
5695,
358,
12175,
7712,
8395,
468,
5783,
1221,
3071,
326,
8625,
353,
13895,
635,
6728,
326,
613,
18,
309,
486,
3300,
1... |
"descip": "lstring", | "descrip": "lstring", | def set_time(self, gps): self.time, self.time_ns = gps.seconds, gps.nanoseconds | de4cff3b3d0d37d0d76122616e471b43eed70b1b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5758/de4cff3b3d0d37d0d76122616e471b43eed70b1b/lsctables.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
67,
957,
12,
2890,
16,
20985,
4672,
365,
18,
957,
16,
365,
18,
957,
67,
2387,
273,
20985,
18,
7572,
16,
20985,
18,
13569,
23272,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
67,
957,
12,
2890,
16,
20985,
4672,
365,
18,
957,
16,
365,
18,
957,
67,
2387,
273,
20985,
18,
7572,
16,
20985,
18,
13569,
23272,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
if self.content[5]: r.append(`self.content[5]`) | if categories: r.append(`categories`) | def __repr__(self): r = [] r.append("{") if self.content[3] is None: r.append("*") else: r.append(self.content[3]) if self.content[4]: r.append(" ") r.append(`self.content[4]`) elif self.content[5]: r.append(" {}") if self.content[5]: r.append(`self.content[5]`) r.append("}") return "".join(r) | fe0476adf5e7d348b399ba81e2e3a2176d84d524 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/2408/fe0476adf5e7d348b399ba81e2e3a2176d84d524/wordfilter.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
12715,
972,
12,
2890,
4672,
436,
273,
5378,
436,
18,
6923,
2932,
95,
7923,
309,
365,
18,
1745,
63,
23,
65,
353,
599,
30,
436,
18,
6923,
2932,
7388,
13,
469,
30,
436,
18,
6923,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
12715,
972,
12,
2890,
4672,
436,
273,
5378,
436,
18,
6923,
2932,
95,
7923,
309,
365,
18,
1745,
63,
23,
65,
353,
599,
30,
436,
18,
6923,
2932,
7388,
13,
469,
30,
436,
18,
6923,
... |
text = "" | text = u"" | def page_link(page, variant=model.Text.HEAD, create=False, link=True, icon=True, icon_size=16): if create: page = page.encode('utf-8').replace("<", "<") page = page.replace(">", ">") page = page.replace("&", "&") url = urllib.quote(page) url = "/page/new?title=%s" % url url = instance_url(c.instance, path=url) return "<a class='page_link new' href='%s'>%s</a>" % (url, page) if page.is_deleted(): link = False if not link: return cgi.escape(page.title) text = "" _text = page.variant_head(variant) if icon: text += "<img class='dgb_icon' src='%s' /> " % text_icon(_text, page, size=icon_size) text += "<a class='page_link exists' href='%s'>%s</a>" title = page.title if variant != model.Text.HEAD: title = "%s <code>(%s)</code>" % (title, variant) return text % (entity_url(_text), cgi.escape(title)) | d3aabe3fe9dc3859e3a5d92f8d18e6ed2121896a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10360/d3aabe3fe9dc3859e3a5d92f8d18e6ed2121896a/helpers.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1363,
67,
1232,
12,
2433,
16,
5437,
33,
2284,
18,
1528,
18,
12458,
16,
752,
33,
8381,
16,
1692,
33,
5510,
16,
4126,
33,
5510,
16,
4126,
67,
1467,
33,
2313,
4672,
309,
752,
30,
1363,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1363,
67,
1232,
12,
2433,
16,
5437,
33,
2284,
18,
1528,
18,
12458,
16,
752,
33,
8381,
16,
1692,
33,
5510,
16,
4126,
33,
5510,
16,
4126,
67,
1467,
33,
2313,
4672,
309,
752,
30,
1363,
... |
tag = wc.filter.HtmlTags.check_spelling(tag, self.url) | tag = wc.filter.html.check_spelling(tag, self.url) | def _start_element (self, tag, attrs, startend): tag = wc.filter.HtmlTags.check_spelling(tag, self.url) self.stack.append(tag) if not self.valid: return if tag in rss_allowed: self.outbuf.write(u"<%s" % tag) if attrs: quote = wc.HtmlParser.htmllib.quote_attrval for attr in attrs: if attr in rss_allowed[tag]: val = attrs[attr] self.outbuf.write(u' %s="%s"' % (attr, quote(val))) if startend: self.outbuf.write(u"/>") else: self.outbuf.write(u">") elif not startend: self.valid = False self.stack = [tag] | a7d64b9a3c3e6a8d17da2470e8d1a16a5f926407 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3948/a7d64b9a3c3e6a8d17da2470e8d1a16a5f926407/RssHtmlFilter.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1937,
67,
2956,
261,
2890,
16,
1047,
16,
3422,
16,
787,
409,
4672,
1047,
273,
14310,
18,
2188,
18,
2620,
18,
1893,
67,
1752,
1165,
310,
12,
2692,
16,
365,
18,
718,
13,
365,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1937,
67,
2956,
261,
2890,
16,
1047,
16,
3422,
16,
787,
409,
4672,
1047,
273,
14310,
18,
2188,
18,
2620,
18,
1893,
67,
1752,
1165,
310,
12,
2692,
16,
365,
18,
718,
13,
365,
18,
... |
console.debug("Found manifest: " + os.path.join(root, name)) | console.debug("Found manifest: " + repr(os.path.join(root, name))) | def scanRepository(self): console.info("Scanning %s" %self.dir) console.indent() demoDir = "%sdemo%s" %(os.sep,os.sep) manifestPaths = [] for root, dirs, files in os.walk(self.dir, topdown=True): for name in files[:]: #if "Bugs" in root: # dirs = [] # files = [] if name == "Manifest.json" and root != self.dir: console.debug("Found manifest: " + os.path.join(root, name)) manifestPath = os.path.join(root,name) if not demoDir in manifestPath: manifestPaths.append(manifestPath) dirs = [] files = [] console.info("Found %s manifests" %len(manifestPaths)) console.outdent() return manifestPaths | 484d1421cf7e72b8ccae73c5eacd181e34d1beac /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5718/484d1421cf7e72b8ccae73c5eacd181e34d1beac/repository.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4135,
3305,
12,
2890,
4672,
2983,
18,
1376,
2932,
1541,
10903,
738,
87,
6,
738,
2890,
18,
1214,
13,
2983,
18,
9355,
1435,
21477,
1621,
273,
2213,
87,
27928,
9,
87,
6,
8975,
538,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4135,
3305,
12,
2890,
4672,
2983,
18,
1376,
2932,
1541,
10903,
738,
87,
6,
738,
2890,
18,
1214,
13,
2983,
18,
9355,
1435,
21477,
1621,
273,
2213,
87,
27928,
9,
87,
6,
8975,
538,
18,
... |
raise bb.runqueue.TaskFailure(self.failed_fnids) for fnid in self.failed_fnids: | raise bb.runqueue.TaskFailure(self.rqexe.failed_fnids) for fnid in self.rqexe.failed_fnids: | def execute_runqueue(self): """ Run the tasks in a queue prepared by rqdata.prepare() Upon failure, optionally try to recover the build using any alternate providers (if the abort on failure configuration option isn't set) """ | 6739a629e96a93c037e4d863e7f842a317c0c571 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8123/6739a629e96a93c037e4d863e7f842a317c0c571/runqueue.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1836,
67,
2681,
4000,
12,
2890,
4672,
3536,
1939,
326,
4592,
316,
279,
2389,
8208,
635,
22227,
892,
18,
9366,
1435,
587,
500,
5166,
16,
8771,
775,
358,
5910,
326,
1361,
1450,
1281,
12184... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1836,
67,
2681,
4000,
12,
2890,
4672,
3536,
1939,
326,
4592,
316,
279,
2389,
8208,
635,
22227,
892,
18,
9366,
1435,
587,
500,
5166,
16,
8771,
775,
358,
5910,
326,
1361,
1450,
1281,
12184... |
def check_sizeof(self, o, size): | def check_sizeof(self, o, size, size2=None): """Check size of o. Possible are size and optionally size2).""" | def check_sizeof(self, o, size): result = sys.getsizeof(o) msg = 'wrong size for %s: got %d, expected %d' \ % (type(o), result, size) self.assertEqual(result, size, msg) | dc1b20e8ed636de34b1eedeb54eb2d3fe02c96cf /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8125/dc1b20e8ed636de34b1eedeb54eb2d3fe02c96cf/test_sys.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
1467,
792,
12,
2890,
16,
320,
16,
963,
16,
963,
22,
33,
7036,
4672,
3536,
1564,
963,
434,
320,
18,
25433,
854,
963,
471,
8771,
963,
22,
13,
12123,
563,
273,
2589,
18,
588,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
1467,
792,
12,
2890,
16,
320,
16,
963,
16,
963,
22,
33,
7036,
4672,
3536,
1564,
963,
434,
320,
18,
25433,
854,
963,
471,
8771,
963,
22,
13,
12123,
563,
273,
2589,
18,
588,
... |
id = garbageIds[i] | idx = garbageIndices[i] | def run(self): # do the garbage collection wasOn = gcDebugOn() oldFlags = gc.get_debug() if not wasOn: gc.set_debug(gc.DEBUG_SAVEALL) gc.collect() self.garbage = list(gc.garbage) # only yield if there's more time-consuming work to do, # if there's no garbage, give instant feedback if len(self.garbage) > 0: yield None # don't repr the garbage list if we don't have to if self.notify.getDebug(): self.notify.debug('self.garbage == %s' % safeRepr(self.garbage)) del gc.garbage[:] if not wasOn: gc.set_debug(oldFlags) | 233501d0a6e3d831ba670b198f335bccbf84ffb5 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8543/233501d0a6e3d831ba670b198f335bccbf84ffb5/GarbageReport.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
468,
741,
326,
15340,
1849,
1703,
1398,
273,
8859,
2829,
1398,
1435,
1592,
5094,
273,
8859,
18,
588,
67,
4148,
1435,
309,
486,
1703,
1398,
30,
8859,
18,
542,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
468,
741,
326,
15340,
1849,
1703,
1398,
273,
8859,
2829,
1398,
1435,
1592,
5094,
273,
8859,
18,
588,
67,
4148,
1435,
309,
486,
1703,
1398,
30,
8859,
18,
542,
67,
... |
sys.stdout.write('\n') | stream.write('\n') | def unix_getpass(prompt='Password: '): """Prompt for a password, with echo turned off. Restore terminal settings at end. """ try: fd = sys.stdin.fileno() except: return default_getpass(prompt) old = termios.tcgetattr(fd) # a copy to save new = old[:] new[3] = new[3] & ~termios.ECHO # 3 == 'lflags' try: termios.tcsetattr(fd, termios.TCSADRAIN, new) passwd = _raw_input(prompt) finally: termios.tcsetattr(fd, termios.TCSADRAIN, old) sys.stdout.write('\n') return passwd | 1f1f631dd29e292259f19f3b5e641e5ae0d429db /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/1f1f631dd29e292259f19f3b5e641e5ae0d429db/getpass.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9753,
67,
588,
5466,
12,
13325,
2218,
3913,
30,
296,
4672,
3536,
15967,
364,
279,
2201,
16,
598,
3376,
21826,
3397,
18,
225,
11197,
8651,
1947,
622,
679,
18,
3536,
225,
775,
30,
5194,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9753,
67,
588,
5466,
12,
13325,
2218,
3913,
30,
296,
4672,
3536,
15967,
364,
279,
2201,
16,
598,
3376,
21826,
3397,
18,
225,
11197,
8651,
1947,
622,
679,
18,
3536,
225,
775,
30,
5194,
... |
for ixurl in ( 'http://www.publications.parliament.uk/pa/cm/cmregmem/memi02.htm', 'http://www.publications.parliament.uk/pa/cm200304/cmregmem/memi02.htm' ): ur = urllib.urlopen(ixurl) content = ur.read() ur.close(); allurls = re.findall('<a href="([^>]*)">(?i)', content) for url in allurls: if url.find("memi02") >= 0: url = urlparse.urljoin(ixurl, url) ur = urllib.urlopen(url) content = ur.read() ur.close(); content = content.replace(" ", " ") alldates = re.findall('<[Bb]>(\d+ [A-Z][a-z]* \d\d\d\d)', content) if len(alldates) != 1: print alldates raise Exception, 'Date match failed, expected one got %d\n%s' % (len(alldates), url) date = mx.DateTime.DateTimeFrom(alldates[0]).date urls.append((date, url)) | ixurl = 'http://www.publications.parliament.uk/pa/cm/cmregmem/memi02.htm' ur = urllib.urlopen(ixurl) content = ur.read() ur.close(); allurls = re.findall('<a href="([^>]*)">(?i)', content) for url in allurls: if url.find("100203/") >= 0: continue if url.find("memi02") >= 0: url = urlparse.urljoin(ixurl, url) ur = urllib.urlopen(url) content = ur.read() ur.close(); content = content.replace(" ", " ") alldates = re.findall('<[Bb]>(\d+ [A-Z][a-z]* \d\d\d\d)', content) if len(alldates) != 1: print alldates raise Exception, 'Date match failed, expected one got %d\n%s' % (len(alldates), url) date = mx.DateTime.DateTimeFrom(alldates[0]).date urls.append((date, url)) | def FindRegmemPages(): urls = [] # Meta index is here: 'http://www.publications.parliament.uk/pa/cm/cmhocpap.htm' # We just grab some specific session index pages as it is too hard otherwise. # The first URL /pa/cm/cmregmem/memi02.htm should get all new regmems as they # arrive anyway. for ixurl in ( 'http://www.publications.parliament.uk/pa/cm/cmregmem/memi02.htm', 'http://www.publications.parliament.uk/pa/cm200304/cmregmem/memi02.htm' ): # print "IXURL", ixurl ur = urllib.urlopen(ixurl) content = ur.read() ur.close(); # <A HREF="/pa/cm199900/cmregmem/memi02.htm">Register # of Members' Interests November 2000</A> allurls = re.findall('<a href="([^>]*)">(?i)', content) for url in allurls: #print url if url.find("memi02") >= 0: url = urlparse.urljoin(ixurl, url) # find date ur = urllib.urlopen(url) content = ur.read() ur.close(); # <B>14 May 2001 (Dissolution)</B> content = content.replace(" ", " ") alldates = re.findall('<[Bb]>(\d+ [A-Z][a-z]* \d\d\d\d)', content) if len(alldates) != 1: print alldates raise Exception, 'Date match failed, expected one got %d\n%s' % (len(alldates), url) date = mx.DateTime.DateTimeFrom(alldates[0]).date #print date, url urls.append((date, url)) return urls | 3090b80192d51dd16b12afda6222144c3eeacce2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8428/3090b80192d51dd16b12afda6222144c3eeacce2/regmempullgluepages.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4163,
1617,
3917,
5716,
13332,
6903,
273,
5378,
468,
6565,
770,
353,
2674,
30,
296,
2505,
2207,
5591,
18,
482,
1012,
18,
1065,
549,
14773,
18,
19445,
19,
8774,
19,
7670,
19,
7670,
76,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4163,
1617,
3917,
5716,
13332,
6903,
273,
5378,
468,
6565,
770,
353,
2674,
30,
296,
2505,
2207,
5591,
18,
482,
1012,
18,
1065,
549,
14773,
18,
19445,
19,
8774,
19,
7670,
19,
7670,
76,
... |
self.wfile.flush() | if not self.wfile.closed: self.wfile.flush() | def finish(self): self.wfile.flush() self.wfile.close() self.rfile.close() | 4cedc1e84ee174cebaa59295690cfd5c6f10a0b7 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/4cedc1e84ee174cebaa59295690cfd5c6f10a0b7/SocketServer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4076,
12,
2890,
4672,
309,
486,
365,
18,
91,
768,
18,
12204,
30,
365,
18,
91,
768,
18,
11330,
1435,
365,
18,
91,
768,
18,
4412,
1435,
365,
18,
86,
768,
18,
4412,
1435,
2,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4076,
12,
2890,
4672,
309,
486,
365,
18,
91,
768,
18,
12204,
30,
365,
18,
91,
768,
18,
11330,
1435,
365,
18,
91,
768,
18,
4412,
1435,
365,
18,
86,
768,
18,
4412,
1435,
2,
-100,
-10... |
def _outKeyValFill(key, val): | def _outKeyValFill(self, key, val): | def _outKeyValFill(key, val): """ Return a key value pair in the common two column output format. """ nxt = ' ' * (len(key) - 2) + ': ' return fill(val, width=self.term.columns, initial_indent=key, subsequent_indent=nxt) | 6c840a0fcf28824ef33acdb2c99eb45017550eb9 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5445/6c840a0fcf28824ef33acdb2c99eb45017550eb9/output.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
659,
653,
3053,
8026,
12,
2890,
16,
498,
16,
1244,
4672,
3536,
2000,
279,
498,
460,
3082,
316,
326,
2975,
2795,
1057,
876,
740,
18,
3536,
29465,
273,
296,
296,
380,
261,
1897,
12,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
659,
653,
3053,
8026,
12,
2890,
16,
498,
16,
1244,
4672,
3536,
2000,
279,
498,
460,
3082,
316,
326,
2975,
2795,
1057,
876,
740,
18,
3536,
29465,
273,
296,
296,
380,
261,
1897,
12,... |
base = 'http://pida.vm.bytemark.co.uk/projects/pida/' | base = 'http://tracos.org/pida/' | def maketicket(self, summary, description, name='Pida Bug report'): base = 'http://pida.vm.bytemark.co.uk/projects/pida/' url = '%s/newticket' % base.rstrip('/') postdata={'reporter': name, 'summary': summary, 'description': description, 'action': 'create', 'status': 'new', 'milestone': '0.3',} data = urllib.urlencode(postdata) request = urllib2.Request(url, data=data) try: page = urllib2.urlopen(request).read() except: page = '' if 'id="ticket"' in page: number = page.split('<title>', 1)[-1].split(' ', 1)[0] number = number.strip('#') reply = ('New ticket at: %s/ticket/%s' % (base.rstrip('/'), number)) else: reply = 'Posting a new bug report failed.' self._print(reply) | cec65c3b312309df057a69c2390296fa98d394db /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2233/cec65c3b312309df057a69c2390296fa98d394db/debugwindow.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
29796,
278,
29378,
12,
2890,
16,
4916,
16,
2477,
16,
508,
2218,
12478,
69,
16907,
2605,
11,
4672,
1026,
273,
296,
2505,
2207,
313,
1077,
538,
18,
3341,
19,
6610,
69,
2473,
880,
273,
19... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
29796,
278,
29378,
12,
2890,
16,
4916,
16,
2477,
16,
508,
2218,
12478,
69,
16907,
2605,
11,
4672,
1026,
273,
296,
2505,
2207,
313,
1077,
538,
18,
3341,
19,
6610,
69,
2473,
880,
273,
19... |
self.player, self.queue_audio, self.convert, | self.player, self.volume, self.convert, | def change_effect(self, effect_name, effect_type): | 205cfae3a6252fa8d3de771366e5bfba15a8f3b4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7705/205cfae3a6252fa8d3de771366e5bfba15a8f3b4/sltv.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2549,
67,
13867,
12,
2890,
16,
5426,
67,
529,
16,
5426,
67,
723,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2549,
67,
13867,
12,
2890,
16,
5426,
67,
529,
16,
5426,
67,
723,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.