rem stringlengths 2 226k | add stringlengths 0 227k | context stringlengths 8 228k | meta stringlengths 156 215 | input_ids list | attention_mask list | labels list |
|---|---|---|---|---|---|---|
flt_pattern =regex.compile('\([0-9]+\.[0-9]+\)') | flt_pattern =regex.compile(':\([0-9]+\.[0-9]+\)') | def __init__(self,*args): """Return a new date-time object | b753086612169dfe46169785b0f71dcbe4435f62 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9658/b753086612169dfe46169785b0f71dcbe4435f62/DateTime.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
14,
1968,
4672,
3536,
990,
279,
394,
1509,
17,
957,
733,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
14,
1968,
4672,
3536,
990,
279,
394,
1509,
17,
957,
733,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
Return False if the peak time of the sim is "near" the burst event. | Return False (injection matches coinc) if the peak time of the sim is "near" the burst event. | def ExcessPowerNearCoincCompare(sim, burst): """ Return False if the peak time of the sim is "near" the burst event. """ return not SimBurstUtils.burst_is_near_injection(sim, burst.start_time, burst.start_time_ns, burst.duration, burst.ifo) | f23162e4f63afbd6a2032de2d75ca44643717584 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3592/f23162e4f63afbd6a2032de2d75ca44643717584/ligolw_binjfind.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1312,
614,
13788,
24686,
27055,
71,
8583,
12,
9812,
16,
27966,
4672,
3536,
2000,
1083,
261,
31969,
1885,
13170,
71,
13,
309,
326,
11148,
813,
434,
326,
3142,
353,
315,
27862,
6,
326,
279... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1312,
614,
13788,
24686,
27055,
71,
8583,
12,
9812,
16,
27966,
4672,
3536,
2000,
1083,
261,
31969,
1885,
13170,
71,
13,
309,
326,
11148,
813,
434,
326,
3142,
353,
315,
27862,
6,
326,
279... |
encoding= 'utf-8' return encoding | '''good as any''' return 'utf-8' | def hardCoded (self, page): # good as any encoding= 'utf-8' | 4806be63c3fd4332aab2afc601356d226d51686c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5759/4806be63c3fd4332aab2afc601356d226d51686c/url.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7877,
1085,
72,
261,
2890,
16,
1363,
4672,
468,
7494,
487,
1281,
2688,
33,
296,
3158,
17,
28,
11,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7877,
1085,
72,
261,
2890,
16,
1363,
4672,
468,
7494,
487,
1281,
2688,
33,
296,
3158,
17,
28,
11,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
if not os.path.exists(conf): | if new_conf: | def save_config(args, config): conf = DEFAULT_CONF if args.config: conf = args.config logging.debug("Saving configuration to %s" % conf) confdir = os.path.dirname(conf) if confdir and not os.path.exists(confdir): loggin.info("Config directory does not exist: %s" % confdir) os.mkdirs(confdir) f = open(conf, 'w') yaml.dump(f, config) f.close() if not os.path.exists(conf): logging.info("Adding new configuration file to vcs: %s" % conf) cmd = config['vcs_add'] if 'vcs_add' in config else DARCS_ADD check_call(shlex.split(cmd % {'file': conf})) | a7e08cef125a289cadd5348fcf3aaabad28d18d4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5132/a7e08cef125a289cadd5348fcf3aaabad28d18d4/musdex.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1923,
67,
1425,
12,
1968,
16,
642,
4672,
2195,
273,
3331,
67,
14497,
309,
833,
18,
1425,
30,
2195,
273,
833,
18,
1425,
2907,
18,
4148,
2932,
24660,
1664,
358,
738,
87,
6,
738,
2195,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1923,
67,
1425,
12,
1968,
16,
642,
4672,
2195,
273,
3331,
67,
14497,
309,
833,
18,
1425,
30,
2195,
273,
833,
18,
1425,
2907,
18,
4148,
2932,
24660,
1664,
358,
738,
87,
6,
738,
2195,
... |
""" SMTP 'help' command. Returns help text from server """ | """ SMTP 'help' command. Returns help text from server """ | def help(self): """ SMTP 'help' command. Returns help text from server """ | e73ba6c2cf417ec058a05564b3e178c42b6b9b58 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/e73ba6c2cf417ec058a05564b3e178c42b6b9b58/smtplib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2809,
12,
2890,
4672,
3536,
18102,
296,
5201,
11,
1296,
18,
2860,
2809,
977,
628,
1438,
3536,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2809,
12,
2890,
4672,
3536,
18102,
296,
5201,
11,
1296,
18,
2860,
2809,
977,
628,
1438,
3536,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
scopeId = self._getNextScopeId() | def removeInterest(self, handle, event=None, auto=False): """ Stop looking in a (set of) zone(s) """ | 1d7c9f7e76af06d8da3ff3abcb9e944ed12b0889 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8543/1d7c9f7e76af06d8da3ff3abcb9e944ed12b0889/DoInterestManager.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1206,
29281,
12,
2890,
16,
1640,
16,
871,
33,
7036,
16,
3656,
33,
8381,
4672,
3536,
5131,
7849,
316,
279,
261,
542,
434,
13,
4157,
12,
87,
13,
3536,
2,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1206,
29281,
12,
2890,
16,
1640,
16,
871,
33,
7036,
16,
3656,
33,
8381,
4672,
3536,
5131,
7849,
316,
279,
261,
542,
434,
13,
4157,
12,
87,
13,
3536,
2,
-100,
-100,
-100,
-100,
-100,
... | |
else | else: | def __init__(data = None) if data == None: quickfix.IntField.__init__(self, 665) else quickfix.IntField.__init__(self, 665, data) | 484890147d4b23aac4b9d0e85e84fceab7e137c3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8819/484890147d4b23aac4b9d0e85e84fceab7e137c3/quickfix_fields.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
892,
273,
599,
13,
309,
501,
422,
599,
30,
9549,
904,
18,
1702,
974,
16186,
2738,
972,
12,
2890,
16,
1666,
9222,
13,
469,
30,
9549,
904,
18,
1702,
974,
16186,
27... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
892,
273,
599,
13,
309,
501,
422,
599,
30,
9549,
904,
18,
1702,
974,
16186,
2738,
972,
12,
2890,
16,
1666,
9222,
13,
469,
30,
9549,
904,
18,
1702,
974,
16186,
27... |
log.write("exit code=%s\n" % e.args) | if e.args == 0: err = '' else: err = '***Error ' log.write("%sexit code=%s\n" % (err, e.args)) | def doPerformanceTests(hardhatScript, mode, workingDir, outputDir, buildVersion, log): chandlerDir = os.path.join(workingDir, "chandler") testDir = os.path.join(chandlerDir, 'tools', 'QATestScripts', 'Performance') logDir = os.path.join(chandlerDir, 'test_profile') chandlerLog = os.path.join(logDir, 'chandler.log') FuncTestLog = os.path.join(logDir, 'FunctionalTestSuite.log') if buildenv['version'] == 'debug': python = buildenv['python_d'] pythonOpts = '' elif buildenv['version'] == 'release': python = buildenv['python'] pythonOpts = '-O' os.chdir(chandlerDir) result = 'success' testFiles = glob.glob(os.path.join(testDir, 'Perf*.py')) for testFile in testFiles: args = [python, pythonOpts, os.path.join(chandlerDir, 'Chandler.py'), '--create', '--profileDir=%s' % logDir, '--scriptFile=%s' % testFile] try: try: os.remove('chandler.log') except OSError: pass outputlist = hardhatutil.executeCommandReturnOutput(args) # Hack: find if any line contains '#TINDERBOX# Status = FAILED' and # if so raise the exception to signal test failure for line in outputList: if line.find('#TINDERBOX# Status = FAIL') >= 0 or \ line.find('#TINDERBOX# Status = UNCHECKED') >= 0: raise hardhatutil.ExternalCommandErrorWithOutputList([0, outputList]) hardhatutil.dumpOutputList(outputlist, log) except hardhatutil.ExternalCommandErrorWithOutputList, e: print "perf tests failed", e log.write("***Error during performance tests***\n") log.write("Test log:\n") hardhatutil.dumpOutputList(e.outputList, log) log.write("exit code=%s\n" % e.args) log.write("NOTE: If the tests themselves passed but the exit code\n") log.write(" reports failure, it means a shutdown problem.\n") log.write("chandler.log:\n") try: CopyLog(chandlerLog, log) except: pass forceBuildNextCycle(log, workingDir) return "test_failed" except Exception, e: print "perf tests failed", e log.write("***Error during performance tests***\n") log.write("Exception:\n") log.write(str(e) + "\n") forceBuildNextCycle(log, workingDir) return "test_failed" else: log.write("Performance tests exit code=0\n") return result | 7c8a3c8d36a170f993438c7ccb3157b64b91acb9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/7c8a3c8d36a170f993438c7ccb3157b64b91acb9/newchandler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
25024,
14650,
12,
20379,
11304,
3651,
16,
1965,
16,
5960,
1621,
16,
21412,
16,
1361,
1444,
16,
613,
4672,
462,
464,
749,
1621,
273,
1140,
18,
803,
18,
5701,
12,
20478,
1621,
16,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
25024,
14650,
12,
20379,
11304,
3651,
16,
1965,
16,
5960,
1621,
16,
21412,
16,
1361,
1444,
16,
613,
4672,
462,
464,
749,
1621,
273,
1140,
18,
803,
18,
5701,
12,
20478,
1621,
16,
3... |
if 0 and hasattr(object, '__all__'): visible = lambda key, all=object.__all__: key in all else: visible = lambda key: key[:1] != '_' | def docmodule(self, object, name=None): """Produce HTML documentation for a module object.""" name = object.__name__ # ignore the passed-in name parts = split(name, '.') links = [] for i in range(len(parts)-1): links.append( '<a href="%s.html"><font color="#ffffff">%s</font></a>' % (join(parts[:i+1], '.'), parts[i])) linkedname = join(links + parts[-1:], '.') head = '<big><big><strong>%s</strong></big></big>' % linkedname try: path = inspect.getabsfile(object) filelink = '<a href="file:%s">%s</a>' % (path, path) except TypeError: filelink = '(built-in)' info = [] if hasattr(object, '__version__'): version = str(object.__version__) if version[:11] == '$' + 'Revision: ' and version[-1:] == '$': version = strip(version[11:-1]) info.append('version %s' % self.escape(version)) if hasattr(object, '__date__'): info.append(self.escape(str(object.__date__))) if info: head = head + ' (%s)' % join(info, ', ') result = self.heading( head, '#ffffff', '#7799ee', '<a href=".">index</a><br>' + filelink) | 5a804edd3cacf2a790ff3c872adad22f95e7a604 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/5a804edd3cacf2a790ff3c872adad22f95e7a604/pydoc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
997,
2978,
12,
2890,
16,
733,
16,
508,
33,
7036,
4672,
3536,
25884,
3982,
7323,
364,
279,
1605,
733,
12123,
508,
273,
733,
16186,
529,
972,
468,
2305,
326,
2275,
17,
267,
508,
2140,
27... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
997,
2978,
12,
2890,
16,
733,
16,
508,
33,
7036,
4672,
3536,
25884,
3982,
7323,
364,
279,
1605,
733,
12123,
508,
273,
733,
16186,
529,
972,
468,
2305,
326,
2275,
17,
267,
508,
2140,
27... | |
elif passwd: | elif login and passwd: | def cmd_notAuthorized(self, request, response, form, objs, user): login, passwd, perm_login = [form.getfirst(name) for name in \ ('login', 'passwd', 'perm_login')] | 9a63f4b640cb3e156691a5026f514d786952821b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1871/9a63f4b640cb3e156691a5026f514d786952821b/qSecurity.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1797,
67,
902,
15341,
12,
2890,
16,
590,
16,
766,
16,
646,
16,
12721,
16,
729,
4672,
3925,
16,
19093,
16,
4641,
67,
5819,
273,
306,
687,
18,
588,
3645,
12,
529,
13,
364,
508,
316,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1797,
67,
902,
15341,
12,
2890,
16,
590,
16,
766,
16,
646,
16,
12721,
16,
729,
4672,
3925,
16,
19093,
16,
4641,
67,
5819,
273,
306,
687,
18,
588,
3645,
12,
529,
13,
364,
508,
316,
... |
student_code_thread = Student_code_runner(vessels[0], filename,student_file_data) | student_code_thread = Student_code_runner(vessels[0],student_file_name, student_file_data) | def run_webserver_tests(filename,name_ip_tuples): # get a tuple for each test [(filename,file_content_str)...] test_tuples = get_tests() # get the student file to grade file_obj = open(TO_GRADE_DIR+"/"+filename,'r') student_file_data = file_obj.read() file_obj.close() #get list of hosts host_list =[] for (host,ip) in name_ip_tuples: host_list.append(host) print "initializing vessels" # initialize seattle nodes on each host try: is_init, vessels = install_autograder.initialize(host_list, "autograder") except Exception: clean_up() raise else: if not is_init: raise Exception, 'failed to intialize seattle nodes '+str(vessels) #create a thread to run student code student_code_thread = Student_code_runner(vessels[0], filename,student_file_data) #TODO, reset all nodes after each test webserver_host_name, webserver_ip = name_ip_tuples[0] for (test_name,test_data) in test_tuples: # run the students webserver student_code_thread.run() testargs =test_name+" "+webserver_ip+"63173" print "running test "+test_name+" against "+filename # run the test try: test_running,error = install_autograder.run_target(vessels[1], test_name,test_data,testargs) except Exception: clean_up() raise else: if not test_running: clean_up() raise Exception, "test code "+test_name+"failed to start\n"+error # get vessel logs try: #server_log = install_autograder.showlog_vessel(vessels[0]) test_log = install_autograder.showlog_vessel(vessels[1]) except Exception: clean_up() raise # stop the repy code install_autograder.stop_target(vessel[1]) student_code_thread.stop() results = score(test_log,test_name) save_grade(results,filename) #reset the nodes # mv file to the graded directory shutil.move(TO_GRADE_DIR+"/"+filename, GRADED_DIR+"/"+filename) # clean up install_autograder.tear_down() for test_name,test_obj in test_tuples: test_obj.close() | f448ac0aa00383d8363134c56e7549a55e26dc82 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7995/f448ac0aa00383d8363134c56e7549a55e26dc82/autograder_runner.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
67,
1814,
2266,
502,
67,
16341,
12,
3459,
16,
529,
67,
625,
67,
17705,
4672,
225,
468,
336,
279,
3193,
364,
1517,
1842,
306,
12,
3459,
16,
768,
67,
1745,
67,
701,
13,
2777,
65,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
67,
1814,
2266,
502,
67,
16341,
12,
3459,
16,
529,
67,
625,
67,
17705,
4672,
225,
468,
336,
279,
3193,
364,
1517,
1842,
306,
12,
3459,
16,
768,
67,
1745,
67,
701,
13,
2777,
65,... |
self.fileTimeout = gConfig.getValue('/Resources/StorageElements/FileTimeout',30) self.filesPerCall = gConfig.getValue('/Resources/StorageElements/FilesPerCall',20) | self.fileTimeout = gConfig.getValue( '/Resources/StorageElements/FileTimeout', 30 ) self.filesPerCall = gConfig.getValue( '/Resources/StorageElements/FilesPerCall', 20 ) | def __init__(self,storageName,protocol,path,host,port,spaceToken,wspath): self.isok = True self.gfal = False self.lcg_util = False | 70e66af095cb6701e39b1e701e4a2ce4d012b4f7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/70e66af095cb6701e39b1e701e4a2ce4d012b4f7/SRM2Storage.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
5697,
461,
16,
8373,
16,
803,
16,
2564,
16,
655,
16,
2981,
1345,
16,
4749,
803,
4672,
365,
18,
291,
601,
273,
1053,
365,
18,
75,
74,
287,
273,
1083,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
5697,
461,
16,
8373,
16,
803,
16,
2564,
16,
655,
16,
2981,
1345,
16,
4749,
803,
4672,
365,
18,
291,
601,
273,
1053,
365,
18,
75,
74,
287,
273,
1083,
... |
t=apsw.VFSFile("", "testfile", [apsw.SQLITE_OPEN_CREATE|apsw.SQLITE_OPEN_READWRITE,0]) | t=apsw.VFSFile("", "testfile", [apsw.SQLITE_OPEN_MAIN_DB|apsw.SQLITE_OPEN_CREATE|apsw.SQLITE_OPEN_READWRITE,0]) | def foo(): t=apsw.VFSFile("", "testfile", [apsw.SQLITE_OPEN_CREATE|apsw.SQLITE_OPEN_READWRITE,0]) apsw.faultdict["xCloseFails"]=True del t gc.collect() | f0f139ca1c73176ffc374f8eb6f63ed6e2cd34b3 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/993/f0f139ca1c73176ffc374f8eb6f63ed6e2cd34b3/tests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8431,
13332,
268,
33,
438,
5328,
18,
58,
4931,
812,
2932,
3113,
315,
3813,
768,
3113,
306,
438,
5328,
18,
3997,
5398,
67,
11437,
67,
19803,
67,
2290,
96,
438,
5328,
18,
3997,
5398,
67,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8431,
13332,
268,
33,
438,
5328,
18,
58,
4931,
812,
2932,
3113,
315,
3813,
768,
3113,
306,
438,
5328,
18,
3997,
5398,
67,
11437,
67,
19803,
67,
2290,
96,
438,
5328,
18,
3997,
5398,
67,... |
_format_results(report, keyval) | self.format_results(report, keyval) | def execute(self, iterations = 1, args = ''): vars = 'TMPDIR=\"%s\" RESULTDIR=\"%s\"' % (self.tmpdir, self.resultsdir) profilers = self.job.profilers keyval = open(self.resultsdir + '/keyval', 'w') | f6b475aee14766a1d794cb5f0587e1b784504429 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/10349/f6b475aee14766a1d794cb5f0587e1b784504429/unixbench.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1836,
12,
2890,
16,
11316,
273,
404,
16,
833,
273,
875,
4672,
4153,
273,
296,
28259,
4537,
5189,
9,
87,
2412,
17210,
4537,
5189,
9,
87,
2412,
11,
738,
261,
2890,
18,
5645,
1214,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1836,
12,
2890,
16,
11316,
273,
404,
16,
833,
273,
875,
4672,
4153,
273,
296,
28259,
4537,
5189,
9,
87,
2412,
17210,
4537,
5189,
9,
87,
2412,
11,
738,
261,
2890,
18,
5645,
1214,
16,
... |
self.elementTypeInfo = eltype.getdef.elementTypeInfo self.pointerToSh = eltype.getdef.pointerToSh self.pointerToShj = eltype.getdef.pointerToShj self.pointerToGauss = eltype.getdef.pointerToGauss self.numberVolumeElementNodes = eltype.getdef.numberElementNodes self.numberVolumeElementGaussPoints = eltype.getdef.numberElementGaussPoints self.numberVolumeElementEquations = eltype.getdef.numberElementEquations | self.elementTypeInfo = eltype.elementTypeInfo self.pointerToSh = eltype.pointerToSh self.pointerToShj = eltype.pointerToShj self.pointerToGauss = eltype.pointerToGauss self.numberVolumeElementNodes = eltype.numberVolumeElementNodes self.numberVolumeElementGaussPoints = eltype.numberVolumeElementGaussPoints self.numberVolumeElementEquations = eltype.numberVolumeElementEquations self.connectivitySize = self.numberVolumeElements*self.numberVolumeElementNodes | self.pointerToLgdef = None | fd3da5a98efec13d16ce111ad2d0286cfd326cfa /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8645/fd3da5a98efec13d16ce111ad2d0286cfd326cfa/Lithomop3d_setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
365,
18,
10437,
774,
48,
75,
536,
273,
599,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
365,
18,
10437,
774,
48,
75,
536,
273,
599,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
return "User" | self.state = "User" | def _ebLogin(self, failure): self.transport.write("Authentication failed\n") self.transport.write("Username: ") return "User" | 968ceb0acb1314669023ba3a3923149b7aa11bd1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12595/968ceb0acb1314669023ba3a3923149b7aa11bd1/telnet.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
24008,
5358,
12,
2890,
16,
5166,
4672,
365,
18,
13049,
18,
2626,
2932,
6492,
2535,
64,
82,
7923,
365,
18,
13049,
18,
2626,
2932,
8575,
30,
9369,
327,
315,
1299,
6,
2,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
24008,
5358,
12,
2890,
16,
5166,
4672,
365,
18,
13049,
18,
2626,
2932,
6492,
2535,
64,
82,
7923,
365,
18,
13049,
18,
2626,
2932,
8575,
30,
9369,
327,
315,
1299,
6,
2,
-100,
-100,
... |
else: boto.log.warning('No SimpleDB domain set, persistance is disabled') | else: boto.log.warning('No SimpleDB domain set, persistance is disabled') | def __init__(self, domain_name=None, aws_access_key_id=None, aws_secret_access_key=None, debug=0): self.domain_name = domain_name self.aws_access_key_id = aws_access_key_id self.aws_secret_access_key = aws_secret_access_key self.domain = None self.sdb = None self.s3 = None if not self.domain_name: self.domain_name = self.DefaultDomainName if self.domain_name: boto.log.info('No SimpleDB domain set, using default_domain: %s' % self.domain_name) else: boto.log.warning('No SimpleDB domain set, persistance is disabled') if self.domain_name: self.sdb = boto.connect_sdb(aws_access_key_id=self.aws_access_key_id, aws_secret_access_key=self.aws_secret_access_key, debug=debug) self.domain = self.sdb.lookup(self.domain_name) if not self.domain: self.domain = self.sdb.create_domain(self.domain_name) | 352230c26fd25891703b58fc69769a3e6c36b17f /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1098/352230c26fd25891703b58fc69769a3e6c36b17f/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
2461,
67,
529,
33,
7036,
16,
2287,
67,
3860,
67,
856,
67,
350,
33,
7036,
16,
2287,
67,
5875,
67,
3860,
67,
856,
33,
7036,
16,
1198,
33,
20,
4672,
365... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
2461,
67,
529,
33,
7036,
16,
2287,
67,
3860,
67,
856,
67,
350,
33,
7036,
16,
2287,
67,
5875,
67,
3860,
67,
856,
33,
7036,
16,
1198,
33,
20,
4672,
365... |
title = lines.pop(0) | title = lines.pop(0).strip() | def blosxom_entry_parser(filename, request): """ Open up a *.txt file and read its contents. The first line becomes the title of the entry. The other lines are the body of the entry. @param filename: A filename to extract data and metadata from @type filename: string @param request: A standard request object @type request: L{Pyblosxom.pyblosxom.Request} object @returns: A dict containing parsed data and meta data with the particular file (and plugin) @rtype: dict """ config = request.getConfiguration() entryData = {} lines = open(filename).readlines() # the file has nothing in it... so we're going to return # a blank entry data object. if len(lines) == 0: return { "title": "", "body": "" } # NOTE: you can probably use the next bunch of lines verbatim # for all entryparser plugins. this pulls the first line off as # the title, the next bunch of lines that start with # as # metadata lines, and then everything after that is the body # of the entry. title = lines.pop(0) entryData['title'] = title # absorb meta data lines which begin with a # while lines and lines[0].startswith("#"): meta = lines.pop(0) meta = meta[1:].strip() # remove the hash meta = meta.split(" ", 2) entryData[meta[0]] = meta[1] # Call the preformat function args = {'parser': entryData.get('parser', config.get('parser', 'plain')), 'story': lines, 'request': request} entryData['body'] = tools.run_callback('preformat', args, donefunc = lambda x:x != None, defaultfunc = lambda x: ''.join(x['story'])) # Call the postformat callbacks tools.run_callback('postformat', {'request': request, 'entry_data': entryData}) return entryData | b29e819b89b4ef1e254f34f68090a27cadd2b7b9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11836/b29e819b89b4ef1e254f34f68090a27cadd2b7b9/pyblosxom.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
324,
383,
30319,
362,
67,
4099,
67,
4288,
12,
3459,
16,
590,
4672,
3536,
3502,
731,
279,
380,
18,
5830,
585,
471,
855,
2097,
2939,
18,
225,
1021,
1122,
980,
12724,
326,
2077,
434,
326,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
324,
383,
30319,
362,
67,
4099,
67,
4288,
12,
3459,
16,
590,
4672,
3536,
3502,
731,
279,
380,
18,
5830,
585,
471,
855,
2097,
2939,
18,
225,
1021,
1122,
980,
12724,
326,
2077,
434,
326,... |
secsRemaining = | secsRemaining = timerTime-passesdTime | def _updateTimerCb( self ): nowTime = time.time() passedTime = nowTime - self.timerStartTime print( "updateTimer, passed:", passedTime ) | a2ca8daca04eab073f1d77b9c429e2a44ce76e17 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/6451/a2ca8daca04eab073f1d77b9c429e2a44ce76e17/ui.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2725,
6777,
15237,
12,
365,
262,
30,
2037,
950,
273,
813,
18,
957,
1435,
2275,
950,
273,
2037,
950,
300,
365,
18,
12542,
13649,
1172,
12,
315,
2725,
6777,
16,
2275,
2773,
16,
2275... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2725,
6777,
15237,
12,
365,
262,
30,
2037,
950,
273,
813,
18,
957,
1435,
2275,
950,
273,
2037,
950,
300,
365,
18,
12542,
13649,
1172,
12,
315,
2725,
6777,
16,
2275,
2773,
16,
2275... |
P.legend(tnames) | P.legend([r"$t=" + str(ts) + "$" for ts in self.tresult[tindex]]) | def plotkcrosssection(self, tindex=None, fig=None, show=True, varindex=None, klist=None, kfunction=None, saveplot=False): """Plot results for different ks in 3d plot. Can only plot a single variable at a time.""" #Test whether model has run yet if self.runcount == 0: raise ModelError("Model has not been run yet, cannot plot results!") #Test whether model has k variable dependence try: self.yresult[0,0,0] #Does this exist? except IndexError, er: raise ModelError("This model does not have any k variable to plot! Got " + er.message) if varindex is None: varindex = 0 #Set variable to plot if klist is None: klist = N.arange(len(self.k)) #Plot all ks if tindex is None: tindex = N.arange(0,len(self.tresult), 1000) #Selection of time slices #Set names for t slices tnames = str(self.tresult[tindex]) if fig is None: fig = P.figure() #Create figure else: P.figure(fig.number) #Plot figure, default is semilogx for k P.semilogx(self.k[klist], self.yresult[tindex,varindex,klist].transpose()) #Create legends and axis names P.xlabel(r"$k$") P.legend(tnames) #Should we show it now or just return it without showing? if show: P.show() #Should we save the plot somewhere? if saveplot: self.saveplot(fig) return fig | feb4e34368105dfcc1fe5b7e8ded99dd0e6b64e2 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7283/feb4e34368105dfcc1fe5b7e8ded99dd0e6b64e2/cosmomodels.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3207,
79,
14653,
3464,
12,
2890,
16,
268,
1615,
33,
7036,
16,
4291,
33,
7036,
16,
2405,
33,
5510,
16,
569,
1615,
33,
7036,
16,
417,
1098,
33,
7036,
16,
417,
915,
33,
7036,
16,
1923,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3207,
79,
14653,
3464,
12,
2890,
16,
268,
1615,
33,
7036,
16,
4291,
33,
7036,
16,
2405,
33,
5510,
16,
569,
1615,
33,
7036,
16,
417,
1098,
33,
7036,
16,
417,
915,
33,
7036,
16,
1923,
... |
memory=ram, os=os, vcpus=vcpus, | os=os, vcpus=vcpus, | def create(request, cluster_slug=None): """ Create a new instance Store in DB and Create on given cluster """ user = request.user if not(user.is_superuser or user.perms_on_any(Cluster, ['admin', 'create_vm'])): return render_403(request, 'You do not have permission to create virtual \ machines') if cluster_slug is not None: cluster = get_object_or_404(Cluster, slug=cluster_slug) else: cluster = None if request.method == 'POST': form = NewVirtualMachineForm(user, None, request.POST) if form.is_valid(): data = form.cleaned_data owner = data['owner'] cluster = data['cluster'] hostname = data['hostname'] disk_template = data['disk_template'] # Default to not pass in pnode and snode # since these will be set if the form is correct pnode = None snode = None os = data['os'] name_check = data['name_check'] iallocator = data['iallocator'] # Hidden fields iallocator_hostname = None if 'iallocator_hostname' in data: iallocator_hostname = data['iallocator_hostname'] # BEPARAMS vcpus = data['vcpus'] disk_size = data['disk_size'] ram = data['ram'] nicmode = data['nicmode'] nictype = data['nictype'] # HVPARAMS kernelpath = data['kernelpath'] rootpath = data['rootpath'] serialconsole = data['serialconsole'] bootorder = data['bootorder'] imagepath = data['imagepath'] # If iallocator was not checked do not pass in the iallocator # name. If iallocator was checked don't pass snode,pnode. if not iallocator: iallocator_hostname = None pnode = data['pnode'] # If drbd is being used assign the secondary node if disk_template == 'drbd' and pnode is not None: snode = data['snode'] try: job_id = cluster.rapi.CreateInstance('create', hostname, disk_template, [{"size": disk_size, }],[{nicmode: nictype, }], memory=ram, os=os, vcpus=vcpus, pnode=pnode, snode=snode, name_check=name_check, ip_check=name_check, iallocator=iallocator_hostname, hvparams={'kernel_path': kernelpath, \ 'root_path': rootpath, \ 'serial_console':serialconsole, \ 'boot_order':bootorder, \ 'cdrom_image_path':imagepath}) # Wait for job to process as the error will not happen # right away sleep(2) jobstatus = cluster.rapi.GetJobStatus(job_id) # raise an exception if there was an error in the job if jobstatus["status"] == 'error': raise GanetiApiError(jobstatus["opresult"]) vm = VirtualMachine(cluster=cluster, owner=owner, hostname=hostname, disk_size=disk_size, ram=ram, virtual_cpus=vcpus) vm.ignore_cache = True vm.save() job = Job.objects.create(job_id=job_id, obj=vm, cluster=cluster) VirtualMachine.objects.filter(id=vm.id).update(last_job=job) # log information about creating the machine log_action(user, vm, "created") # grant admin permissions to the owner data['grantee'].grant('admin', vm) return HttpResponseRedirect( \ reverse('instance-detail', args=[cluster.slug, vm.hostname])) except GanetiApiError, e: msg = 'Error creating virtual machine on this cluster: %s' % e form._errors["cluster"] = form.error_class([msg]) elif request.method == 'GET': form = NewVirtualMachineForm(user, cluster) return render_to_response('virtual_machine/create.html', { 'form': form }, context_instance=RequestContext(request), ) | 54130827bcd3073cf49bad837dfe0705307f5ebe /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10509/54130827bcd3073cf49bad837dfe0705307f5ebe/virtual_machine.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
12,
2293,
16,
2855,
67,
6436,
33,
7036,
4672,
3536,
1788,
279,
394,
791,
4994,
316,
2383,
471,
1788,
603,
864,
2855,
3536,
729,
273,
590,
18,
1355,
309,
486,
12,
1355,
18,
291,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
12,
2293,
16,
2855,
67,
6436,
33,
7036,
4672,
3536,
1788,
279,
394,
791,
4994,
316,
2383,
471,
1788,
603,
864,
2855,
3536,
729,
273,
590,
18,
1355,
309,
486,
12,
1355,
18,
291,
... |
f.seek(0) f.truncate(1) expect(f.tell(), 0) expect(len(f.read()), 1) | f.seek(0) f.truncate(1) expect(f.tell(), 0) expect(len(f.read()), 1) | def expect(got_this, expect_this): if test_support.verbose: print '%r =?= %r ...' % (got_this, expect_this), if got_this != expect_this: if test_support.verbose: print 'no' raise test_support.TestFailed, 'got %r, but expected %r' %\ (got_this, expect_this) else: if test_support.verbose: print 'yes' | e9e1243574bc55762c3c6470699e478f1fdc056d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/e9e1243574bc55762c3c6470699e478f1fdc056d/test_largefile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4489,
12,
13212,
67,
2211,
16,
4489,
67,
2211,
4672,
309,
1842,
67,
13261,
18,
11369,
30,
1172,
1995,
86,
273,
35,
33,
738,
86,
25956,
738,
261,
13212,
67,
2211,
16,
4489,
67,
2211,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4489,
12,
13212,
67,
2211,
16,
4489,
67,
2211,
4672,
309,
1842,
67,
13261,
18,
11369,
30,
1172,
1995,
86,
273,
35,
33,
738,
86,
25956,
738,
261,
13212,
67,
2211,
16,
4489,
67,
2211,
... |
Make some modifications to list, as part of parsing a 'subvariants' block. @param list: List that will be processed @param name: Name that will be prepended to the dictionary name @param dep_list: List of dependencies to be added to the list dictionaries @param add_to_shortname: Whether we'll add a shortname parameter to the dictionaries. | Make some modifications to list, as part of parsing a 'subvariants' block. @param list: List to be processed @param name: Name to be appended to the dictionary's 'name' key @param dep_list: List of dependencies to be added to the dictionary's 'depend' key @param add_to_shortname: Boolean indicating whether name should be appended to the dictionary's 'shortname' as well | def __modify_list_subvariants(self, list, name, dep_list, add_to_shortname): """ Make some modifications to list, as part of parsing a 'subvariants' block. | 5cfc252eb7878a541fecc40b0cbb270db39d4d08 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12268/5cfc252eb7878a541fecc40b0cbb270db39d4d08/kvm_config.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
17042,
67,
1098,
67,
1717,
15886,
12,
2890,
16,
666,
16,
508,
16,
5993,
67,
1098,
16,
527,
67,
869,
67,
28650,
4672,
3536,
4344,
2690,
17953,
358,
666,
16,
487,
1087,
434,
5811,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
17042,
67,
1098,
67,
1717,
15886,
12,
2890,
16,
666,
16,
508,
16,
5993,
67,
1098,
16,
527,
67,
869,
67,
28650,
4672,
3536,
4344,
2690,
17953,
358,
666,
16,
487,
1087,
434,
5811,
... |
raise MDError, "No Package ID found for package %s, not going to add it" % po | raise MDError, "No Package ID found for package %s, not going to" \ " add it" % po | def read_in_package(self, rpmfile, pkgpath=None, reldir=None): """rpmfile == relative path to file from self.packge_dir""" remote_package = False baseurl = self.conf.baseurl | 473cb91d36b5c88e3a55fa60404eb172a1c6e32b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9557/473cb91d36b5c88e3a55fa60404eb172a1c6e32b/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
855,
67,
267,
67,
5610,
12,
2890,
16,
25228,
768,
16,
3475,
803,
33,
7036,
16,
1279,
1214,
33,
7036,
4672,
3536,
86,
7755,
768,
422,
3632,
589,
358,
585,
628,
365,
18,
2920,
908,
67,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
855,
67,
267,
67,
5610,
12,
2890,
16,
25228,
768,
16,
3475,
803,
33,
7036,
16,
1279,
1214,
33,
7036,
4672,
3536,
86,
7755,
768,
422,
3632,
589,
358,
585,
628,
365,
18,
2920,
908,
67,... |
+ headwordTrad[i] + "' to '" + entity \ | + entry.HeadwordTraditional[i] + "' to '" + entity \ | def checkCharacterReading(entryList, readingName, readingOptions={}, ignoreFifthTone=False): for entry in entryList: headwordTrad, headwordSimp, reading = entry if headwordTrad != headwordSimp: headword = headwordTrad + "'/'" + headwordSimp else: headword = headwordTrad try: operator = getReadingOperator(readingName, readingOptions) entities = operator.decompose(reading) except exception.DecompositionError: print ("WARNING: can't parse line '" \ + headwordTrad + "', '" + headwordSimp + "', '" + reading \ + "'").encode(default_encoding) continue entitiesFiltered = [] for entity in entities: if re.match(r"\s+$", entity): continue entitiesFiltered.append(entity) if (len(entitiesFiltered) != len(headwordTrad)) \ or (len(entitiesFiltered) != len(headwordSimp)): print ("WARNING: can't parse line '" \ + headwordTrad + "', '" + headwordSimp + "', '" + reading \ + "'").encode(default_encoding) continue for i, entity in enumerate(entitiesFiltered): if getReadingOperator(readingName).isReadingEntity(entity): if headwordTrad[i] != headwordSimp[i]: charList = [headwordTrad[i], headwordSimp[i]] else: charList = [headwordTrad[i]] for char in charList: validReading = True try: readingList = getCJK().getReadingForCharacter(char, readingName, **readingOptions) if not hasReading(entity, readingList, readingName, ignoreFifthTone): print (char + " " + entity + ", known readings: " \ + ', '.join(readingList) + "; for headword '" \ + headword + "'").encode(default_encoding) except exception.NoInformationError: pass else: # Check mapping of non-Pinyin entities. They either map # to the same character again (e.g. ellipsis: ...) or have # a different form described by table NON_PINYIN_MAPPING if headwordTrad[i] != entity \ and (headwordTrad[i] not in NON_PINYIN_MAPPING \ or NON_PINYIN_MAPPING[headwordTrad[i]] != entity): print ("WARNING: invalid mapping of entity '" \ + headwordTrad[i] + "' to '" + entity \ + "'; for headword '" + headword + "'")\ .encode(default_encoding) elif headwordSimp[i] != entity \ and (headwordSimp[i] not in NON_PINYIN_MAPPING \ or NON_PINYIN_MAPPING[headwordSimp[i]] != entity): print ("WARNING: invalid mapping of entity '" \ + headwordSimp[i] + "' to '" + entity \ + "'; for headword '" + headword + "'")\ .encode(default_encoding) | 03ee2902c9101f1e847d6ed9b78fc134a514e71a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11128/03ee2902c9101f1e847d6ed9b78fc134a514e71a/checkcedict.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
7069,
15714,
12,
4099,
682,
16,
6453,
461,
16,
6453,
1320,
28793,
2305,
42,
430,
451,
56,
476,
33,
8381,
4672,
364,
1241,
316,
1241,
682,
30,
910,
1095,
1609,
72,
16,
910,
1095,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
7069,
15714,
12,
4099,
682,
16,
6453,
461,
16,
6453,
1320,
28793,
2305,
42,
430,
451,
56,
476,
33,
8381,
4672,
364,
1241,
316,
1241,
682,
30,
910,
1095,
1609,
72,
16,
910,
1095,
... |
n = wd + self.firstDayOfWeek + 1 | n = wd + self.firstDayOfWeek - 1 | def DrawMonth(self, dc, startDate, y, highlightDate = False): """ draw a single month return the updated value of y """ dc.SetTextForeground(wx.BLACK); | b1890ae729cf70fa2a0da432c9780afb8031468e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/b1890ae729cf70fa2a0da432c9780afb8031468e/minical.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10184,
5445,
12,
2890,
16,
6744,
16,
12572,
16,
677,
16,
8839,
1626,
273,
1083,
4672,
3536,
3724,
279,
2202,
3138,
327,
326,
3526,
460,
434,
677,
3536,
6744,
18,
694,
1528,
23206,
12,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10184,
5445,
12,
2890,
16,
6744,
16,
12572,
16,
677,
16,
8839,
1626,
273,
1083,
4672,
3536,
3724,
279,
2202,
3138,
327,
326,
3526,
460,
434,
677,
3536,
6744,
18,
694,
1528,
23206,
12,
... |
data = Numeric.array(data) length, dof = Numeric.shape(data) | data = Numeric.array(data) length, dof = Numeric.shape(data) | def spline(data, quality): """Return an interpolated trajectory from data. The returned value will be either an AMC object or a Numeric array, depending on what type the original data is. quality is the number of interpolated points to insert between each point in the initial data. """ # Special case: The input data is an AMC object. For each bone in the AMC # object create a spline. Return an AMC object. if data.__class__ == AMC: interpolated = AMC() for bone, motion in data.bones.iteritems(): interpolated.bones[bone] = spline(motion, quality) return interpolated data = Numeric.array(data) length, dof = Numeric.shape(data) interpolated = Numeric.empty((length * quality, dof)) # This is the range of times we'll be using for the vast majority of the # splining process. times = Numeric.arange(2, 3, 1. / quality) # This function is used to generate the intermediate points from the # constants and the time. f = lambda c: lambda t: c[0] + c[1] * t + c[2] * t**2 + c[3] * t**3 for frame in range(length - 3): # Generate matrices and solve for the constants for this section of the # data. A, b = __getMatrix(data[frame:frame + 4], dof) Ainv = inverse(A) z = [Numeric.matrixmultiply(Ainv, x) for x in b] for degree in dof: # Special case: At the beginning of the trajectory or the end we use # the beginning or end of the spline to interpolate. Normally we # only use the middle interval of the spline to interpolate. if frame == 0: interpolated[frame:frame + quality, degree] = map(f(z[dof][:4]), Numeric.arange(1, 2, 1. / quality)) elif frame == length - 4: interpolated[frame, degree] = map(f(z[dof][-4:]), Numeric.arange(3, 4, 1. / quality)) interpolated[frame, degree] = map(f(z[dof][4:8]), times) return interpolated | 698295475485f542abc417ee5d41526ef721d04a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6757/698295475485f542abc417ee5d41526ef721d04a/Interpolate.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
22826,
12,
892,
16,
9312,
4672,
3536,
990,
392,
24985,
27016,
628,
501,
18,
1021,
2106,
460,
903,
506,
3344,
392,
432,
20022,
733,
578,
279,
16980,
526,
16,
8353,
603,
4121,
618,
326,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
22826,
12,
892,
16,
9312,
4672,
3536,
990,
392,
24985,
27016,
628,
501,
18,
1021,
2106,
460,
903,
506,
3344,
392,
432,
20022,
733,
578,
279,
16980,
526,
16,
8353,
603,
4121,
618,
326,
... |
simple.GL_ATTRIB_STACK_DEPTH : (1,), simple.GL_AUTO_NORMAL : (1,), simple.GL_AUX_BUFFERS : (1,), simple.GL_BLEND : (1,), simple.GL_BLEND_COLOR : (4,), simple.GL_BLEND_DST_ALPHA : (1,), simple.GL_BLEND_DST_RGB : (1,), simple.GL_BLEND_EQUATION_RGB : (1,), simple.GL_BLEND_EQUATION_ALPHA : (1,), simple.GL_BLEND_SRC_RGB : (1,), simple.GL_BLEND_SRC_ALPHA : (1,), simple.GL_BLUE_BIAS : (1,), simple.GL_BLUE_BITS : (1,), simple.GL_BLUE_SCALE : (1,), simple.GL_CLIENT_ACTIVE_TEXTURE : (1,), simple.GL_CLIENT_ACTIVE_TEXTURE_ARB : (1,), simple.GL_CLIENT_ATTRIB_STACK_DEPTH : (1,), simple.GL_CLIP_PLANE0 : (1,), simple.GL_CLIP_PLANE1 : (1,), simple.GL_CLIP_PLANE2 : (1,), simple.GL_CLIP_PLANE3 : (1,), simple.GL_CLIP_PLANE4 : (1,), simple.GL_CLIP_PLANE5 : (1,), simple.GL_COLOR_ARRAY : (1,), simple.GL_COLOR_ARRAY_BUFFER_BINDING : (1,), simple.GL_COLOR_ARRAY_SIZE : (1,), simple.GL_COLOR_ARRAY_STRIDE : (1,), simple.GL_COLOR_ARRAY_TYPE : (1,), simple.GL_COLOR_CLEAR_VALUE : (4,), simple.GL_COLOR_LOGIC_OP : (1,), simple.GL_COLOR_MATERIAL : (1,), simple.GL_COLOR_MATERIAL_FACE : (1,), simple.GL_COLOR_MATERIAL_PARAMETER : (1,), simple.GL_COLOR_MATRIX : (4, 4), simple.GL_COLOR_MATRIX_STACK_DEPTH : (1,), simple.GL_COLOR_SUM : (1,), simple.GL_COLOR_TABLE : (1,), simple.GL_COLOR_WRITEMASK : (4,), | simple.GL_ATTRIB_STACK_DEPTH : (1,), simple.GL_AUTO_NORMAL : (1,), simple.GL_AUX_BUFFERS : (1,), simple.GL_BLEND : (1,), simple.GL_BLEND_COLOR : (4,), simple.GL_BLEND_DST_ALPHA : (1,), simple.GL_BLEND_DST_RGB : (1,), simple.GL_BLEND_EQUATION_RGB : (1,), simple.GL_BLEND_EQUATION_ALPHA : (1,), simple.GL_BLEND_SRC_RGB : (1,), simple.GL_BLEND_SRC_ALPHA : (1,), simple.GL_BLUE_BIAS : (1,), simple.GL_BLUE_BITS : (1,), simple.GL_BLUE_SCALE : (1,), simple.GL_CLIENT_ACTIVE_TEXTURE : (1,), simple.GL_CLIENT_ACTIVE_TEXTURE_ARB : (1,), simple.GL_CLIENT_ATTRIB_STACK_DEPTH : (1,), simple.GL_CLIP_PLANE0 : (1,), simple.GL_CLIP_PLANE1 : (1,), simple.GL_CLIP_PLANE2 : (1,), simple.GL_CLIP_PLANE3 : (1,), simple.GL_CLIP_PLANE4 : (1,), simple.GL_CLIP_PLANE5 : (1,), simple.GL_COLOR_ARRAY : (1,), simple.GL_COLOR_ARRAY_BUFFER_BINDING : (1,), simple.GL_COLOR_ARRAY_SIZE : (1,), simple.GL_COLOR_ARRAY_STRIDE : (1,), simple.GL_COLOR_ARRAY_TYPE : (1,), simple.GL_COLOR_CLEAR_VALUE : (4,), simple.GL_COLOR_LOGIC_OP : (1,), simple.GL_COLOR_MATERIAL : (1,), simple.GL_COLOR_MATERIAL_FACE : (1,), simple.GL_COLOR_MATERIAL_PARAMETER : (1,), simple.GL_COLOR_MATRIX : (4, 4), simple.GL_COLOR_MATRIX_STACK_DEPTH : (1,), simple.GL_COLOR_SUM : (1,), simple.GL_COLOR_TABLE : (1,), simple.GL_COLOR_WRITEMASK : (4,), | def glGetDoublev( pname ): "Natural writing of glGetDoublev using standard ctypes" output = c_double*sizes.get( pname ) result = output() result = platform.OpenGL.glGetDoublev( pname, byref(result) ) return Numeric.array( result ) | 779beb78d8242dd156d555119ccab07062fb4bf3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1545/779beb78d8242dd156d555119ccab07062fb4bf3/glget.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5118,
967,
5265,
90,
12,
19952,
262,
30,
315,
23779,
3766,
7410,
434,
5118,
967,
5265,
90,
1450,
4529,
6983,
6,
876,
273,
276,
67,
9056,
14,
11914,
18,
588,
12,
19952,
262,
563,
273,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5118,
967,
5265,
90,
12,
19952,
262,
30,
315,
23779,
3766,
7410,
434,
5118,
967,
5265,
90,
1450,
4529,
6983,
6,
876,
273,
276,
67,
9056,
14,
11914,
18,
588,
12,
19952,
262,
563,
273,
... |
if kk == 13 and not (event.ShiftDown() or event.HasModifiers()): | if kk == 13 and not (shiftDown or event.HasModifiers()): | def OnKeyDown(self, event): #if Preferences.handleSpecialEuropeanKeys: # self.handleSpecialEuropeanKeys(event, Preferences.euroKeysCountry) | 819b5d970b4a9a604bebc6d5927c5a6b42025791 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4325/819b5d970b4a9a604bebc6d5927c5a6b42025791/ShellEditor.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2755,
653,
4164,
12,
2890,
16,
871,
4672,
468,
430,
28310,
18,
4110,
12193,
29197,
304,
2396,
30,
468,
565,
365,
18,
4110,
12193,
29197,
304,
2396,
12,
2575,
16,
28310,
18,
73,
19321,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2755,
653,
4164,
12,
2890,
16,
871,
4672,
468,
430,
28310,
18,
4110,
12193,
29197,
304,
2396,
30,
468,
565,
365,
18,
4110,
12193,
29197,
304,
2396,
12,
2575,
16,
28310,
18,
73,
19321,
... |
self.al = AcceptLanguageType.decode("da, en-gb;q=0.8") | self.al = AcceptLanguageType.decode("pt, en-gb;q=0.8") | def setUp(self): self.al = AcceptLanguageType.decode("da, en-gb;q=0.8") | aaaaaa8e5437bf30ca25c8dc79b11640ed33dd21 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12681/aaaaaa8e5437bf30ca25c8dc79b11640ed33dd21/test_i18n.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
24292,
12,
2890,
4672,
365,
18,
287,
273,
8662,
3779,
559,
18,
3922,
2932,
2414,
16,
570,
17,
4490,
31,
85,
33,
20,
18,
28,
7923,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
24292,
12,
2890,
4672,
365,
18,
287,
273,
8662,
3779,
559,
18,
3922,
2932,
2414,
16,
570,
17,
4490,
31,
85,
33,
20,
18,
28,
7923,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
self.problem("'%s' as well as '%s'" % (new[code].asasciilink(), pl.asasciilink())) | self.problem(err) | def assemble(self, returnonquestion = False, askall = False): new = {} for pl in self.done.keys(): code = pl.code() if code == wikipedia.mylang and pl.exists() and not pl.isRedirectPage() and not pl.isEmpty(): if pl != self.inpl: if returnonquestion: return None self.problem('Someone refers to %s with us' % pl.asasciilink()) if globalvar.autonomous: return None elif pl.exists() and not pl.isRedirectPage(): if new.has_key(code) and new[code] is None: print "NOTE: Ignoring %s"%(pl.asasciilink()) elif new.has_key(code) and new[code] != pl: if returnonquestion: return None self.problem("'%s' as well as '%s'" % (new[code].asasciilink(), pl.asasciilink())) if globalvar.autonomous: return None # beep before asking question if globalvar.bell: sys.stdout.write('\07') while 1: answer = raw_input("Use (f)ormer or (l)atter or (n)either or (g)ive up?") if answer.startswith('f'): break elif answer.startswith('l'): new[pl.code()] = pl break elif answer.startswith('n'): new[pl.code()] = None break elif answer.startswith('g'): # Give up return None elif code in ('zh-tw','zh-cn') and new.has_key('zh') and new['zh'] is not None: print "NOTE: Ignoring %s, using %s"%(new['zh'].asasciilink(),pl.asasciilink()) if self.ask(askall, pl): new['zh'] = None # Remove the global zh link new[code] = pl # Add the more precise one elif code == 'zh' and ( (new.has_key('zh-tw') and new['zh-tw'] is not None) or (new.has_key('zh-cn') and new['zh-cn'] is not None)): print "NOTE: Ignoring %s"%(pl.asasciilink()) pass # do not add global zh if there is a specific zh-tw or zh-cn elif code not in new: if self.ask(askall, pl): new[code] = pl | 09fea2b0b2f663153db90a0c2f46ba6a88409a3a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4404/09fea2b0b2f663153db90a0c2f46ba6a88409a3a/interwiki.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
19540,
12,
2890,
16,
327,
265,
4173,
273,
1083,
16,
6827,
454,
273,
1083,
4672,
394,
273,
2618,
364,
886,
316,
365,
18,
8734,
18,
2452,
13332,
981,
273,
886,
18,
710,
1435,
309,
981,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
19540,
12,
2890,
16,
327,
265,
4173,
273,
1083,
16,
6827,
454,
273,
1083,
4672,
394,
273,
2618,
364,
886,
316,
365,
18,
8734,
18,
2452,
13332,
981,
273,
886,
18,
710,
1435,
309,
981,
... |
self.logging.error( job.runningJob['schedulerId'] + \ | self.logging.error( selJob.runningJob['schedulerId'] + \ | def getOutput( self, obj, outdir='' ): """ retrieve job output """ | 55a0d373215e2d2c4f99fe71353cbd62af3d7838 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8886/55a0d373215e2d2c4f99fe71353cbd62af3d7838/SchedulerGLite.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11062,
12,
365,
16,
1081,
16,
15398,
2218,
11,
262,
30,
3536,
4614,
1719,
876,
3536,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11062,
12,
365,
16,
1081,
16,
15398,
2218,
11,
262,
30,
3536,
4614,
1719,
876,
3536,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
self.spanWidth = kwargs.get('spanWidth', 5) self.spanHeight = kwargs.get('spanHeight', 2) | self.spanColumns = kwargs.get('spanColumns', 5) self.spanRows = kwargs.get('spanRows', 2) | def __init__(self, im, mode=None, **kwargs): self.nodeColumns = kwargs.get('nodeColumns', 8) self.nodeRows = kwargs.get('nodeRows', 4) self.lineSpacing = kwargs.get('lineSpacing', 2) self.nodePadding = kwargs.get('nodePadding', 2) self.cellSize = kwargs.get('cellSize', 16) self.spanWidth = kwargs.get('spanWidth', 5) self.spanHeight = kwargs.get('spanHeight', 2) self.fill = kwargs.get('fill', (0, 0, 0)) | 9bf3d37b12467067461bdc0be583ddbbb935cb32 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/81/9bf3d37b12467067461bdc0be583ddbbb935cb32/blockdiag.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
709,
16,
1965,
33,
7036,
16,
2826,
4333,
4672,
365,
18,
2159,
3380,
273,
1205,
18,
588,
2668,
2159,
3380,
2187,
1725,
13,
365,
18,
2159,
4300,
273,
1205,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
709,
16,
1965,
33,
7036,
16,
2826,
4333,
4672,
365,
18,
2159,
3380,
273,
1205,
18,
588,
2668,
2159,
3380,
2187,
1725,
13,
365,
18,
2159,
4300,
273,
1205,... |
for path, s_id, e_id in map(None, paths, start_ids, end_ids): | path_len = len(paths) s_id_len = len(start_ids) e_id_len = len(end_ids) if path_len != s_id_len and s_id_len != e_id_len: raise RuntimeError("The lengths of paths, start_ids, and end_ids "\ +"must be the same") else: pass for i in range(path_len): path = paths[i] s_id = start_ids[i] e_id = end_ids[i] | def filter_pixels(obj, paths, start_ids, end_ids, **kwargs): """ This function filters a set of pixels based on starting and ending IDs for each data path. Parameters: ---------- -> obj is a SOM that needs to be filtered -> paths is a list of tuples containing the NeXus data paths and signals -> start_ids is a list of tuples containing the starting ID pairs (i,j) for each data path in paths -> end_ids is a list of tuples containing the ending ID pairs (i,j) for each data path in paths Returns: ------- <- A tuple of containing the SOM with the pixels in the filter range and SOM containing the rest of the pixels not filtered Exceptions: ---------- <- TypeError is raised if obj is not a SOM <- RuntimeError is raised if the number of data paths, starting IDs and ending IDs are not all identical """ # import the helper functions import hlr_utils # set up for working through data (result, res_descr) = hlr_utils.empty_result(obj) o_descr = hlr_utils.get_descr(obj) if o_descr != "SOM": raise TypeError, "Only SOM operations supported" else: pass if len(paths) != len(start_ids) and len(start_ids) != len(end_ids): raise RuntimeError, "Data paths, starting ids and ending ids need "\ +"to have the same length." else: pass result = hlr_utils.copy_som_attr(result, res_descr, obj, o_descr) # Next three try/except blocks look for difference between list and tuple try: paths.reverse() paths.reverse() except AttributeError: temp = paths paths = [] paths.append(temp) try: start_ids.reverse() start_ids.reverse() except AttributeError: temp = start_ids start_ids = [] start_ids.append(temp) try: end_ids.reverse() end_ids.reverse() except AttributeError: temp = end_ids end_ids = [] end_ids.append(temp) for path, s_id, e_id in map(None, paths, start_ids, end_ids): bank = path[0].split('/')[-1] for i in range(s_id[0], e_id[0]): for j in range(s_id[1], e_id[1]): id = (bank, (i, j)) length = hlr_utils.get_length(obj) nabove = length + 1 nbelow = 0 index = nbelow - 1 while nabove - nbelow > 1: middle = (nabove + nbelow) / 2 if id == obj[middle-1].id: index = middle - 1 break if id < obj[middle-1].id: nabove = middle else: nbelow = middle if index != -1: so = obj.pop(index) hlr_utils.result_insert(result, res_descr, so, None, "all") return (result, obj) | 29a15f400db9ebd08a8a99ae7637655a9392d7b2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/763/29a15f400db9ebd08a8a99ae7637655a9392d7b2/hlr_filter_pixels.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1034,
67,
21084,
12,
2603,
16,
2953,
16,
787,
67,
2232,
16,
679,
67,
2232,
16,
2826,
4333,
4672,
3536,
1220,
445,
3415,
279,
444,
434,
8948,
2511,
603,
5023,
471,
11463,
7115,
364,
151... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1034,
67,
21084,
12,
2603,
16,
2953,
16,
787,
67,
2232,
16,
679,
67,
2232,
16,
2826,
4333,
4672,
3536,
1220,
445,
3415,
279,
444,
434,
8948,
2511,
603,
5023,
471,
11463,
7115,
364,
151... |
self.addChannel(channel) | for channel in bot.channels: self.addChannel(channel) for channel in bot.requestedChannels: self.addChannel(channel) | def botDisconnected(self, bot): """Called when one of our bots has been disconnected""" log.msg("Bot %r on server %r disconnected" % (bot.nickname, self.host)) del self.bots[bot.nickname] for channel in self.channels.keys(): if self.channels[channel] == bot: # This is a channel formerly serviced by the bot that was just disconnected. # Remove it from our list of active channels, but try to get a bot back # in there ASAP. del self.channels[channel] self.addChannel(channel) | 0e59b8074dbb9d1a5fd85936037861e1d4bb2334 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9478/0e59b8074dbb9d1a5fd85936037861e1d4bb2334/IRC.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2512,
26303,
12,
2890,
16,
2512,
4672,
3536,
8185,
1347,
1245,
434,
3134,
2512,
87,
711,
2118,
17853,
8395,
613,
18,
3576,
2932,
6522,
738,
86,
603,
1438,
738,
86,
17853,
6,
738,
261,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2512,
26303,
12,
2890,
16,
2512,
4672,
3536,
8185,
1347,
1245,
434,
3134,
2512,
87,
711,
2118,
17853,
8395,
613,
18,
3576,
2932,
6522,
738,
86,
603,
1438,
738,
86,
17853,
6,
738,
261,
... |
file.Write("%s %s(%s);\n" % (func.return_type, func.original_name, func.MakeTypedOriginalArgString(""))) file.Write("\n") def WriteGLES2ImplementationHeader(self, func, file): """Writes the GLES2 Implemention.""" | def WriteGLES2ImplementationDeclaration(self, func, file): """Writes the GLES2 Implemention declaration.""" file.Write("%s %s(%s);\n" % (func.return_type, func.original_name, func.MakeTypedOriginalArgString(""))) file.Write("\n") | 4f0c32d419b2a9783e1458192f5c0d3d79813764 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9392/4f0c32d419b2a9783e1458192f5c0d3d79813764/build_gles2_cmd_buffer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2598,
43,
11386,
22,
13621,
6094,
12,
2890,
16,
1326,
16,
585,
4672,
3536,
8368,
326,
611,
11386,
22,
10886,
285,
8266,
12123,
585,
18,
3067,
27188,
87,
738,
87,
9275,
87,
20472,
82,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2598,
43,
11386,
22,
13621,
6094,
12,
2890,
16,
1326,
16,
585,
4672,
3536,
8368,
326,
611,
11386,
22,
10886,
285,
8266,
12123,
585,
18,
3067,
27188,
87,
738,
87,
9275,
87,
20472,
82,
6... | |
def check(ok, *args): if not ok: raise TestFailed, " ".join(map(str, args)) | def check(ok, *args): if not ok: raise TestFailed, " ".join(map(str, args)) | 5af2f7454daec9b8526ce5437783137f17a0bce3 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8546/5af2f7454daec9b8526ce5437783137f17a0bce3/test_contains.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
225,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
225,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... | |
if cur_width < cur_height: ratio = float(new_width)/cur_width else: ratio = float(new_height)/cur_height | ratio = max(float(new_width)/cur_width,float(new_height)/cur_height) | def create_size(self, photosize): if self.size_exists(photosize): return if not os.path.isdir(self.cache_path()): os.makedirs(self.cache_path()) try: im = Image.open(self.get_image_filename()) except IOError: return if im.size == photosize.size(): shutil.copy(self.get_image_filename(), self._get_SIZE_path(photosize.name)) return cur_width, cur_height = im.size new_width, new_height = photosize.size() if photosize.crop: if cur_width < cur_height: ratio = float(new_width)/cur_width else: ratio = float(new_height)/cur_height x = (cur_width * ratio) y = (cur_height * ratio) xd = abs(new_width - x) yd = abs(new_height - y) x_diff = int(xd / 2) y_diff = int(yd / 2) if self.crop_from == 'top': box = (x_diff, 0, (x-x_diff), new_height) elif self.crop_from == 'left': box = (0, y_diff, new_width, (y-y_diff)) elif self.crop_from == 'bottom': box = (x_diff, yd, (x-x_diff), y) # y - yd = new_height elif self.crop_from == 'right': box = (xd, y_diff, x, (y-y_diff)) # x - xd = new_width else: box = (x_diff, y_diff, (x-x_diff), (y-y_diff)) resized = im.resize((int(x), int(y)), Image.ANTIALIAS).crop(box) else: if not new_width == 0 and not new_height == 0: if cur_width > cur_height: ratio = float(new_width)/cur_width else: ratio = float(new_height)/cur_height else: if new_width == 0: ratio = float(new_height)/cur_height else: ratio = float(new_width)/cur_width resized = im.resize((int(cur_width*ratio), int(cur_height*ratio)), Image.ANTIALIAS) if self.filter_set is not None: filter_set = self.filter_set.filters.all() elif photosize.filter_set is not None: filter_set = list(photosize.filter_set.filters.all()) else: filter_set = None if filter_set is not None: for f in filter_set: filter = getattr(ImageFilter, f.name, None) if filter is not None: resized = resized.filter(filter) resized_filename = getattr(self, "get_%s_path" % photosize.name)() try: if im.format == 'JPEG': resized.save(resized_filename, 'JPEG', quality=photosize.quality, optimize=True) else: resized.save(resized_filename) except IOError, e: if os.path.isfile(resized_filename): os.unlink(resized_filename) raise e | 97d8ab46617aec21cd957429b4350e0ccc04fece /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/10711/97d8ab46617aec21cd957429b4350e0ccc04fece/models.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
1467,
12,
2890,
16,
18185,
554,
4672,
309,
365,
18,
1467,
67,
1808,
12,
29121,
554,
4672,
327,
309,
486,
1140,
18,
803,
18,
291,
1214,
12,
2890,
18,
2493,
67,
803,
1435,
467... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
1467,
12,
2890,
16,
18185,
554,
4672,
309,
365,
18,
1467,
67,
1808,
12,
29121,
554,
4672,
327,
309,
486,
1140,
18,
803,
18,
291,
1214,
12,
2890,
18,
2493,
67,
803,
1435,
467... |
target = 'snow' red, green, blue = colordb.find_byname(target) print target, ':', red, green, blue, hex(rrggbb) name, aliases = colordb.find_byrgb((red, green, blue)) | red, green, blue = rgbtuple = colordb.find_byname(target) print target, ':', red, green, blue, triplet_to_rrggbb(rgbtuple) name, aliases = colordb.find_byrgb(rgbtuple) | def triplet_to_pmwrgb(rgbtuple): return map(operator.__div__, rgbtuple, _maxtuple) | a5a018fbd49b29da3726905388e68c2e56b49cfd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/a5a018fbd49b29da3726905388e68c2e56b49cfd/ColorDB.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6882,
1469,
67,
869,
67,
7755,
91,
10948,
12,
10948,
8052,
4672,
327,
852,
12,
9497,
16186,
2892,
972,
16,
6917,
8052,
16,
389,
1896,
8052,
13,
565,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6882,
1469,
67,
869,
67,
7755,
91,
10948,
12,
10948,
8052,
4672,
327,
852,
12,
9497,
16186,
2892,
972,
16,
6917,
8052,
16,
389,
1896,
8052,
13,
565,
2,
-100,
-100,
-100,
-100,
-100,
-1... |
box = (x_diff, y_diff, (x-x_diff), (y-y_diff)) resized = im.resize((x, y), Image.ANTIALIAS).crop(box) | box = (x_diff, y_diff, int(x-x_diff), int(y-y_diff)) resized = im.resize((int(x), int(y)), Image.ANTIALIAS).crop(box) | def create_size(self, photosize): if self.size_exists(photosize): return if not os.path.isdir(self.cache_path()): os.makedirs(self.cache_path()) try: im = Image.open(self.get_image_filename()) except IOError: return cur_width, cur_height = im.size new_width, new_height = photosize.size() if photosize.crop: if cur_width < cur_height: ratio = float(new_width)/cur_width else: ratio = float(new_height)/cur_height x = (cur_width * ratio) y = (cur_height * ratio) x_diff = int(abs((new_width - x) / 2)) y_diff = int(abs((new_height - y) / 2)) box = (x_diff, y_diff, (x-x_diff), (y-y_diff)) resized = im.resize((x, y), Image.ANTIALIAS).crop(box) else: if not new_width == 0 and not new_height == 0: if cur_width > cur_height: ratio = float(new_width)/cur_width else: ratio = float(new_height)/cur_height else: if new_width == 0: ratio = float(new_height)/cur_height else: ratio = float(new_width)/cur_width resized = im.resize((int(cur_width*ratio), int(cur_height*ratio)), Image.ANTIALIAS) resized.save(getattr(self, "get_%s_path" % photosize.name)()) | 6ce77f65ba89351df98ecb232da58d216c1960f3 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/10711/6ce77f65ba89351df98ecb232da58d216c1960f3/models.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
1467,
12,
2890,
16,
18185,
554,
4672,
309,
365,
18,
1467,
67,
1808,
12,
29121,
554,
4672,
327,
309,
486,
1140,
18,
803,
18,
291,
1214,
12,
2890,
18,
2493,
67,
803,
1435,
467... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
1467,
12,
2890,
16,
18185,
554,
4672,
309,
365,
18,
1467,
67,
1808,
12,
29121,
554,
4672,
327,
309,
486,
1140,
18,
803,
18,
291,
1214,
12,
2890,
18,
2493,
67,
803,
1435,
467... |
gLogger.info("%s.getTransformations: Obtained %d transformations to process" % (AGENT_NAME,len(transformations))) | gLogger.info( "%s.getTransformations: Obtained %d transformations to process" % ( AGENT_NAME, len( transformations ) ) ) | def getTransformations(self): # Obtain the transformations to be executed transName = self.am_getOption('Transformation','All') if transName == 'All': gLogger.info("%s.getTransformations: Initializing general purpose agent." % AGENT_NAME) res = self.transDB.getTransformations({'Status':['Active','Completing','Flush']},extraParams=True) if not res['OK']: gLogger.error("%s.getTransformations: Failed to get transformations." % AGENT_NAME, res['Message']) return res transformations = res['Value'] gLogger.info("%s.getTransformations: Obtained %d transformations to process" % (AGENT_NAME,len(transformations))) else: gLogger.info("%s.getTransformations: Initializing for transformation %s." % (AGENT_NAME,transName)) res = self.transDB.getTransformation(transName,extraParams=True) if not res['OK']: gLogger.error("%s.getTransformations: Failed to get transformation." % AGENT_NAME, res['Message']) return res transformations = [res['Value']] return S_OK(transformations) | 5b98b3b45ce744b0fa8f66e9d9812d0b7910a0ab /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/5b98b3b45ce744b0fa8f66e9d9812d0b7910a0ab/TransformationAgent.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
4059,
1012,
12,
2890,
4672,
468,
24850,
326,
19245,
358,
506,
7120,
906,
461,
273,
365,
18,
301,
67,
588,
1895,
2668,
15292,
17023,
1595,
6134,
309,
906,
461,
422,
296,
1595,
4278,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
4059,
1012,
12,
2890,
4672,
468,
24850,
326,
19245,
358,
506,
7120,
906,
461,
273,
365,
18,
301,
67,
588,
1895,
2668,
15292,
17023,
1595,
6134,
309,
906,
461,
422,
296,
1595,
4278,
... |
def __init__(self, updd, bus_name, object_path='/Updatesd'): | def __init__(self, updd, bus_name, object_path='/Updatesd', allowshutdown = False): | def __init__(self, updd, bus_name, object_path='/Updatesd'): dbus.service.Object.__init__(self, bus_name, object_path) self.updd = updd self.allowshutdown = False | 407777395791145c466bec07939913193d3f17e6 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5445/407777395791145c466bec07939913193d3f17e6/yum-updatesd.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
731,
449,
16,
5766,
67,
529,
16,
733,
67,
803,
2218,
19,
5121,
72,
2187,
1699,
15132,
273,
1083,
4672,
21866,
18,
3278,
18,
921,
16186,
2738,
972,
12,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
731,
449,
16,
5766,
67,
529,
16,
733,
67,
803,
2218,
19,
5121,
72,
2187,
1699,
15132,
273,
1083,
4672,
21866,
18,
3278,
18,
921,
16186,
2738,
972,
12,
... |
'Improved login form') self.assertEqual(datagrid.rows[2]['object'].summary, | def testDashboard5(self): """Testing dashboard view (mine)""" self.client.login(username='doc', password='doc') | d72238b7393eae1f5069fef4eba2d906a7ba3bdf /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1600/d72238b7393eae1f5069fef4eba2d906a7ba3bdf/tests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
14830,
25,
12,
2890,
4672,
3536,
22218,
11825,
1476,
261,
3081,
15574,
365,
18,
2625,
18,
5819,
12,
5053,
2218,
2434,
2187,
2201,
2218,
2434,
6134,
2,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
14830,
25,
12,
2890,
4672,
3536,
22218,
11825,
1476,
261,
3081,
15574,
365,
18,
2625,
18,
5819,
12,
5053,
2218,
2434,
2187,
2201,
2218,
2434,
6134,
2,
-100,
-100,
-100,
-100,
-100,
... | |
self.extcmds[0] = 'recent' | self.extcmds.insert(0, 'recent') | def parseCommands(self): """reads self.cmds and parses them out to make sure that the requested base command + argument makes any sense at all""" | 96e6ed5e0bb7909b018bf0d0fd8f70140bee9c1e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5445/96e6ed5e0bb7909b018bf0d0fd8f70140bee9c1e/cli.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
9127,
12,
2890,
4672,
3536,
11904,
365,
18,
24680,
471,
8940,
2182,
596,
358,
1221,
3071,
716,
326,
3764,
1026,
1296,
397,
1237,
7297,
1281,
12764,
622,
777,
8395,
2,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
9127,
12,
2890,
4672,
3536,
11904,
365,
18,
24680,
471,
8940,
2182,
596,
358,
1221,
3071,
716,
326,
3764,
1026,
1296,
397,
1237,
7297,
1281,
12764,
622,
777,
8395,
2,
-100,
-100,
-... |
self._changepath(f, 'n') | self._changepath(f, 'n', True) | def normal(self, f): 'mark a file normal and clean' self._dirty = True self._changepath(f, 'n') s = os.lstat(self._join(f)) self._map[f] = ('n', s.st_mode, s.st_size, s.st_mtime, 0) if self._copymap.has_key(f): del self._copymap[f] | 77c027da3db930123dc583c05ad09b55a604f4b9 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/11312/77c027da3db930123dc583c05ad09b55a604f4b9/dirstate.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2212,
12,
2890,
16,
284,
4672,
296,
3355,
279,
585,
2212,
471,
2721,
11,
365,
6315,
18013,
273,
1053,
365,
6315,
3427,
803,
12,
74,
16,
296,
82,
2187,
1053,
13,
272,
273,
1140,
18,
8... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2212,
12,
2890,
16,
284,
4672,
296,
3355,
279,
585,
2212,
471,
2721,
11,
365,
6315,
18013,
273,
1053,
365,
6315,
3427,
803,
12,
74,
16,
296,
82,
2187,
1053,
13,
272,
273,
1140,
18,
8... |
owner = ClusterUser(id=1, name='foobar') | owner = ClusterUser(id=32, name='foobar') | def test_non_trivial(self): """ Test instantiating a VirtualMachine with extra parameters """ # Define cluster for use vm_hostname='vm.test.org' cluster = Cluster(hostname='test.osuosl.bak', slug='OSL_TEST') cluster.save() owner = ClusterUser(id=1, name='foobar') # Cluster vm = VirtualMachine(cluster=cluster, hostname=vm_hostname) vm.save() self.assertTrue(vm.id) self.assertEqual('vm.test.org', vm.hostname) self.assertFalse(vm.error) vm.delete() # Multiple vm = VirtualMachine(cluster=cluster, hostname=vm_hostname, virtual_cpus=3, ram=512, disk_size=5120, owner=owner) vm.save() self.assertTrue(vm.id) self.assertEqual('vm.test.org', vm.hostname) self.assertEqual(512, vm.ram) self.assertEqual(5120, vm.disk_size) self.assertEqual('foobar', vm.owner.name) self.assertFalse(vm.error) # Remove cluster Cluster.objects.all().delete(); | c24ff4471dfca799ba7ddabc41abb048929bc8fd /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10509/c24ff4471dfca799ba7ddabc41abb048929bc8fd/virtual_machine.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
5836,
67,
313,
20109,
12,
2890,
4672,
3536,
7766,
5934,
22906,
279,
18452,
598,
2870,
1472,
3536,
468,
13184,
2855,
364,
999,
4268,
67,
10358,
2218,
3489,
18,
3813,
18,
3341,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
5836,
67,
313,
20109,
12,
2890,
4672,
3536,
7766,
5934,
22906,
279,
18452,
598,
2870,
1472,
3536,
468,
13184,
2855,
364,
999,
4268,
67,
10358,
2218,
3489,
18,
3813,
18,
3341,
1... |
serv_args = (self.evt, 2) | serv_args = (self.evt, 1) | def setUp(self): self.evt = threading.Event() # start server thread to handle requests serv_args = (self.evt, 2) threading.Thread(target=http_server, args=serv_args).start() | 15c974651f7bb8e54a008359cf306a8e8ec13c12 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8546/15c974651f7bb8e54a008359cf306a8e8ec13c12/test_xmlrpc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
24292,
12,
2890,
4672,
365,
18,
73,
11734,
273,
17254,
18,
1133,
1435,
468,
787,
1438,
2650,
358,
1640,
3285,
13515,
67,
1968,
273,
261,
2890,
18,
73,
11734,
16,
404,
13,
17254,
18,
38... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
24292,
12,
2890,
4672,
365,
18,
73,
11734,
273,
17254,
18,
1133,
1435,
468,
787,
1438,
2650,
358,
1640,
3285,
13515,
67,
1968,
273,
261,
2890,
18,
73,
11734,
16,
404,
13,
17254,
18,
38... |
('?post', SIOC['id'], '?id')]) opt = sparql.GraphPattern([('?post', SIOC['title'], '?postTitle'), | ('?post', SIOC['title'], '?postTitle'), | def query(self): try: self.graph = self.loadMailingList(self.uri) self.loadAdditionalData(self.uri) print 'Total triples loaded:', len(self.graph) sparqlGr = sparql.sparqlGraph.SPARQLGraph(self.graph) select = ('?post', '?id', '?postTitle', '?userName') where = sparql.GraphPattern([('?post', RDF['type'], SIOC['Post']), ('?post', SIOC['id'], '?id')]) opt = sparql.GraphPattern([('?post', SIOC['title'], '?postTitle'), ('?post', SIOC['has_creator'], '?user'), ('?user', SIOC['name'], '?userName')]) posts = sparqlGr.query(select, where, opt) return posts except Exception, details: gsr.messageBar('unknow problem parsing RDF at ' + self.uri) print 'parsing exception:', str(details) return None | 5b54a91a94e4f31ef0b42d45e24c1d892de27e0d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5712/5b54a91a94e4f31ef0b42d45e24c1d892de27e0d/gsr.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
843,
12,
2890,
4672,
775,
30,
365,
18,
4660,
273,
365,
18,
945,
6759,
310,
682,
12,
2890,
18,
1650,
13,
365,
18,
945,
10552,
751,
12,
2890,
18,
1650,
13,
225,
1172,
296,
5269,
22388,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
843,
12,
2890,
4672,
775,
30,
365,
18,
4660,
273,
365,
18,
945,
6759,
310,
682,
12,
2890,
18,
1650,
13,
365,
18,
945,
10552,
751,
12,
2890,
18,
1650,
13,
225,
1172,
296,
5269,
22388,... |
return channel_vals.__iter__ | return channel_vals.__iter__() | def __iter__(self): # To make thread safe channel_vals = self._channels.values() return channel_vals.__iter__ | 2b8a06c8c17561d3abc9207149d9506cdeb9343a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5028/2b8a06c8c17561d3abc9207149d9506cdeb9343a/channel.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2165,
972,
12,
2890,
4672,
468,
2974,
1221,
2650,
4183,
1904,
67,
4524,
273,
365,
6315,
9114,
18,
2372,
1435,
327,
1904,
67,
4524,
16186,
2165,
972,
2,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2165,
972,
12,
2890,
4672,
468,
2974,
1221,
2650,
4183,
1904,
67,
4524,
273,
365,
6315,
9114,
18,
2372,
1435,
327,
1904,
67,
4524,
16186,
2165,
972,
2,
-100,
-100,
-100,
-100,
-100... |
def urlopen(url, data=None): | def urlopen(url, data=None, proxies=None): | def urlopen(url, data=None): """urlopen(url [, data]) -> open file-like object""" global _urlopener if not _urlopener: _urlopener = FancyURLopener() if data is None: return _urlopener.open(url) else: return _urlopener.open(url, data) | 22806d26422dea44e8ab7064525004b4cf13790b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/22806d26422dea44e8ab7064525004b4cf13790b/urllib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
19613,
12,
718,
16,
501,
33,
7036,
16,
13263,
33,
7036,
4672,
3536,
295,
18589,
12,
718,
306,
16,
501,
5717,
317,
1696,
585,
17,
5625,
733,
8395,
2552,
389,
295,
16884,
708,
309,
486,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
19613,
12,
718,
16,
501,
33,
7036,
16,
13263,
33,
7036,
4672,
3536,
295,
18589,
12,
718,
306,
16,
501,
5717,
317,
1696,
585,
17,
5625,
733,
8395,
2552,
389,
295,
16884,
708,
309,
486,
... |
imclass = object.im_class | def docroutine(self, object, name=None, mod=None, funcs={}, classes={}, methods={}, cl=None): """Produce HTML documentation for a function or method object.""" realname = object.__name__ name = name or realname anchor = (cl and cl.__name__ or '') + '-' + name note = '' skipdocs = 0 if inspect.ismethod(object): if cl: imclass = object.im_class if imclass is not cl: url = '%s.html#%s-%s' % ( imclass.__module__, imclass.__name__, name) note = ' from <a href="%s">%s</a>' % ( url, classname(imclass, mod)) skipdocs = 1 else: note = (object.im_self and ' method of %s instance' + object.im_self.__class__ or ' unbound %s method' % object.im_class.__name__) object = object.im_func | 6dcfa3884d3e4a6fe8dc57eec4d8df1980cda04e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/6dcfa3884d3e4a6fe8dc57eec4d8df1980cda04e/pydoc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
997,
22640,
12,
2890,
16,
733,
16,
508,
33,
7036,
16,
681,
33,
7036,
16,
15630,
28793,
3318,
28793,
2590,
28793,
927,
33,
7036,
4672,
3536,
25884,
3982,
7323,
364,
279,
445,
578,
707,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
997,
22640,
12,
2890,
16,
733,
16,
508,
33,
7036,
16,
681,
33,
7036,
16,
15630,
28793,
3318,
28793,
2590,
28793,
927,
33,
7036,
4672,
3536,
25884,
3982,
7323,
364,
279,
445,
578,
707,
... | |
if self.serverVariables['CONTENT_TYPE'] == 'text/xml': | if self.serverVariables['CONTENT_TYPE'][:8] == 'text/xml': | def __init__(self, rawRequest): self.serverVariables = rawRequest['env'] #print self.serverVariables self.serverVariables.setdefault('HTTP_ACCEPT_LANGUAGE', '') self.serverVariables.setdefault('HTTP_USER_AGENT', '') self.serverVariables.setdefault('PATH_INFO', '/') if self.serverVariables.setdefault('QUERY_STRING', ''): self.queryString = parse_qs(self.serverVariables['QUERY_STRING']) else: self.queryString = {} self.cookies = Cookie.SimpleCookie() if self.serverVariables.has_key('HTTP_COOKIE'): self.cookies.load(self.serverVariables['HTTP_COOKIE']) self.input = cStringIO.StringIO(rawRequest['inp']) self.interface = rawRequest['if'] self.method = self.queryString.get('cmd', [''])[0] self.form = None self.type = 'http' if self.serverVariables['REQUEST_METHOD'] == 'POST': if self.serverVariables['CONTENT_TYPE'] == 'text/xml': # xmlrpc request? method_match = re.search(XMLRPC_METHOD, self.input.getvalue()) if method_match: self.method = method_match.groups()[0] self.type = 'xmlrpc' else: # http form post self.form = FieldStorage(fp=self.input, environ=self.serverVariables) | c2a34621d245663eed4cce6e85ce6e20bc5ea0fd /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/2504/c2a34621d245663eed4cce6e85ce6e20bc5ea0fd/request.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1831,
691,
4672,
365,
18,
3567,
6158,
273,
1831,
691,
3292,
3074,
3546,
225,
468,
1188,
365,
18,
3567,
6158,
225,
365,
18,
3567,
6158,
18,
542,
1886,
266... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1831,
691,
4672,
365,
18,
3567,
6158,
273,
1831,
691,
3292,
3074,
3546,
225,
468,
1188,
365,
18,
3567,
6158,
225,
365,
18,
3567,
6158,
18,
542,
1886,
266... |
[(word: 111, word: 111, None)] | [(word: 111, word: 111, word: 1)] | def rauzy_graph(self, n): r""" Returns the Rauzy graph of the factors of length n of self. | fbc8adcb81ea8c5edc4559a9852a8804df3117df /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/fbc8adcb81ea8c5edc4559a9852a8804df3117df/word.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
767,
89,
21832,
67,
4660,
12,
2890,
16,
290,
4672,
436,
8395,
2860,
326,
534,
8377,
21832,
2667,
434,
326,
14490,
434,
769,
290,
434,
365,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
767,
89,
21832,
67,
4660,
12,
2890,
16,
290,
4672,
436,
8395,
2860,
326,
534,
8377,
21832,
2667,
434,
326,
14490,
434,
769,
290,
434,
365,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
print "Warning html of file not converted: %s" % stderr | def to_html(self, outfile=None): if self.__html_output__ is not None: return self.__html_output__ print "OfficeDocument.to_html" | 5acf607fe866693c0a2b88b17b7895c564d2f2bd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12681/5acf607fe866693c0a2b88b17b7895c564d2f2bd/office.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
358,
67,
2620,
12,
2890,
16,
8756,
33,
7036,
4672,
309,
365,
16186,
2620,
67,
2844,
972,
353,
486,
599,
30,
327,
365,
16186,
2620,
67,
2844,
972,
1172,
315,
30126,
2519,
18,
869,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
358,
67,
2620,
12,
2890,
16,
8756,
33,
7036,
4672,
309,
365,
16186,
2620,
67,
2844,
972,
353,
486,
599,
30,
327,
365,
16186,
2620,
67,
2844,
972,
1172,
315,
30126,
2519,
18,
869,
67,
... | |
self.topic.post_count -= 1 self.topic.save() self.topic.forum.post_count -= 1 | if self_id == head_post_id: self.topic.forum.post_count -= 1 + self.topic.posts.all().count() self.topic.delete() else: self.topic.post_count -= 1 self.topic.save() self.topic.forum.post_count -= 1 | def delete(self, *args, **kwargs): self_id = self.id head_post_id = self.topic.posts.order_by('created')[0].id super(Post, self).delete(*args, **kwargs) | 3efa43027a001644543a2edab16cb7e31f4562fa /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12938/3efa43027a001644543a2edab16cb7e31f4562fa/models.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1430,
12,
2890,
16,
380,
1968,
16,
2826,
4333,
4672,
365,
67,
350,
273,
365,
18,
350,
910,
67,
2767,
67,
350,
273,
365,
18,
10476,
18,
11660,
18,
1019,
67,
1637,
2668,
4824,
6134,
63... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1430,
12,
2890,
16,
380,
1968,
16,
2826,
4333,
4672,
365,
67,
350,
273,
365,
18,
350,
910,
67,
2767,
67,
350,
273,
365,
18,
10476,
18,
11660,
18,
1019,
67,
1637,
2668,
4824,
6134,
63... |
deps.add(os.path.normpath(path)) | base_dependency_name = groups[2] path = os.path.normpath(path) if os.path.exists(path): deps.add(path) else: found_include = False for idir in CYTHON_INCLUDE_DIRS: new_path = os.path.normpath(idir + base_dependency_name) if os.path.exists(new_path): deps.add(new_path) found_include = True break if not found_include: raise IOError, "could not find dependency %s included in %s."%(path, filename) | def parse_deps(self, filename, verify=True): """ Open a Cython file and extract all of its dependencies. | 04b161befbb9b35fa223fc05ffd8eb265efc4eea /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9417/04b161befbb9b35fa223fc05ffd8eb265efc4eea/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
14877,
12,
2890,
16,
1544,
16,
3929,
33,
5510,
4672,
3536,
3502,
279,
385,
18490,
585,
471,
2608,
777,
434,
2097,
5030,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
14877,
12,
2890,
16,
1544,
16,
3929,
33,
5510,
4672,
3536,
3502,
279,
385,
18490,
585,
471,
2608,
777,
434,
2097,
5030,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
" %s(%s, len + 1, &len, bucket->GetDataAs<GLchar*>(0, len + 1));\n" % (func.GetGLFunctionName(), id_arg.name)) | " %s %s = GetSharedMemoryAs<%s>(\n" % (dest_arg.type, dest_arg.name, dest_arg.type)) file.Write( " c.%s_shm_id, c.%s_shm_offset, %s);\n" % (dest_arg.name, dest_arg.name, bufsize_arg.name)) for arg in all_but_last_2_args + [dest_arg]: arg.WriteValidationCode(file) func.WriteValidationCode(file) func.WriteHandlerImplementation(file) | def WriteServiceImplementation(self, func, file): """Overrriden from TypeHandler.""" file.Write( "error::Error GLES2DecoderImpl::Handle%s(\n" % func.name) file.Write( " uint32 immediate_data_size, const gles2::%s& c) {\n" % func.name) args = func.GetCmdArgs() id_arg = args[0] bucket_arg = args[1] id_arg.WriteGetCode(file) bucket_arg.WriteGetCode(file) id_arg.WriteValidationCode(file) file.Write(" GLint len = 0;\n") file.Write(" %s(%s, %s, &len);\n" % ( func.GetInfo('get_len_func'), id_arg.name, func.GetInfo('get_len_enum'))) file.Write(" Bucket* bucket = CreateBucket(%s);\n" % bucket_arg.name) file.Write(" bucket->SetSize(len + 1);\n"); file.Write( " %s(%s, len + 1, &len, bucket->GetDataAs<GLchar*>(0, len + 1));\n" % (func.GetGLFunctionName(), id_arg.name)) file.Write(" return error::kNoError;\n") file.Write("}\n") file.Write("\n") | b8fb1c2876060305792c0b72943407a8dd011c60 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5060/b8fb1c2876060305792c0b72943407a8dd011c60/build_gles2_cmd_buffer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2598,
1179,
13621,
12,
2890,
16,
1326,
16,
585,
4672,
3536,
22042,
1691,
275,
628,
1412,
1503,
12123,
585,
18,
3067,
12,
315,
1636,
2866,
668,
611,
11386,
22,
7975,
2828,
2866,
3259,
9,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2598,
1179,
13621,
12,
2890,
16,
1326,
16,
585,
4672,
3536,
22042,
1691,
275,
628,
1412,
1503,
12123,
585,
18,
3067,
12,
315,
1636,
2866,
668,
611,
11386,
22,
7975,
2828,
2866,
3259,
9,
... |
try: listdef = self.listdefs[self.position_key] except KeyError: listdef = ListDef(self, s_ImpossibleValue) for e in x: listdef.generalize(self.annotation_from_example(e)) self.listdefs[self.position_key] = listdef | listdef = ListDef(self, s_ImpossibleValue) for e in x: listdef.generalize(self.annotation_from_example(e)) self.listdefs[self.position_key] = listdef | def annotation_from_example(self, x): """The most precise SomeValue instance that contains the mutable value x.""" # convert unbound methods to the underlying function if hasattr(x, 'im_self') and x.im_self is None: x = x.im_func assert not hasattr(x, 'im_self') if x is sys: # special case constant sys to someobject return SomeObject() tp = type(x) if issubclass(tp, Symbolic): # symbolic constants support result = x.annotation() return result if tp is bool: result = SomeBool() elif tp is int: result = SomeInteger(nonneg = x>=0) elif issubclass(tp, str): # py.lib uses annotated str subclasses if len(x) == 1: result = SomeChar() else: result = SomeString() elif tp is unicode and len(x) == 1: result = SomeUnicodeCodePoint() elif tp is tuple: result = SomeTuple(items = [self.annotation_from_example(e) for e in x]) elif tp is float: result = SomeFloat() elif tp is list: # try to get listdef try: listdef = self.listdefs[self.position_key] except KeyError: listdef = ListDef(self, s_ImpossibleValue) for e in x: listdef.generalize(self.annotation_from_example(e)) self.listdefs[self.position_key] = listdef result = SomeList(listdef) elif tp is dict or tp is r_dict: try: dictdef = self.dictdefs[self.position_key] except KeyError: dictdef = DictDef(self, s_ImpossibleValue, s_ImpossibleValue, is_r_dict = tp is r_dict) if tp is r_dict: s_eqfn = self.immutablevalue(x.key_eq) s_hashfn = self.immutablevalue(x.key_hash) dictdef.dictkey.update_rdict_annotations(s_eqfn, s_hashfn) for ek, ev in x.iteritems(): dictdef.generalize_key(self.annotation_from_example(ek)) dictdef.generalize_value(self.annotation_from_example(ev)) result = SomeDict(dictdef) elif ishashable(x) and x in BUILTIN_ANALYZERS: _module = getattr(x,"__module__","unknown") result = SomeBuiltin(BUILTIN_ANALYZERS[x], methodname="%s.%s" % (_module, x.__name__)) elif extregistry.is_registered(x, self.policy): entry = extregistry.lookup(x, self.policy) result = entry.compute_annotation_bk(self) | e5ed9f77e0095f6d554c54e2bbe2055cdd00540a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6934/e5ed9f77e0095f6d554c54e2bbe2055cdd00540a/bookkeeper.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3204,
67,
2080,
67,
8236,
12,
2890,
16,
619,
4672,
3536,
1986,
4486,
31293,
10548,
620,
791,
716,
1914,
326,
16074,
460,
619,
12123,
468,
1765,
30177,
2590,
358,
326,
6808,
445,
309,
385... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3204,
67,
2080,
67,
8236,
12,
2890,
16,
619,
4672,
3536,
1986,
4486,
31293,
10548,
620,
791,
716,
1914,
326,
16074,
460,
619,
12123,
468,
1765,
30177,
2590,
358,
326,
6808,
445,
309,
385... |
gLogger.debug(infoStr) | gLogger.debug( infoStr ) | def __importExternals(self): if (self.lcg_util) and (self.gfal): return S_OK() try: import lcg_util infoStr = 'Using lcg_util from: %s' % lcg_util.__file__ gLogger.debug(infoStr) infoStr = "The version of lcg_utils is %s" % lcg_util.lcg_util_version() gLogger.debug(infoStr) except Exception,x: errStr = "SRM2Storage.__init__: Failed to import lcg_util" gLogger.exception(errStr,'',x) ISOK = False return S_ERROR(errStr) try: import gfalthr as gfal infoStr = "Using gfalthr from: %s" % gfal.__file__ gLogger.debug(infoStr) infoStr = "The version of gfalthr is %s" % gfal.gfal_version() gLogger.debug(infoStr) except Exception,x: errStr = "SRM2Storage.__init__: Failed to import gfalthr: %s." % (x) gLogger.warn(errStr) try: import gfal infoStr = "Using gfal from: %s" % gfal.__file__ gLogger.debug(infoStr) infoStr = "The version of gfal is %s" % gfal.gfal_version() gLogger.debug(infoStr) except Exception,x: errStr = "SRM2Storage.__init__: Failed to import gfal" gLogger.exception(errStr,'',x) ISOK = False return S_ERROR(errStr) self.lcg_util = lcg_util self.gfal = gfal return S_OK() | 70e66af095cb6701e39b1e701e4a2ce4d012b4f7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/70e66af095cb6701e39b1e701e4a2ce4d012b4f7/SRM2Storage.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
5666,
29764,
1031,
12,
2890,
4672,
309,
261,
2890,
18,
17704,
75,
67,
1367,
13,
471,
261,
2890,
18,
75,
74,
287,
4672,
327,
348,
67,
3141,
1435,
775,
30,
1930,
9109,
75,
67,
13... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
5666,
29764,
1031,
12,
2890,
4672,
309,
261,
2890,
18,
17704,
75,
67,
1367,
13,
471,
261,
2890,
18,
75,
74,
287,
4672,
327,
348,
67,
3141,
1435,
775,
30,
1930,
9109,
75,
67,
13... |
return SQL.getProcedure(self.realConn, procName) | return Oracle.getProcedure(self.realConn, procName) | def getProcedure(self, procName): return SQL.getProcedure(self.realConn, procName) | 591612f2a264a9c64f034f016bd3c52359f40201 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2304/591612f2a264a9c64f034f016bd3c52359f40201/oracleconn.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3570,
8153,
12,
2890,
16,
5418,
461,
4672,
327,
3063,
18,
588,
17213,
12,
2890,
18,
7688,
3543,
16,
5418,
461,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3570,
8153,
12,
2890,
16,
5418,
461,
4672,
327,
3063,
18,
588,
17213,
12,
2890,
18,
7688,
3543,
16,
5418,
461,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
security.declareProtected(ModifyPortalContent, 'setRecipient_email') | security.declareProtected(ModifyPortalContent, 'setRecipientEmail') | def setBCCRecipients(self, value, **kwargs): self.setExpressionsField('bcc_recipients', value, **kwargs) | 1726ea530e255e00923a42c26528b516eef4ebb9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1807/1726ea530e255e00923a42c26528b516eef4ebb9/PloneFormMailer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
15268,
6743,
22740,
12,
2890,
16,
460,
16,
2826,
4333,
4672,
365,
18,
542,
8927,
974,
2668,
31345,
67,
27925,
2187,
460,
16,
2826,
4333,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
15268,
6743,
22740,
12,
2890,
16,
460,
16,
2826,
4333,
4672,
365,
18,
542,
8927,
974,
2668,
31345,
67,
27925,
2187,
460,
16,
2826,
4333,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
def __addsub(self, other, fn): | def __add__(self,other,fn): | def __addsub(self, other, fn): # First check if argument is a scalar if isscalarlike(other): # Now we would add this scalar to every element. raise NotImplementedError, 'adding a scalar to a CSC or CSR ' \ 'matrix is not supported' elif isspmatrix(other): if (other.shape != self.shape): raise ValueError, "inconsistent shapes" return self._binopt(other,fn) elif isdense(other): # Convert this matrix to a dense matrix and add them return other + self.todense() else: raise NotImplemented | f6118bd5eab2f5ec319329b322947271865d4bd5 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12971/f6118bd5eab2f5ec319329b322947271865d4bd5/sparse.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
225,
1652,
1001,
1289,
972,
12,
2890,
16,
3011,
16,
4293,
4672,
468,
5783,
866,
309,
1237,
353,
279,
4981,
309,
3385,
3473,
5625,
12,
3011,
4672,
468,
4494,
732,
4102,
527,
333,
4981,
358,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
225,
1652,
1001,
1289,
972,
12,
2890,
16,
3011,
16,
4293,
4672,
468,
5783,
866,
309,
1237,
353,
279,
4981,
309,
3385,
3473,
5625,
12,
3011,
4672,
468,
4494,
732,
4102,
527,
333,
4981,
358,
3... |
for label, view in self.menus[menu_name]: | for label, view in self.menus[menu_name]: | def render(self, menu_name, depth=0, active=None): """ The render() method returns a HTML string suitable for use as a menu bar on a given page. menu_name: the label of a menu, as specified at class initialization. depth (kw): an integer specifying how far into the tree this is being rendered. Usually the render() method takes care of the whole menu, but it may be occasionally useful to delegate a single sub-menu to the SimpleMenu class. Note that this parameter only affects the CSS class. active(kw): the active label, if any, in the menu. | 6d3257b8c4a1776d5526acbc55ae4b47e2cf5dde /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/4778/6d3257b8c4a1776d5526acbc55ae4b47e2cf5dde/menu.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1743,
12,
2890,
16,
3824,
67,
529,
16,
3598,
33,
20,
16,
2695,
33,
7036,
4672,
3536,
1021,
1743,
1435,
707,
1135,
279,
3982,
533,
10631,
364,
999,
487,
279,
3824,
4653,
603,
279,
864,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1743,
12,
2890,
16,
3824,
67,
529,
16,
3598,
33,
20,
16,
2695,
33,
7036,
4672,
3536,
1021,
1743,
1435,
707,
1135,
279,
3982,
533,
10631,
364,
999,
487,
279,
3824,
4653,
603,
279,
864,
... |
shutil.rmtree(tmpdir) | shutil.rmtree(tmpdir) | def transfer_recipe_logs(self, recipe_id): """ If Cache is turned on then move the recipes logs to there final place """ if self.conf.get("CACHE",False): tmpdir = tempfile.mkdtemp(dir=self.basepath) # Move logs to tmp directory layout mylogs = self.hub.recipes.files(recipe_id) trlogs = [] for mylog in mylogs: server = '%s/%s' % (self.conf.get("ARCHIVE_SERVER"), mylog['filepath']) basepath = '%s/%s' % (self.conf.get("ARCHIVE_BASEPATH"), mylog['filepath']) mysrc = '%s/%s/%s' % (mylog['basepath'], mylog['path'], mylog['filename']) mydst = '%s/%s/%s/%s' % (tmpdir, mylog['filepath'], mylog['path'], mylog['filename']) if not os.path.exists(os.path.dirname(mydst)): os.makedirs(os.path.dirname(mydst)) try: os.link(mysrc,mydst) trlogs.append(mylog) except OSError: pass # rsync the logs to there new home rc = self.rsync('%s/' % tmpdir, '%s' % self.conf.get("ARCHIVE_RSYNC")) if rc == 0: # if the logs have been transfered then tell the server the new location for mylog in trlogs: server = '%s/%s' % (self.conf.get("ARCHIVE_SERVER"), mylog['filepath']) basepath = '%s/%s' % (self.conf.get("ARCHIVE_BASEPATH"), mylog['filepath']) mysrc = '%s/%s/%s' % (mylog['basepath'], mylog['path'], mylog['filename']) self.hub.recipes.change_file(mylog['tid'], server, basepath) self.rm(mysrc) try: self.removedirs('%s/%s' % (mylog['basepath'], mylog['path'])) except OSError: # Its ok if it fails, dir may not be empty yet pass # get rid of our tmpdir. shutil.rmtree(tmpdir) | c9a6bd212553375856eecd65c1fa6ac25c2b552f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14755/c9a6bd212553375856eecd65c1fa6ac25c2b552f/proxy.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7412,
67,
3927,
3151,
67,
10011,
12,
2890,
16,
16100,
67,
350,
4672,
3536,
971,
4379,
353,
21826,
603,
1508,
3635,
326,
5863,
281,
5963,
358,
1915,
727,
3166,
3536,
309,
365,
18,
3923,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7412,
67,
3927,
3151,
67,
10011,
12,
2890,
16,
16100,
67,
350,
4672,
3536,
971,
4379,
353,
21826,
603,
1508,
3635,
326,
5863,
281,
5963,
358,
1915,
727,
3166,
3536,
309,
365,
18,
3923,
... |
ext.sources.extend(fficonfig['ffi_sources']) | ext.sources.extend(os.path.join(ffi_srcdir, f) for f in fficonfig['ffi_sources']) | def configure_ctypes(self, ext): if not self.use_system_libffi: if sys.platform == 'darwin': return self.configure_ctypes_darwin(ext) | f6b48cb836d441bdc70cc93ffced86d6221bc7ee /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8125/f6b48cb836d441bdc70cc93ffced86d6221bc7ee/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5068,
67,
299,
989,
12,
2890,
16,
1110,
4672,
309,
486,
365,
18,
1202,
67,
4299,
67,
2941,
1403,
77,
30,
309,
2589,
18,
9898,
422,
296,
28076,
4278,
327,
365,
18,
14895,
67,
299,
989... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5068,
67,
299,
989,
12,
2890,
16,
1110,
4672,
309,
486,
365,
18,
1202,
67,
4299,
67,
2941,
1403,
77,
30,
309,
2589,
18,
9898,
422,
296,
28076,
4278,
327,
365,
18,
14895,
67,
299,
989... |
navtrail = '''<a class=navtrail href="%s/admin/index.%s.html">%s</a> > <a class=navtrail href="%s/admin/bibformat/bibformatadmin.py/?ln=%s">%s</a> ''' % (weburl, ln, _("Admin Area"), weburl, ln, _("BibFormat Admin")) | navtrail = '''<a class=navtrail href="%s/admin/index.%s.html">%s</a> > <a class=navtrail href="%s/admin/bibformat/bibformatadmin.py?ln=%s">%s</a> ''' % (weburl, ln, _("Admin Area"), weburl, ln, _("BibFormat Admin")) | def getnavtrail(previous = '', ln=cdslang): """Get the navtrail""" previous = wash_url_argument(previous, 'str') ln = wash_language(ln) _ = gettext_set_language(ln) navtrail = '''<a class=navtrail href="%s/admin/index.%s.html">%s</a> > <a class=navtrail href="%s/admin/bibformat/bibformatadmin.py/?ln=%s">%s</a> ''' % (weburl, ln, _("Admin Area"), weburl, ln, _("BibFormat Admin")) navtrail = navtrail + previous return navtrail | 4aba3f30f10872d4adc5f7ac0e2e0bae44ed9754 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12027/4aba3f30f10872d4adc5f7ac0e2e0bae44ed9754/bibformatadminlib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
11589,
15565,
12,
11515,
273,
10226,
7211,
33,
4315,
2069,
539,
4672,
3536,
967,
326,
8775,
15565,
8395,
2416,
273,
341,
961,
67,
718,
67,
3446,
12,
11515,
16,
296,
701,
6134,
7211,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
11589,
15565,
12,
11515,
273,
10226,
7211,
33,
4315,
2069,
539,
4672,
3536,
967,
326,
8775,
15565,
8395,
2416,
273,
341,
961,
67,
718,
67,
3446,
12,
11515,
16,
296,
701,
6134,
7211,... |
context=None, target_language=None, default=None): | def translate (self, domain, msgid, mapping=None, """Interpolates and translate TAL expression.""" context=None, target_language=None, default=None): _msg = self.gettext(msgid) wc.log.debug(wc.LOG_TAL, "TRANSLATED %r %r", msgid, _msg) return wc.webgui.TAL.TALInterpreter.interpolate(_msg, mapping) | 5f13f232088408097e4d2517a71797db42dfb5f3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3948/5f13f232088408097e4d2517a71797db42dfb5f3/webconfig.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4204,
261,
2890,
16,
2461,
16,
24389,
16,
2874,
33,
7036,
16,
3536,
15665,
815,
471,
4204,
399,
1013,
2652,
12123,
389,
3576,
273,
365,
18,
29967,
12,
3576,
350,
13,
14310,
18,
1330,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4204,
261,
2890,
16,
2461,
16,
24389,
16,
2874,
33,
7036,
16,
3536,
15665,
815,
471,
4204,
399,
1013,
2652,
12123,
389,
3576,
273,
365,
18,
29967,
12,
3576,
350,
13,
14310,
18,
1330,
1... | |
print 'Q1' | def OnLeftClick(self, x, y, keys=0, attachment=0): shape = self.GetShape() canvas = shape.GetCanvas() dc = wx.ClientDC(canvas) canvas.PrepareDC(dc) if shape.Selected(): print 'Q1' shape.deselect(notify = True) else: shape.select(notify = True) | 8425b42d626c1d63eaccb8b24d93b33ec53d1571 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/3683/8425b42d626c1d63eaccb8b24d93b33ec53d1571/page.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2755,
3910,
6563,
12,
2890,
16,
619,
16,
677,
16,
1311,
33,
20,
16,
6042,
33,
20,
4672,
2179,
273,
365,
18,
967,
8500,
1435,
5953,
273,
2179,
18,
967,
12971,
1435,
6744,
273,
7075,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2755,
3910,
6563,
12,
2890,
16,
619,
16,
677,
16,
1311,
33,
20,
16,
6042,
33,
20,
4672,
2179,
273,
365,
18,
967,
8500,
1435,
5953,
273,
2179,
18,
967,
12971,
1435,
6744,
273,
7075,
1... | |
self._debug(NIGHTMARE, "wait") | def feed (self, data): if self.state=='parse': if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': return data = self.inbuf.getvalue() self.inbuf.close() self.inbuf = StringIO() if data: self._debug(NIGHTMARE, "feed", `data`) HtmlParser.feed(self, data) else: self._debug(NIGHTMARE, "feed") else: self._debug(NIGHTMARE, "wait") self.inbuf.write(data) | b6cdd9629ec41bf387c3e531943b23aec78afb66 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3948/b6cdd9629ec41bf387c3e531943b23aec78afb66/Rewriter.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4746,
261,
2890,
16,
501,
4672,
309,
365,
18,
2019,
18920,
2670,
4278,
309,
365,
18,
7048,
329,
30,
365,
18,
7048,
329,
273,
374,
2529,
4385,
16,
365,
18,
7048,
4385,
273,
365,
18,
7... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4746,
261,
2890,
16,
501,
4672,
309,
365,
18,
2019,
18920,
2670,
4278,
309,
365,
18,
7048,
329,
30,
365,
18,
7048,
329,
273,
374,
2529,
4385,
16,
365,
18,
7048,
4385,
273,
365,
18,
7... | |
self.cp.add_section(self.section) | if self.section is not None: self.cp.add_section(self.section) | def parse_config_files(self): self.cp.read(self.files) if not self.cp.has_section(self.section): self.cp.add_section(self.section) | b9086fdf39941343fde65551b8667e635e05f14f /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/14925/b9086fdf39941343fde65551b8667e635e05f14f/system_info.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
1425,
67,
2354,
12,
2890,
4672,
365,
18,
4057,
18,
896,
12,
2890,
18,
2354,
13,
309,
486,
365,
18,
4057,
18,
5332,
67,
3464,
12,
2890,
18,
3464,
4672,
365,
18,
4057,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
1425,
67,
2354,
12,
2890,
4672,
365,
18,
4057,
18,
896,
12,
2890,
18,
2354,
13,
309,
486,
365,
18,
4057,
18,
5332,
67,
3464,
12,
2890,
18,
3464,
4672,
365,
18,
4057,
18,
... |
item_amount_over = float(item['amount']) - float(category['balance']) | item_amount_over = item_amount - category_balance | def _update_category_balance(self, item): "Update the category balance from the item amount" if int(item['type']) == 1: table = "ExpenseCategory" elif int(item['type']) == 2: table = "BillCategory" elif int(item['type']) == 3: table = "SavingCategory" category_select = "select * from %s where id = :id;" % (table) category_update = "update %s set balance = :balance where id = :id;" % (table) | c1b405e7b76ee374ecb1e3fca3072ef6e0007414 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14271/c1b405e7b76ee374ecb1e3fca3072ef6e0007414/view.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2725,
67,
4743,
67,
12296,
12,
2890,
16,
761,
4672,
315,
1891,
326,
3150,
11013,
628,
326,
761,
3844,
6,
309,
509,
12,
1726,
3292,
723,
19486,
422,
404,
30,
1014,
273,
315,
424,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2725,
67,
4743,
67,
12296,
12,
2890,
16,
761,
4672,
315,
1891,
326,
3150,
11013,
628,
326,
761,
3844,
6,
309,
509,
12,
1726,
3292,
723,
19486,
422,
404,
30,
1014,
273,
315,
424,
... |
if d not in sys.path: sys.path.insert(0,d) | def find_test_files(L,d,N): n = os.path.basename(d) if n!='test' : return for n in filter(lambda n: n[-3:]=='.py',N): fn = os.path.normcase(os.path.normpath(os.path.join(d,n))) if os.path.isfile(fn): L.append(fn) | d144de574b55b6230e50ff50376a0264613d6725 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7053/d144de574b55b6230e50ff50376a0264613d6725/runtests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1104,
67,
3813,
67,
2354,
12,
48,
16,
72,
16,
50,
4672,
290,
273,
1140,
18,
803,
18,
13909,
12,
72,
13,
309,
290,
5,
2218,
3813,
11,
294,
327,
364,
290,
316,
1034,
12,
14661,
290,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1104,
67,
3813,
67,
2354,
12,
48,
16,
72,
16,
50,
4672,
290,
273,
1140,
18,
803,
18,
13909,
12,
72,
13,
309,
290,
5,
2218,
3813,
11,
294,
327,
364,
290,
316,
1034,
12,
14661,
290,
... | |
There is NO WARRANTY, to the extent permitted by law.""" % Common.APP_VERSION | There is NO WARRANTY, to the extent permitted by law.""" % APP_VERSION | def process_cmd_line(): """Parse the command line and execute given commands.""" # TRANSLATORS: This is the command line usage. Don't translate # %prog, but do translate usage, options, cleaner, and option. # More information about the command line is here # http://bleachbit.sourceforge.net/documentation/command-line usage = _("usage: %prog [options] cleaner.option1 cleaner.option2") parser = optparse.OptionParser(usage) parser.add_option("-l", "--list-cleaners", action = "store_true", help = _("list cleaners")) parser.add_option("-d", "--delete", action = "store_true", help = _("delete files and make other permanent changes")) parser.add_option("--sysinfo", action = "store_true", help = _("show system information")) parser.add_option("-p", "--preview", action = "store_true", help = _("preview files to be deleted and other changes")) parser.add_option("-v", "--version", action = "store_true", help = _("output version information and exit")) parser.add_option('-o', '--overwrite', action = 'store_true', help = _('overwrite files to hide contents')) (options, args) = parser.parse_args() if options.version: print """ | c0cb8114a155664217d64603d86ba5ee239056de /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7853/c0cb8114a155664217d64603d86ba5ee239056de/CLI.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
67,
4172,
67,
1369,
13332,
3536,
3201,
326,
1296,
980,
471,
1836,
864,
4364,
12123,
468,
14996,
4559,
3575,
55,
30,
1220,
353,
326,
1296,
980,
4084,
18,
225,
7615,
1404,
4204,
468,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
67,
4172,
67,
1369,
13332,
3536,
3201,
326,
1296,
980,
471,
1836,
864,
4364,
12123,
468,
14996,
4559,
3575,
55,
30,
1220,
353,
326,
1296,
980,
4084,
18,
225,
7615,
1404,
4204,
468,... |
settername = 'set' + 'DBus' + qt4name[0].upper() + qt4name[1:] | settername = set + gettername[0].upper() + gettername[1:] | def do_prop(self, prop): name = prop.getAttribute('name') qt4name = get_qt4_name(prop) qt4propname = 'dbus' + qt4name[0].upper() + qt4name[1:] access = prop.getAttribute('access') gettername = qt4propname settername = None | 9a5761bc4322fb97a8d63bad19a65a7811b662f7 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7593/9a5761bc4322fb97a8d63bad19a65a7811b662f7/qt4-client-gen.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
5986,
12,
2890,
16,
2270,
4672,
508,
273,
2270,
18,
588,
1499,
2668,
529,
6134,
25672,
24,
529,
273,
336,
67,
23311,
24,
67,
529,
12,
5986,
13,
25672,
24,
5986,
529,
273,
29... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
5986,
12,
2890,
16,
2270,
4672,
508,
273,
2270,
18,
588,
1499,
2668,
529,
6134,
25672,
24,
529,
273,
336,
67,
23311,
24,
67,
529,
12,
5986,
13,
25672,
24,
5986,
529,
273,
29... |
self.hboxlayout61 = QtGui.QHBoxLayout() self.hboxlayout61.setMargin(0) self.hboxlayout61.setSpacing(6) self.hboxlayout61.setObjectName("hboxlayout61") | self.hboxlayout64 = QtGui.QHBoxLayout() self.hboxlayout64.setMargin(0) self.hboxlayout64.setSpacing(6) self.hboxlayout64.setObjectName("hboxlayout64") | def setupUi(self, UserPrefsDialog): UserPrefsDialog.setObjectName("UserPrefsDialog") UserPrefsDialog.resize(QtCore.QSize(QtCore.QRect(0,0,609,642).size()).expandedTo(UserPrefsDialog.minimumSizeHint())) | 30deb2219e016847d91b37e801c50b75fbd1c3cf /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/11221/30deb2219e016847d91b37e801c50b75fbd1c3cf/UserPrefsDialog.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
13943,
12,
2890,
16,
2177,
1386,
2556,
6353,
4672,
2177,
1386,
2556,
6353,
18,
542,
16707,
2932,
1299,
1386,
2556,
6353,
7923,
2177,
1386,
2556,
6353,
18,
15169,
12,
23310,
4670,
18,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
13943,
12,
2890,
16,
2177,
1386,
2556,
6353,
4672,
2177,
1386,
2556,
6353,
18,
542,
16707,
2932,
1299,
1386,
2556,
6353,
7923,
2177,
1386,
2556,
6353,
18,
15169,
12,
23310,
4670,
18,... |
conj = t.conjugate() for row in conj: for i in range(1, len(row)): if row[i] <= row[i-1]: return False | if i > 0 and entry <= t[i-1][j]: return False if self.max_entry is not None and row[-1] > self.max_entry: return False | def __contains__(self, x): """ TESTS:: sage: [[1,2],[1]] in SemistandardTableaux() False sage: SST = SemistandardTableaux() sage: all([st in SST for st in StandardTableaux(4)]) True sage: [[1,1],[2]] in SemistandardTableaux() True """ if x not in Tableaux(): return False else: t = Tableau(x) #Check to make sure the first position is 1 for row in t: for i in row: if not isinstance(i, (int, Integer)): return False | dc756d26ea922387ee74781c3b99359e739e8c01 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9890/dc756d26ea922387ee74781c3b99359e739e8c01/tableau.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
12298,
972,
12,
2890,
16,
619,
4672,
3536,
22130,
55,
2866,
225,
272,
410,
30,
12167,
21,
16,
22,
6487,
63,
21,
13563,
316,
15661,
376,
2958,
1388,
18196,
1435,
1083,
272,
410,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
12298,
972,
12,
2890,
16,
619,
4672,
3536,
22130,
55,
2866,
225,
272,
410,
30,
12167,
21,
16,
22,
6487,
63,
21,
13563,
316,
15661,
376,
2958,
1388,
18196,
1435,
1083,
272,
410,
3... |
prints('Using proxies: %s' % proxies) | prints('Using proxies:', proxies) | def get_proxies(): proxies = {} for q in ('http', 'ftp'): proxy = os.environ.get(q+'_proxy', None) if not proxy: continue if proxy.startswith(q+'://'): proxy = proxy[7:] proxies[q] = proxy if iswindows: try: winreg = __import__('_winreg') settings = winreg.OpenKey(winreg.HKEY_CURRENT_USER, 'Software\\Microsoft\\Windows' '\\CurrentVersion\\Internet Settings') proxy = winreg.QueryValueEx(settings, "ProxyEnable")[0] if proxy: server = str(winreg.QueryValueEx(settings, 'ProxyServer')[0]) if ';' in server: for p in server.split(';'): protocol, address = p.split('=') proxies[protocol] = address else: proxies['http'] = server proxies['ftp'] = server settings.Close() except Exception, e: prints('Unable to detect proxy settings: %s' % str(e)) for x in list(proxies): if len(proxies[x]) < 5: prints('Removing invalid', x, 'proxy:', proxies[x]) del proxies[x] if proxies: prints('Using proxies: %s' % proxies) return proxies | 8819fe25883871bd70e6d6b147aedd10f32b09e5 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9125/8819fe25883871bd70e6d6b147aedd10f32b09e5/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
20314,
606,
13332,
13263,
273,
2618,
225,
364,
1043,
316,
7707,
2505,
2187,
296,
11727,
11,
4672,
2889,
273,
225,
1140,
18,
28684,
18,
588,
12,
85,
6797,
67,
5656,
2187,
599,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
20314,
606,
13332,
13263,
273,
2618,
225,
364,
1043,
316,
7707,
2505,
2187,
296,
11727,
11,
4672,
2889,
273,
225,
1140,
18,
28684,
18,
588,
12,
85,
6797,
67,
5656,
2187,
599,
... |
smap = {'h':2, 'm':1, 'l':0} | smap = {'h':3, 'm':2, 'l':1, 'f':0} | def sortFunc(x, y): if not str(x).isdigit(): smap = {'h':2, 'm':1, 'l':0} return cmp(smap[y[0]], smap[x[0]]) else: return cmp (int(y), int(x)) | 28a20f0ce2e5b0def972f36c16ceda1f7997bb39 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7242/28a20f0ce2e5b0def972f36c16ceda1f7997bb39/Actor.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1524,
2622,
12,
92,
16,
677,
4672,
309,
486,
609,
12,
92,
2934,
291,
11052,
13332,
272,
1458,
273,
13666,
76,
4278,
23,
16,
296,
81,
4278,
22,
16,
296,
80,
4278,
21,
16,
296,
74,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1524,
2622,
12,
92,
16,
677,
4672,
309,
486,
609,
12,
92,
2934,
291,
11052,
13332,
272,
1458,
273,
13666,
76,
4278,
23,
16,
296,
81,
4278,
22,
16,
296,
80,
4278,
21,
16,
296,
74,
4... |
print """WARNING: Your imap server uses commas as the folder delimiter. This may cause unpredictable errors.""" folders.append(fol[m.end()+5:-1]) | print "WARNING: Your imap server uses a comma as the " \ "folder delimiter. This may cause unpredictable " \ "errors." folders.append(fol[m.end()+4:].strip('"')) | def folder_list(self): '''Return a alphabetical list of all folders available on the server''' response = self.list() if response[0] != "OK": return [] all_folders = response[1] folders = [] for fol in all_folders: # Sigh. Some servers may give us back the folder name as a # literal, so we need to crunch this out. if isinstance(fol, ()): r = re.compile(r"{\d+}") m = r.search(fol[0]) if not m: # Something is wrong here! Skip this folder continue fol = '%s"%s"' % (fol[0][:m.start()], fol[1]) r = re.compile(r"\(([\w\\ ]*)\) ") m = r.search(fol) if not m: # Something is not good with this folder, so skip it. continue name_attributes = fol[:m.end()-1] # IMAP is a truly odd protocol. The delimiter is # only the delimiter for this particular folder - each # folder *may* have a different delimiter self.folder_delimiter = fol[m.end()+1:m.end()+2] # a bit of a hack, but we really need to know if this is # the case if self.folder_delimiter == ',': print """WARNING: Your imap server uses commas as the folder delimiter. This may cause unpredictable errors.""" folders.append(fol[m.end()+5:-1]) folders.sort() return folders | ccd6023de3c36b59a81e3402155d544e2dfd2686 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9857/ccd6023de3c36b59a81e3402155d544e2dfd2686/sb_imapfilter.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3009,
67,
1098,
12,
2890,
4672,
9163,
990,
279,
10877,
1706,
666,
434,
777,
9907,
2319,
603,
326,
1438,
26418,
766,
273,
365,
18,
1098,
1435,
309,
766,
63,
20,
65,
480,
315,
3141,
6877... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3009,
67,
1098,
12,
2890,
4672,
9163,
990,
279,
10877,
1706,
666,
434,
777,
9907,
2319,
603,
326,
1438,
26418,
766,
273,
365,
18,
1098,
1435,
309,
766,
63,
20,
65,
480,
315,
3141,
6877... |
out('%s\\textit{(%s=%s)}\n\n' % (' '*12, lhs, rhs)) | out('%s{\\it (%s=%s)}\n\n' % (' '*12, lhs, rhs)) | def write_func_list_box(self, out, var_doc): func_doc = var_doc.value is_inherited = (var_doc.overrides not in (None, UNKNOWN)) | 90de12e26850e253bd23f33820a3991b0c1c0a83 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/3512/90de12e26850e253bd23f33820a3991b0c1c0a83/latex.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
67,
644,
67,
1098,
67,
2147,
12,
2890,
16,
596,
16,
569,
67,
2434,
4672,
1326,
67,
2434,
273,
569,
67,
2434,
18,
1132,
353,
67,
6018,
329,
273,
261,
1401,
67,
2434,
18,
19775,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
67,
644,
67,
1098,
67,
2147,
12,
2890,
16,
596,
16,
569,
67,
2434,
4672,
1326,
67,
2434,
273,
569,
67,
2434,
18,
1132,
353,
67,
6018,
329,
273,
261,
1401,
67,
2434,
18,
19775,
... |
def getCatalogIsLink( self, lfn, singleFile=False, catalogs=[] ): | def getCatalogIsLink( self, lfn, singleFile = False, catalogs = [] ): | def getCatalogIsLink( self, lfn, singleFile=False, catalogs=[] ): """ Determine whether the path is registered as a link in the FileCatalog 'lfn' is the paths to check (can be a single path or list of paths) """ if singleFile: return self._executeSingleFileCatalogFunction( lfn, 'isLink', catalogs=catalogs ) else: return self._executeFileCatalogFunction( lfn, 'isLink', catalogs=catalogs ) | 8cfc8d124489ba672254aee3d1fff4f748757c6c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/8cfc8d124489ba672254aee3d1fff4f748757c6c/ReplicaManager.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1927,
4392,
2520,
2098,
12,
365,
16,
328,
4293,
16,
2202,
812,
273,
1083,
16,
26455,
273,
5378,
262,
30,
3536,
10229,
2856,
326,
589,
353,
4104,
487,
279,
1692,
316,
326,
1387,
9769,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1927,
4392,
2520,
2098,
12,
365,
16,
328,
4293,
16,
2202,
812,
273,
1083,
16,
26455,
273,
5378,
262,
30,
3536,
10229,
2856,
326,
589,
353,
4104,
487,
279,
1692,
316,
326,
1387,
9769,
2... |
self.inclusions.append(inclusions) | self.inclusions.append(inclusions) | def updateArgs(self, *args, **keywords): """ The default way to update a class is to override any provided keywords. Subclasses which have the ability to provide more intelligent handling can override this method. This method is invoked automatically by recipe.py when a recipe references a policy object. It acts rather like __init__ except that it can meaningfully be called more than once for an object. | f1fb636b63d79eefba73e71873bed9081455b3cc /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8747/f1fb636b63d79eefba73e71873bed9081455b3cc/policy.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
2615,
12,
2890,
16,
380,
1968,
16,
2826,
11771,
4672,
3536,
1021,
805,
4031,
358,
1089,
279,
667,
353,
358,
3849,
1281,
2112,
7093,
18,
225,
2592,
4701,
1492,
1240,
326,
7123,
358,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
2615,
12,
2890,
16,
380,
1968,
16,
2826,
11771,
4672,
3536,
1021,
805,
4031,
358,
1089,
279,
667,
353,
358,
3849,
1281,
2112,
7093,
18,
225,
2592,
4701,
1492,
1240,
326,
7123,
358,... |
+ [self.base_ring()(0, absprec = self._relprecs[i] + self._valbase) for i in range(polylen, len(self._relprecs))] | + [self.base_ring()(0, absprec = self._relprecs[i] + self._valaddeds[i] + self._valbase) for i in range(polylen, len(self._relprecs))] while self._list[-1]._is_exact_zero(): self._list.pop() | def _comp_list(self): if self.degree() == -1 and self._valbase == infinity: self._list = [] return self._list polylist = self._poly.list() polylen = len(polylist) self._list = [self.base_ring()(polylist[i], absprec = self._relprecs[i]) << self._valbase for i in range(polylen)] \ + [self.base_ring()(0, absprec = self._relprecs[i] + self._valbase) for i in range(polylen, len(self._relprecs))] | 425d148bd6d48946166c40b08c148463145b804f /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/425d148bd6d48946166c40b08c148463145b804f/polynomial_padic_capped_relative_dense.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2919,
67,
1098,
12,
2890,
4672,
309,
365,
18,
21361,
1435,
422,
300,
21,
471,
365,
6315,
1125,
1969,
422,
27272,
30,
365,
6315,
1098,
273,
5378,
327,
365,
6315,
1098,
7573,
1098,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2919,
67,
1098,
12,
2890,
4672,
309,
365,
18,
21361,
1435,
422,
300,
21,
471,
365,
6315,
1125,
1969,
422,
27272,
30,
365,
6315,
1098,
273,
5378,
327,
365,
6315,
1098,
7573,
1098,
... |
return self._getCachedInfo()['images'][size] | return self._getCachedInfo('images', size) | def getImage(self, size = IMAGE_LARGE): """Returns the associated image URL. * size: The image size. Possible values: o IMAGE_LARGE o IMAGE_MEDIUM o IMAGE_SMALL """ return self._getCachedInfo()['images'][size] | ec1304614ea5bc838c2f5bb83444fc59ac9c26d8 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9926/ec1304614ea5bc838c2f5bb83444fc59ac9c26d8/pylast.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10567,
12,
2890,
16,
963,
273,
10757,
67,
48,
28847,
4672,
3536,
1356,
326,
3627,
1316,
1976,
18,
380,
963,
30,
1021,
1316,
963,
18,
25433,
924,
30,
320,
10757,
67,
48,
28847,
320,
107... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10567,
12,
2890,
16,
963,
273,
10757,
67,
48,
28847,
4672,
3536,
1356,
326,
3627,
1316,
1976,
18,
380,
963,
30,
1021,
1316,
963,
18,
25433,
924,
30,
320,
10757,
67,
48,
28847,
320,
107... |
rdata = self.tryToResolveConflict(oid, oserial, serial, data) if rdata is None: raise POSException.ConflictError( oid=oid, serials=(oserial, serial), data=data) else: data = rdata | data=self.tryToResolveConflict(oid, oserial, serial, data) if not data: raise POSException.ConflictError(oid=oid, serials=(oserial, serial)) | def store(self, oid, serial, data, version, transaction): if transaction is not self._transaction: raise POSException.StorageTransactionError(self, transaction) if version: # we allow a version to be in use although we don't # support versions in the storage. LOG('TemporaryStorage', BLATHER, ('versions in use with TemporaryStorage although Temporary' 'Storage doesnt support versions'), ) self._lock_acquire() try: if self._index.has_key(oid): oserial=self._index[oid] if serial != oserial: rdata = self.tryToResolveConflict(oid, oserial, serial, data) if rdata is None: raise POSException.ConflictError( oid=oid, serials=(oserial, serial), data=data) else: data = rdata else: oserial = serial newserial=self._tid self._tmp.append((oid, data)) now = time.time() self._conflict_cache[(oid, newserial)] = data, now return serial == oserial and newserial or ResolvedSerial finally: self._lock_release() | 1a676b2cb5faf9426357ce4e2e3a36b941b8d277 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9658/1a676b2cb5faf9426357ce4e2e3a36b941b8d277/TemporaryStorage.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1707,
12,
2890,
16,
7764,
16,
2734,
16,
501,
16,
1177,
16,
2492,
4672,
309,
2492,
353,
486,
365,
6315,
7958,
30,
1002,
12511,
503,
18,
3245,
3342,
668,
12,
2890,
16,
2492,
13,
309,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1707,
12,
2890,
16,
7764,
16,
2734,
16,
501,
16,
1177,
16,
2492,
4672,
309,
2492,
353,
486,
365,
6315,
7958,
30,
1002,
12511,
503,
18,
3245,
3342,
668,
12,
2890,
16,
2492,
13,
309,
1... |
iconset += Icon, | if Icon: iconset += Icon, | def run(self, Theme, IconDB, Store, iconset_data): context = iconset_data[2] name = iconset_data[1] | f9700d18c4ee95628d8e1a11f3ae4de00f6c5680 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3955/f9700d18c4ee95628d8e1a11f3ae4de00f6c5680/editor.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
16,
19745,
16,
16011,
2290,
16,
4994,
16,
4126,
542,
67,
892,
4672,
819,
273,
4126,
542,
67,
892,
63,
22,
65,
508,
273,
4126,
542,
67,
892,
63,
21,
65,
2,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
16,
19745,
16,
16011,
2290,
16,
4994,
16,
4126,
542,
67,
892,
4672,
819,
273,
4126,
542,
67,
892,
63,
22,
65,
508,
273,
4126,
542,
67,
892,
63,
21,
65,
2,
-100,
-10... |
fext = self._getExtension(fname) | fext = self._getExtension(fileName) | def load(self, fname): ''' Load data from a csv or a pickle file of the DataStore class. The encoding is determined by the filename's extension: 'csv' : When the filename ends in '.csv' the routine tries to interpret the file as comma seperated values. The variable names must be in the first row Any other extension is considered to mean a file in Python's pickle format. Arguments: fname : filename; string Returns: None ''' # setting the ascii/csv file name used for input #self.DBname = os.getcwd() + '/' + fname | 45268ff49ddb4f963530947d7202a5e0642e2fda /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2720/45268ff49ddb4f963530947d7202a5e0642e2fda/storage.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
12,
2890,
16,
5299,
4672,
9163,
4444,
501,
628,
279,
6101,
578,
279,
13379,
585,
434,
326,
22961,
667,
18,
225,
1021,
2688,
353,
11383,
635,
326,
1544,
1807,
2710,
30,
296,
6715,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
12,
2890,
16,
5299,
4672,
9163,
4444,
501,
628,
279,
6101,
578,
279,
13379,
585,
434,
326,
22961,
667,
18,
225,
1021,
2688,
353,
11383,
635,
326,
1544,
1807,
2710,
30,
296,
6715,
... |
/.../RunChandler... --catch=tests --profileDir=test_profile --parcelPath=tools/cats/DataFiles --catsPerfLog=test_profile/time.log --scriptFile=tools/QATestScripts/Performance/PerfImportCalendar.py --create | /.../chandler... --catch=tests --profileDir=test_profile --parcelPath=tools/cats/DataFiles --catsPerfLog=test_profile/time.log --scriptFile=tools/QATestScripts/Performance/PerfImportCalendar.py --create | def main(options): """ >>> options = parseOptions() >>> options.dryrun = True >>> options.verbose = True >>> main(options) False Try and run a test that does not exist >>> options.single = 'TestFoo.py' >>> main(options) Test(s) not found False Try different single tests single unit test: >>> options.single = 'TestCrypto' >>> main(options) /.../RunPython... application/tests/TestCrypto.py -v - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + False unit test and functional test: >>> options.single = 'TestCrypto,TestSharing' >>> main(options) /.../RunChandler... --create --catch=tests --profileDir=test_profile --parcelPath=tools/cats/DataFiles --chandlerTests=TestSharing -D2 -M0 - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + /.../RunPython... application/tests/TestCrypto.py -v - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + False unit, functional and two perf tests, one of which is a startup test: >>> options.single = 'TestCrypto,TestSharing,PerfImportCalendar,startup_large' >>> main(options) /.../RunChandler... --catch=tests --profileDir=test_profile --parcelPath=tools/cats/DataFiles --catsPerfLog=test_profile/time.log --scriptFile=tools/QATestScripts/Performance/PerfImportCalendar.py --create PerfImportCalendar.py 0.00 - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + | 0.00 ... 0.00 /.../RunChandler... --create --catch=tests --profileDir=test_profile --parcelPath=tools/cats/DataFiles --chandlerTests=TestSharing -D2 -M0 ... /.../RunPython... application/tests/TestCrypto.py -v ... Creating repository for startup time tests ... Startup_with_large_calendar ... ... False Try and specify an invalid mode >>> options.single = '' >>> options.mode = 'foo' >>> main(options) foo removed from mode list foo mode requested but not found -- stopping test run True Run unit tests with --dryrun >>> options.mode = None >>> options.unit = True >>> main(options) /.../RunPython... .../tests/TestReferenceAttributes.py -v ... /.../RunPython... setup.py test -v ... False Run unitSuite with --dryrun >>> options.unit = False >>> options.unitSuite = True >>> main(options) /.../RunPython... tools/run_tests.py -v application i18n osaf repository - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + False Run functional tests with --dryrun >>> options.unit = False >>> options.unitSuite = False >>> options.funcSuite = True >>> main(options) /.../RunChandler... --create --catch=tests --profileDir=test_profile --parcelPath=tools/cats/DataFiles --scriptFile=tools/cats/Functional/FunctionalTestSuite.py -D2 -M0 - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + False Run functional tests each in its on process >>> options.funcSuite = False >>> options.func = True >>> main(options) /.../RunChandler... --create --catch=tests --profileDir=test_profile --parcelPath=tools/cats/DataFiles --chandlerTests=TestAllDayEvent -D2 -M0 - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + ... False Run performance tests with --dryrun >>> options.func = False >>> options.perf = True >>> options.profile = False >>> main(options) /.../RunChandler... --catch=tests --profileDir=test_profile --parcelPath=tools/cats/DataFiles --catsPerfLog=test_profile/time.log --scriptFile=tools/QATestScripts/Performance/PerfImportCalendar.py --create PerfImportCalendar.py ... ... /.../RunChandler... --catch=tests --profileDir=test_profile --parcelPath=tools/cats/DataFiles --catsPerfLog=test_profile/time.log --scriptFile=tools/QATestScripts/Performance/PerfLargeDataResizeCalendar.py --restore=test_profile/__repository__.001 PerfLargeDataResizeCalendar.py ... ... Creating repository for startup time tests ... Startup ... ... False """ checkOptions(options) failed = False if options.mode is None: options.modes = modes = ['release', 'debug'] # silently clear any missing modes if default list is specified for mode in modes: if not os.path.isdir(os.path.join(options.chandlerBin, mode)): options.modes.remove(mode) else: options.mode = options.mode.strip().lower() options.modes = [options.mode] # complain about any missing modes if mode was explicitly stated if not os.path.isdir(os.path.join(options.chandlerBin, options.mode)): options.modes.remove(options.mode) log('%s removed from mode list' % options.mode) if len(options.modes) == 0: log('%s mode requested but not found -- stopping test run' % options.mode) failed = True if not failed: try: # Empty the log file so that we won't be confused by old results later f = open(os.path.join(options.profileDir, 'chandler.log'), 'w') f.close() except IOError: pass # Remove old perf log files (we leave the the latest) for f in glob.glob(os.path.join(options.profileDir, '*.log.*')): try: os.remove(f) except OSError: pass if options.testcase: failed = runTestCase(options) elif options.single: failed = runSingles(options) else: if not options.perf: failed = runLocalizationCheck(options) if options.unit and (not failed or options.noStop): failed = runUnitTests(options) if not failed or options.noStop: if runPluginTests(options): failed = True if options.unitSuite and (not failed or options.noStop): if runUnitSuite(options): failed = True if options.funcSuite and (not failed or options.noStop): if runFuncTest(options): failed = True if options.func and (not failed or options.noStop): if runFuncTestsSingly(options): failed = True if options.recorded and (not failed or options.noStop): if runRecordedScripts(options): failed = True if options.perf and (not failed or options.noStop): if runPerfTests(options): failed = True print 'At end Recorded script status is :', failed if len(failedTests) > 0: log('+-' * 32) log('The following tests failed:') log('\n'.join(failedTests)) log('') return failed | 2268248f0e4728b63ca1842648cbad866d679472 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9228/2268248f0e4728b63ca1842648cbad866d679472/rt.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
2116,
4672,
3536,
4080,
702,
273,
1109,
1320,
1435,
4080,
702,
18,
25011,
2681,
225,
273,
1053,
4080,
702,
18,
11369,
273,
1053,
4080,
2774,
12,
2116,
13,
1083,
225,
6161,
471,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
2116,
4672,
3536,
4080,
702,
273,
1109,
1320,
1435,
4080,
702,
18,
25011,
2681,
225,
273,
1053,
4080,
702,
18,
11369,
273,
1053,
4080,
2774,
12,
2116,
13,
1083,
225,
6161,
471,... |
return "<%s%s id=%r %s, %s active channels>" %( self.__class__.__name__, addr, id, r, i) | return "<%s id=%r %s, %s active channels>" %( self.__class__.__name__, self.id, r, i) | def __repr__(self): """ return string representing gateway type and status. """ if hasattr(self, 'id'): id = self.id else: id = "???" if hasattr(self, 'remoteaddress'): addr = '[%s]' % (self.remoteaddress,) else: addr = '' try: r = (self._receiverthread.isAlive() and "receive-live" or "not-receiving") i = len(self._channelfactory.channels()) except AttributeError: r = "uninitialized" i = "no" return "<%s%s id=%r %s, %s active channels>" %( self.__class__.__name__, addr, id, r, i) | 647c4a943e7d6c2f4b39791d4787ed60759837ca /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7009/647c4a943e7d6c2f4b39791d4787ed60759837ca/gateway.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
12715,
972,
12,
2890,
4672,
3536,
327,
533,
5123,
6878,
618,
471,
1267,
18,
3536,
309,
3859,
12,
2890,
16,
296,
350,
11,
4672,
612,
273,
365,
18,
350,
469,
30,
612,
273,
315,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
12715,
972,
12,
2890,
4672,
3536,
327,
533,
5123,
6878,
618,
471,
1267,
18,
3536,
309,
3859,
12,
2890,
16,
296,
350,
11,
4672,
612,
273,
365,
18,
350,
469,
30,
612,
273,
315,
1... |
def handle(self, verbose=False, make_daemon=False, loop=False, timeout=10, rebuild_index=False, per_page=1000, commit_each=False, *args, **options): | def handle(self, *app_labels, **options): verbose = options['verbose'] make_daemon = options['make_daemon'] loop = options['loop'] timeout = options['timeout'] rebuild_index = options['rebuild_index'] per_page = options['per_page'] commit_each = options['commit_each'] | def handle(self, verbose=False, make_daemon=False, loop=False, timeout=10, rebuild_index=False, per_page=1000, commit_each=False, *args, **options): utils.load_indexes() | 37d63969ac742928cd0a05ebf798f66bfa0d226e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4645/37d63969ac742928cd0a05ebf798f66bfa0d226e/index.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
12,
2890,
16,
380,
2910,
67,
5336,
16,
2826,
2116,
4672,
3988,
273,
702,
3292,
11369,
3546,
225,
1221,
67,
21511,
273,
702,
3292,
6540,
67,
21511,
3546,
2798,
273,
702,
3292,
6498,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
12,
2890,
16,
380,
2910,
67,
5336,
16,
2826,
2116,
4672,
3988,
273,
702,
3292,
11369,
3546,
225,
1221,
67,
21511,
273,
702,
3292,
6540,
67,
21511,
3546,
2798,
273,
702,
3292,
6498,... |
backgroundSubtractedExposure = exposure.Factory(exposure, bbox, true) | backgroundSubtractedExposure = exposure.Factory(exposure, bbox, True) | def estimateBackground(exposure, backgroundPolicy, subtract=True): """Estimate exposure's background using parameters in backgroundPolicy. If subtract is true, make a copy of the exposure and subtract the background. Return background, backgroundSubtractedExposure""" maskedImage = exposure.getMaskedImage() bbox = afwImg.BBox(maskedImage.getXY0(), maskedImage.getWidth(), maskedImage.getHeight()) backgroundSubtractedExposure = exposure.Factory(exposure, bbox, true) image = maskedImage.getImage() background = getBackground(image, backgroundPolicy) del maskedImage del image if not background: raise RuntimeError, "Unable to estimate background for exposure" if not subtract: return background, None copyImage = backgroundSubtractedExposure.getMaskedImage().getImage() copyImage -= background.getImageF() del copyImage return background, backgroundSubtractedExposure | 2868db0dc0ba271f645d5a1b398fa213f2049a4c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6523/2868db0dc0ba271f645d5a1b398fa213f2049a4c/sourceDetection.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11108,
8199,
12,
22739,
16,
5412,
2582,
16,
10418,
33,
5510,
4672,
3536,
13638,
14620,
1807,
5412,
1450,
1472,
316,
5412,
2582,
18,
225,
971,
10418,
353,
638,
16,
1221,
279,
1610,
434,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11108,
8199,
12,
22739,
16,
5412,
2582,
16,
10418,
33,
5510,
4672,
3536,
13638,
14620,
1807,
5412,
1450,
1472,
316,
5412,
2582,
18,
225,
971,
10418,
353,
638,
16,
1221,
279,
1610,
434,
3... |
return html_quote(buf) | return html_quote(buf, self._lang) | def quote(self, buf): return html_quote(buf) | 469b12e8cfc912874019fdd0ddc95060d27b98f4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2120/469b12e8cfc912874019fdd0ddc95060d27b98f4/HyperArch.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3862,
12,
2890,
16,
1681,
4672,
327,
1729,
67,
6889,
12,
4385,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3862,
12,
2890,
16,
1681,
4672,
327,
1729,
67,
6889,
12,
4385,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
return http.Response(stream=message('No such worksheet.')) | return HTMLResponse(stream=message('No such worksheet.')) | def childFactory(self, request, name): worksheet_name = self.name filename = tmp_filename() + '.sws' try: notebook.export_worksheet(worksheet_name, filename) except KeyError: return http.Response(stream=message('No such worksheet.')) r = open(filename, 'rb').read() os.unlink(filename) return static.Data(r, 'application/sage') #return static.File(filename) | 5da3b1cd72cd4a94421b8bc914d7957cdbe6b56d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/5da3b1cd72cd4a94421b8bc914d7957cdbe6b56d/twist.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1151,
1733,
12,
2890,
16,
590,
16,
508,
4672,
13622,
67,
529,
273,
365,
18,
529,
1544,
273,
1853,
67,
3459,
1435,
397,
2418,
87,
4749,
11,
775,
30,
14718,
18,
6530,
67,
1252,
8118,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1151,
1733,
12,
2890,
16,
590,
16,
508,
4672,
13622,
67,
529,
273,
365,
18,
529,
1544,
273,
1853,
67,
3459,
1435,
397,
2418,
87,
4749,
11,
775,
30,
14718,
18,
6530,
67,
1252,
8118,
1... |
if os.path.exists(os.path.join(root_dir, ".git")): | if (os.path.exists(os.path.join(root_dir, ".git")) or os.path.exists(os.path.join(root_dir, ".hg"))): | def RepositoryName(self): """FullName after removing the local path to the repository. | ce282e4b1d1ad36173611564c5ff89d9b7bbb130 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8529/ce282e4b1d1ad36173611564c5ff89d9b7bbb130/cpplint.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6281,
461,
12,
2890,
4672,
3536,
19223,
1839,
9427,
326,
1191,
589,
358,
326,
3352,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6281,
461,
12,
2890,
4672,
3536,
19223,
1839,
9427,
326,
1191,
589,
358,
326,
3352,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
user_info.append(unicode(full_name).encode('utf-8')) | if isinstance(full_name, unicode): full_name = full_name.encode('utf-8') user_info.append(full_name) | def downloadUsersCsv(self, REQUEST=None, RESPONSE=None): """ Return a csv file as a session response. """ if not (REQUEST and RESPONSE): return "" | 6f9e9ee84e324982e556b853b3e9d25c1bca025a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3287/6f9e9ee84e324982e556b853b3e9d25c1bca025a/AuthenticationTool.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4224,
6588,
16759,
12,
2890,
16,
12492,
33,
7036,
16,
20645,
33,
7036,
4672,
3536,
2000,
279,
6101,
585,
487,
279,
1339,
766,
18,
3536,
309,
486,
261,
5519,
471,
20645,
4672,
327,
1408,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4224,
6588,
16759,
12,
2890,
16,
12492,
33,
7036,
16,
20645,
33,
7036,
4672,
3536,
2000,
279,
6101,
585,
487,
279,
1339,
766,
18,
3536,
309,
486,
261,
5519,
471,
20645,
4672,
327,
1408,
... |
res = res_inv = res_cus = {} | res = {} res_inv = {} res_cus = {} | def get_kml(self, cr, uid, context={}): res = res_inv = res_cus = {} address = ' ' coordinates = addresslist = country_list = [] coordinates_text = ' ' number_customer = number_supplier = 0 colors = ['9f8080ff', '9f0000ff'] partner_obj = self.pool.get('res.partner') address_obj= self.pool.get('res.partner.address') partner_ids = partner_obj.search(cr, uid, []) partners = partner_obj.browse(cr, uid, partner_ids) for part in partners: if part.address and part.address[0].country_id and part.address[0].country_id.name: if not string.upper(part.address[0].country_id.name) in country_list: cntry = string.upper(str(part.address[0].country_id.name)) country_name = '' for char in cntry: if char == '&': country_name += 'AND' else: country_name += char country_list.append(country_name) map(lambda x:res.setdefault(x, 0.0), country_list) cr.execute("select sum(l.debit-l.credit), c.name from account_move_line l, res_country c, res_partner_address a, account_account act where l.partner_id = a.partner_id and c.id=a.country_id and l.account_id = act.id and act.type = 'receivable' group by c.name") res_partner = cr.fetchall() list_to = [] for part in res_partner: if part[1]: res[string.upper(part[1])] = part[0] list_to.append(part[0]) | aaaffb7a7ee3c7d260a070bf8a5e141e1c996850 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7339/aaaffb7a7ee3c7d260a070bf8a5e141e1c996850/google_earth.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
79,
781,
12,
2890,
16,
4422,
16,
4555,
16,
819,
12938,
4672,
400,
273,
2618,
400,
67,
5768,
273,
2618,
400,
67,
71,
407,
273,
2618,
1758,
273,
296,
296,
5513,
273,
1758,
109... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
79,
781,
12,
2890,
16,
4422,
16,
4555,
16,
819,
12938,
4672,
400,
273,
2618,
400,
67,
5768,
273,
2618,
400,
67,
71,
407,
273,
2618,
1758,
273,
296,
296,
5513,
273,
1758,
109... |
raise Error, 'Filename too long' | raise Error, 'Filename too long' | def _writeinfo(self, name, finfo): | cada31e8208b1e7201f5afc9f9782089bec7ebe3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/cada31e8208b1e7201f5afc9f9782089bec7ebe3/binhex.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2626,
1376,
12,
2890,
16,
508,
16,
15190,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2626,
1376,
12,
2890,
16,
508,
16,
15190,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
self.conn.send.command_list_begin() | def stream_parse_and_add(self, item): # We need to do different things depending on if this is # a normal stream, pls, m3u, etc.. # Note that we will only download the first 2000 bytes f = None try: request = urllib2.Request(item) opener = urllib2.build_opener() f = opener.open(request).read(2000) except: try: request = urllib2.Request("http://" + item) opener = urllib2.build_opener() f = opener.open(request).read(2000) except: try: request = urllib2.Request("file://" + item) opener = urllib2.build_opener() f = opener.open(request).read(2000) except: pass if f: self.conn.send.command_list_begin() if is_binary(f): # Binary file, just add it: self.conn.send.add(item) else: if "[playlist]" in f: # pls: self.stream_parse_pls(f) elif "#EXTM3U" in f: # extended m3u: self.stream_parse_m3u(f) elif "http://" in f: # m3u or generic list: self.stream_parse_m3u(f) else: # Something else.. self.conn.send.add(item) self.conn.do.command_list_end() else: # Hopefully just a regular stream, try to add it: self.conn.do.add(item) | 0da9815c248298a9268c10f25c59ea81e5bd319d /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2312/0da9815c248298a9268c10f25c59ea81e5bd319d/sonata.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1407,
67,
2670,
67,
464,
67,
1289,
12,
2890,
16,
761,
4672,
468,
1660,
1608,
358,
741,
3775,
9198,
8353,
603,
309,
333,
353,
468,
279,
2212,
1407,
16,
886,
87,
16,
312,
23,
89,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1407,
67,
2670,
67,
464,
67,
1289,
12,
2890,
16,
761,
4672,
468,
1660,
1608,
358,
741,
3775,
9198,
8353,
603,
309,
333,
353,
468,
279,
2212,
1407,
16,
886,
87,
16,
312,
23,
89,
16,
... | |
constantNode.set("value", str(__variantMatchKey(variantValue, variantMap, variantGroup)).lower()) | constantNode.set("value", str(__variantMatchKey(variantValue, variantMap, variantKey)).lower()) | def processVariantIsSet(callNode, variantMap): ''' processes qx.core.Variant.isSet() calls; destructive! re-writes the AST tree passed in [callNode] by replacing choices with the suitable branch ''' if callNode.type != "call": return False params = callNode.getChild("params") if len(params.children) != 2: log("Warning", "Expecting exactly two arguments for qx.core.Variant.isSet. Ignoring this occurrence.", params) return False firstParam = params.getChildByPosition(0) if not isStringLiteral(firstParam): log("Warning", "First argument must be a string literal! Ignoring this occurrence.", firstParam) return False variantGroup = firstParam.get("value"); if not variantGroup in variantMap.keys(): return False secondParam = params.getChildByPosition(1) if isStringLiteral(secondParam): ifcondition = secondParam.parent.parent.parent # normal if then else if ifcondition.type == "expression" and ifcondition.getChildrenLength(True) == 1 and ifcondition.parent.type == "loop": loop = ifcondition.parent variantValue = secondParam.get("value") inlineIfStatement(loop, __variantMatchKey(variantValue, variantMap, variantGroup)) # ternery operator .. ? .. : .. elif ( ifcondition.type == "first" and ifcondition.getChildrenLength(True) == 1 and ifcondition.parent.type == "operation" and ifcondition.parent.get("operator") == "HOOK" ): variantValue = secondParam.get("value") if __variantMatchKey(variantValue, variantMap, variantGroup): repleacement = selectNode(ifcondition, "../second") else: repleacement = selectNode(ifcondition, "../third") replaceChildWithNodes(ifcondition.parent.parent, ifcondition.parent, repleacement.children) else: variantValue = secondParam.get("value") constantNode = tree.Node("constant") constantNode.set("value", str(__variantMatchKey(variantValue, variantMap, variantGroup)).lower()) constantNode.set("constantType", "boolean") constantNode.set("line", callNode.get("line")) callNode.parent.replaceChild(callNode, constantNode) #log("Warning", "Only processing qx.core.Variant.isSet directly inside of an if condition. Ignoring this occurrence.", secondParam) return True log("Warning", "The second parameter of qx.core.Variant.isSet must be a string literal. Ignoring this occurrence.", secondParam) return False | 4e5d66c6835d4497943956c0991b922d8fcff73a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5718/4e5d66c6835d4497943956c0991b922d8fcff73a/variantoptimizer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
9356,
15737,
12,
1991,
907,
16,
5437,
863,
4672,
9163,
8488,
1043,
92,
18,
3644,
18,
9356,
18,
291,
694,
1435,
4097,
31,
23819,
688,
5,
283,
17,
13284,
326,
9183,
2151,
2275,
316... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
9356,
15737,
12,
1991,
907,
16,
5437,
863,
4672,
9163,
8488,
1043,
92,
18,
3644,
18,
9356,
18,
291,
694,
1435,
4097,
31,
23819,
688,
5,
283,
17,
13284,
326,
9183,
2151,
2275,
316... |
actual = list(actual_seq) expected = list(expected_seq) missing, unexpected = unorderable_list_difference(expected, actual) | missing, unexpected = unorderable_list_difference(expected_seq, actual_seq) | def assertCountEqual(self, actual_seq, expected_seq, msg=None): """An unordered sequence specific comparison. It asserts that actual_seq and expected_seq have the same element counts. Equivalent to:: | bb802144887f1b6a5287341da2f8532e520cd8aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12029/bb802144887f1b6a5287341da2f8532e520cd8aa/case.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1815,
1380,
5812,
12,
2890,
16,
3214,
67,
5436,
16,
2665,
67,
5436,
16,
1234,
33,
7036,
4672,
3536,
979,
29573,
3102,
2923,
5826,
18,
2597,
26124,
716,
3214,
67,
5436,
471,
2665,
67,
5... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1815,
1380,
5812,
12,
2890,
16,
3214,
67,
5436,
16,
2665,
67,
5436,
16,
1234,
33,
7036,
4672,
3536,
979,
29573,
3102,
2923,
5826,
18,
2597,
26124,
716,
3214,
67,
5436,
471,
2665,
67,
5... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.