rem stringlengths 1 226k | add stringlengths 0 227k | context stringlengths 6 326k | meta stringlengths 143 403 | input_ids listlengths 256 256 | attention_mask listlengths 256 256 | labels listlengths 128 128 |
|---|---|---|---|---|---|---|
raise LayoutError("Flowable %s too large on page %d" % (f.identity(30), self.page)) | ident = "Flowable %s too large on page %d" % (f.identity(30), self.page) raise LayoutError(ident) | def handle_flowable(self,flowables): '''try to handle one flowable from the front of list flowables.''' | 019ee3bde12b7e3b2adc4e7ec74bba3eb5877041 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7053/019ee3bde12b7e3b2adc4e7ec74bba3eb5877041/doctemplate.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
67,
2426,
429,
12,
2890,
16,
2426,
1538,
4672,
9163,
698,
358,
1640,
1245,
4693,
429,
628,
326,
6641,
434,
666,
4693,
1538,
1093,
6309,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
67,
2426,
429,
12,
2890,
16,
2426,
1538,
4672,
9163,
698,
358,
1640,
1245,
4693,
429,
628,
326,
6641,
434,
666,
4693,
1538,
1093,
6309,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
try: print info['login']['result'] except: print info['error']['code'] print info['error']['info'] return False | return loginerror(info) | def login(self, username, password=False, remember=False, force=False, verify=True, domain=None): """Login to the site remember - saves cookies to a file - the filename will be: hash(username - apibase).cookies the cookies will be saved in the current directory, change cookiepath to use a different location force - forces login over the API even if a cookie file exists and overwrites an existing cookie file if remember is True verify - Checks cookie validity with isLoggedIn() domain - domain name, required for some auth systems like LDAP """ if not force: try: cookiefile = self.cookiepath + str(hash(username+' - '+self.apibase))+'.cookies' self.cookies.load(self, cookiefile, True, True) self.username = username if not verify or self.isLoggedIn(self.username): return True except: pass if not password: from getpass import getpass password = getpass() data = { "action" : "login", "lgname" : username, "lgpassword" : password, } if domain is not None: data["lgdomain"] = domain if self.maxlag < 120: data['maxlag'] = 120 req = api.APIRequest(self, data) info = req.query() if info['login']['result'] == "Success": self.username = username else: try: print info['login']['result'] except: print info['error']['code'] print info['error']['info'] return False if not self.siteinfo: self.setSiteinfo() params = { 'action': 'query', 'meta': 'userinfo', 'uiprop': 'rights', } if self.maxlag < 120: params['maxlag'] = 120 req = api.APIRequest(self, params) info = req.query() user_rights = info['query']['userinfo']['rights'] if 'apihighlimits' in user_rights: self.limit = 5000 if remember: cookiefile = self.cookiepath + str(hash(self.username+' - '+self.apibase))+'.cookies' self.cookies.save(self, cookiefile, True, True) if self.useragent == "python-wikitools/%s" % VERSION: self.useragent = "python-wikitools/%s (User:%s)" % (VERSION, self.username) return True | 6d24d3fd1c052f0c86c4d36cdeae50f08e9b85af /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10544/6d24d3fd1c052f0c86c4d36cdeae50f08e9b85af/wiki.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3925,
12,
2890,
16,
2718,
16,
2201,
33,
8381,
16,
11586,
33,
8381,
16,
2944,
33,
8381,
16,
3929,
33,
5510,
16,
2461,
33,
7036,
4672,
3536,
5358,
358,
326,
2834,
225,
11586,
300,
14649,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3925,
12,
2890,
16,
2718,
16,
2201,
33,
8381,
16,
11586,
33,
8381,
16,
2944,
33,
8381,
16,
3929,
33,
5510,
16,
2461,
33,
7036,
4672,
3536,
5358,
358,
326,
2834,
225,
11586,
300,
14649,... |
return tag('img', src=MetadataLink(context['target'], name).getURL(context)) | return tag('img', src=Link.MetadataLink(context['target'], name).getURL(context)) | def renderData_image(self, context, name, value, mime): """Return an <img> tag linking to the key's value""" return tag('img', src=MetadataLink(context['target'], name).getURL(context)) | ffd4158e0a9c3d3a315b109e5eb370b22fe8a850 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6757/ffd4158e0a9c3d3a315b109e5eb370b22fe8a850/Metadata.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1743,
751,
67,
2730,
12,
2890,
16,
819,
16,
508,
16,
460,
16,
4892,
4672,
3536,
990,
392,
411,
6081,
34,
1047,
27651,
358,
326,
498,
1807,
460,
8395,
327,
1047,
2668,
6081,
2187,
1705,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1743,
751,
67,
2730,
12,
2890,
16,
819,
16,
508,
16,
460,
16,
4892,
4672,
3536,
990,
392,
411,
6081,
34,
1047,
27651,
358,
326,
498,
1807,
460,
8395,
327,
1047,
2668,
6081,
2187,
1705,... |
user_data = 21 | user_data[0] = 21 | def test_22_create_channel(conn, channel_num, profile, received, received_data, close, close_data, user_data, next_data): info ("Called to create channel with profile: " + profile + ", and channel num: " + str (channel_num)) # check beacon if user_data != 20: error ("Expected to find create beacon equal to 20, but found: " + str (user_data)) return None # update beacon user_data = 21 return conn.open_channel (channel_num, profile) | 5e27829fa31cc85835efb1603ce2e1413515bc3b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1306/5e27829fa31cc85835efb1603ce2e1413515bc3b/vortex-regression-client.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
3787,
67,
2640,
67,
4327,
12,
4646,
16,
1904,
67,
2107,
16,
3042,
16,
5079,
16,
5079,
67,
892,
16,
1746,
16,
1746,
67,
892,
16,
729,
67,
892,
16,
1024,
67,
892,
4672,
112... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
3787,
67,
2640,
67,
4327,
12,
4646,
16,
1904,
67,
2107,
16,
3042,
16,
5079,
16,
5079,
67,
892,
16,
1746,
16,
1746,
67,
892,
16,
729,
67,
892,
16,
1024,
67,
892,
4672,
112... |
if alg=="scipy": | if algorithm=="scipy": if prec != 53: raise ValueError, "for the scipy algorithm the precision must be 53" | def hypergeometric_U(alpha,beta,x,alg="pari",prec=53): r""" Default is a wrap of Pari's hyperu(alpha,beta,x) function. Optionally, alg = "scipy" can be used. The confluent hypergeometric function $y = U(a,b,x)$ is defined to be the solution to Kummer's differential equation \[ xy'' + (b-x)y' - ay = 0. \] This satisfies $U(a,b,x) \sim x^{-a}$, as $x\rightarrow \infty$, and is sometimes denoted \verb|x^{-a}2_F_0(a,1+a-b,-1/x)|. This is not the same as Kummer's $M$-hypergeometric function, denoted sometimes as \verb|_1F_1(alpha,beta,x)|, though it satisfies the same DE that $U$ does. WARNING: In the literature, both are called "Kummer confluent hypergeometric" functions. EXAMPLES: sage: hypergeometric_U(1,1,1,"scipy") 0.596347362323... sage: hypergeometric_U(1,1,1) 0.59634736232319... sage: hypergeometric_U(1,1,1,"pari",70) 0.59634736232319407434... """ if alg=="scipy": import scipy.special ans = str(scipy.special.hyperu(float(alpha),float(beta),float(x))) ans = ans.replace("(","") ans = ans.replace(")","") ans = ans.replace("j","*I") return sage_eval(ans) ## For commented out code below, ## f = lambda x: hypergeometric_U(1,1,x) ## P = plot(f,0.1,1) ## seems to hang. I don't know why. from sage.libs.pari.all import pari from sage.interfaces.gp import Gp,gp R,a = _setup(prec) #b = R(pari(alpha).hyperu(beta,x)) #pari.set_real_precision(a) #return b #above has weird bug ans = gp.eval("hyperu(%s,%s,%s)"%(alpha,beta,x)) return R(ans) | 0090c116d1d0d5f9f5ec1cc4c8800a058defd9bb /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/0090c116d1d0d5f9f5ec1cc4c8800a058defd9bb/special.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9512,
10049,
1591,
67,
57,
12,
5429,
16,
5758,
16,
92,
16,
18413,
1546,
1065,
77,
3113,
4036,
33,
8643,
4672,
436,
8395,
2989,
353,
279,
2193,
434,
2280,
77,
1807,
9512,
89,
12,
5429,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9512,
10049,
1591,
67,
57,
12,
5429,
16,
5758,
16,
92,
16,
18413,
1546,
1065,
77,
3113,
4036,
33,
8643,
4672,
436,
8395,
2989,
353,
279,
2193,
434,
2280,
77,
1807,
9512,
89,
12,
5429,
... |
config.DisableLogging() | if not daemon.GetDebugMode(): config.DisableLogging() | def set_signal_image(): global LastStrength global stillWired #keeps us from resetting the wired info over and over global network #declared as global so it is initialized once before it gets used in the if statement below config.DisableLogging() #Check to see we wired profile autoconnect chooser needs to be displayed if daemon.GetNeedWiredProfileChooser() == True: wired_profile_chooser() daemon.SetNeedWiredProfileChooser(False) #Are we using a wired connection? wired_ip = wired.GetWiredIP() if wired.CheckPluggedIn() == True and wired_ip: if stillWired == False: pic.set_from_file("images/wired.png") tooltip.set_tip(eb,language['connected_to_wired'].replace('$A',wired_ip)) stillWired = True lock = '' else: if stillWired == True: #wire must have gotten unplugged pic.set_from_file("images/no-signal.png") tooltip.set_tip(eb,"Wicd - No Connection") stillWired = False wireless_ip = wireless.GetWirelessIP() #If ip returns as None, we are probably returning from hibernation and need to force signal to 0 to avoid crashing if wireless_ip != None: signal = int(wireless.GetCurrentSignalStrength()) else: signal = 0 #only update if the signal strength has changed because doing I/O calls is expensive, #and the icon flickers if (signal != LastStrength or network != wireless.GetCurrentNetwork()) and wireless_ip != None: LastStrength = signal lock = '' #set the string to '' so that when it is put in "high-signal" + lock + ".png", there will be nothing curNetID = wireless.GetCurrentNetworkID() #the network ID needs to be checked because a negative value here will break the tray if signal > 0 and curNetID > -1 and wireless.GetWirelessProperty(curNetID,"encryption"): lock = '-lock' #set the string to '-lock' so that it will display the lock picture network = str(wireless.GetCurrentNetwork()) tooltip.set_tip(eb,language['connected_to_wireless'].replace('$A',network).replace('$B',str(signal)).replace('$C',str(wireless_ip))) if signal > 75: pic.set_from_file("images/high-signal" + lock + ".png") elif signal > 50: pic.set_from_file("images/good-signal" + lock + ".png") elif signal > 25: pic.set_from_file("images/low-signal" + lock + ".png") elif signal > 0: pic.set_from_file("images/bad-signal" + lock + ".png") elif signal == 0: pic.set_from_file("images/no-signal.png") autoreconnect() elif wireless_ip == None and wired_ip == None: pic.set_from_file("images/no-signal") tooltip.set_tip(eb,language['not_connected']) auto_reconnect() if not daemon.GetDebugMode(): config.EnableLogging() return True | 796deb271dfb2b51e85fb4fc484236346e011621 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12280/796deb271dfb2b51e85fb4fc484236346e011621/dapper.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
67,
10420,
67,
2730,
13332,
2552,
6825,
27624,
2552,
4859,
59,
2921,
468,
10102,
87,
584,
628,
30691,
326,
341,
2921,
1123,
1879,
471,
1879,
2552,
2483,
468,
16571,
487,
2552,
1427,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
67,
10420,
67,
2730,
13332,
2552,
6825,
27624,
2552,
4859,
59,
2921,
468,
10102,
87,
584,
628,
30691,
326,
341,
2921,
1123,
1879,
471,
1879,
2552,
2483,
468,
16571,
487,
2552,
1427,
... |
query = "INSERT INTO cmtRECORDCOMMENT (id_bibrec, id_user, body, date_creation, star_score, nb_votes_total, title) " \ "VALUES (%s, %s, %s, %s, %s, %s, %s)" | query = """INSERT INTO cmtRECORDCOMMENT (id_bibrec, id_user, body, date_creation, star_score, nb_votes_total, title) VALUES (%s, %s, %s, %s, %s, %s, %s)""" | def query_add_comment_or_remark(reviews=0, recID=0, uid=-1, msg="", note="", score=0, priority=0, client_ip_address=''): """ Private function Insert a comment/review or remarkinto the database @param recID: record id @param uid: user id @param msg: comment body @param note: comment title @param score: review star score @param priority: remark priority #!FIXME @return integer >0 representing id if successful, integer 0 if not """ current_date = calculate_start_date('0d') #change utf-8 message into general unicode msg = msg.decode('utf-8') note = note.decode('utf-8') #change general unicode back to utf-8 msg = msg.encode('utf-8') note = note.encode('utf-8') query = "INSERT INTO cmtRECORDCOMMENT (id_bibrec, id_user, body, date_creation, star_score, nb_votes_total, title) " \ "VALUES (%s, %s, %s, %s, %s, %s, %s)" params = (recID, uid, msg, current_date, score, 0, note) res = run_sql(query, params) if res: action_code = cfg_webcomment_action_code[reviews and 'ADD_REVIEW' or 'ADD_COMMENT'] action_time = convert_datestruct_to_datetext(time.localtime()) query2 = """INSERT INTO cmtACTIONHISTORY values ('', %i, %i, inet_aton('%s'), '%s', '%s')""" params2 = (recID, uid, client_ip_address, action_time, action_code) run_sql(query2%params2) return int(res) | 1b2b7129fa5990ce4b8d99618b255d82d806d055 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12027/1b2b7129fa5990ce4b8d99618b255d82d806d055/webcomment.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
843,
67,
1289,
67,
3469,
67,
280,
67,
2764,
1313,
12,
266,
7061,
33,
20,
16,
1950,
734,
33,
20,
16,
4555,
29711,
21,
16,
1234,
1546,
3113,
4721,
1546,
3113,
4462,
33,
20,
16,
4394,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
843,
67,
1289,
67,
3469,
67,
280,
67,
2764,
1313,
12,
266,
7061,
33,
20,
16,
1950,
734,
33,
20,
16,
4555,
29711,
21,
16,
1234,
1546,
3113,
4721,
1546,
3113,
4462,
33,
20,
16,
4394,
... |
consumer_key='aonymous', | consumer_key='anonymous', | def RequestAccess(self): """Do all the steps involved with getting an OAuth access token. Return: True if access token was succesfully retrieved and set, otherwise False. """ import gdata.auth import subprocess # Installed applications do not have a pre-registration and so follow # directions for unregistered applications self.SetOAuthInputParameters(gdata.auth.OAuthSignatureMethod.HMAC_SHA1, consumer_key='aonymous', consumer_secret='anonymous') try: request_token = self.FetchOAuthRequestToken() except gdata.service.FetchingOAuthRequestTokenFailed, e: print e['body'].strip() + '; Request token retrieval failed!' return False auth_url = self.GenerateOAuthAuthorizationURL(request_token=request_token) try: browser = config.get('GENERAL', 'auth_browser') or os.getenv('BROWSER') except: browser = None message_format = 'Please log in and/or grant access via your browser%s' +\ ' then hit enter.' if browser: subprocess.call([browser, auth_url]) raw_input(message_format % '') else: print '(Hint: You can automatically launch your browser by adding ' +\ '"auth_browser = <browser>" to your config file under the ' +\ 'GENERAL section, or define the BROWSER environment variable.)' raw_input(message_format % (' at: ' + auth_url)) # This upgrades the token, and if successful, sets the access token try: self.UpgradeToOAuthAccessToken(request_token) except gdata.service.TokenUpgradeFailed: print 'Token upgrade failed! Could not get OAuth access token.' return False else: return True | 857d947836cb7fd3a5a68ae39edee2af0b12045d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/2899/857d947836cb7fd3a5a68ae39edee2af0b12045d/util.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1567,
1862,
12,
2890,
4672,
3536,
3244,
777,
326,
6075,
24589,
598,
8742,
392,
7270,
2006,
1147,
18,
225,
2000,
30,
1053,
309,
2006,
1147,
1703,
8667,
764,
4095,
10295,
471,
444,
16,
354... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1567,
1862,
12,
2890,
4672,
3536,
3244,
777,
326,
6075,
24589,
598,
8742,
392,
7270,
2006,
1147,
18,
225,
2000,
30,
1053,
309,
2006,
1147,
1703,
8667,
764,
4095,
10295,
471,
444,
16,
354... |
tmp_dialog.connect(tmp_dialog, QtCore.SIGNAL("tag_item"), self.tag_item_action) tmp_dialog.connect(tmp_dialog, QtCore.SIGNAL("handle_cancel()"), self.handle_cancel) tmp_dialog.connect(tmp_dialog, QtCore.SIGNAL("open_store_admin_dialog()"), self.show_admin_dialog) | def __configure_tag_dialog(self, store, tmp_dialog): """ given a store and a tag dialog - promote all settings to the dialog """ format_setting = store.get_datestamp_format() | 567eae2b78a9ee56978657c8573b5d42d92a270d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9907/567eae2b78a9ee56978657c8573b5d42d92a270d/tagstore.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
14895,
67,
2692,
67,
12730,
12,
2890,
16,
1707,
16,
1853,
67,
12730,
4672,
3536,
864,
279,
1707,
471,
279,
1047,
6176,
300,
28374,
777,
1947,
358,
326,
6176,
3536,
225,
740,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
14895,
67,
2692,
67,
12730,
12,
2890,
16,
1707,
16,
1853,
67,
12730,
4672,
3536,
864,
279,
1707,
471,
279,
1047,
6176,
300,
28374,
777,
1947,
358,
326,
6176,
3536,
225,
740,
67,
... | |
urllib2.urlopen, "http://www.python.invalid./") | urllib2.urlopen, "http://sadflkjsasf.i.nvali.d/") | def test_bad_address(self): # Make sure proper exception is raised when connecting to a bogus # address. self.assertRaises(IOError, # SF patch 809915: In Sep 2003, VeriSign started # highjacking invalid .com and .net addresses to # boost traffic to their own site. This test # started failing then. One hopes the .invalid # domain will be spared to serve its defined # purpose. # urllib2.urlopen, "http://www.sadflkjsasadf.com/") urllib2.urlopen, "http://www.python.invalid./") | 122388a5a5b02add4f09d2d578f5efc69d410f0a /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8546/122388a5a5b02add4f09d2d578f5efc69d410f0a/test_urllib2_localnet.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
8759,
67,
2867,
12,
2890,
4672,
468,
4344,
3071,
5338,
1520,
353,
11531,
1347,
14244,
358,
279,
324,
28774,
468,
1758,
18,
365,
18,
11231,
12649,
6141,
12,
4294,
668,
16,
468,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
8759,
67,
2867,
12,
2890,
4672,
468,
4344,
3071,
5338,
1520,
353,
11531,
1347,
14244,
358,
279,
324,
28774,
468,
1758,
18,
365,
18,
11231,
12649,
6141,
12,
4294,
668,
16,
468,
... |
"Cannot compiler 'Python.h'. Perhaps you need to "\ | "Cannot compile 'Python.h'. Perhaps you need to "\ | def check_types(config_cmd, ext, build_dir): private_defines = [] public_defines = [] # Expected size (in number of bytes) for each type. This is an # optimization: those are only hints, and an exhaustive search for the size # is done if the hints are wrong. expected = {} expected['short'] = [2] expected['int'] = [4] expected['long'] = [8, 4] expected['float'] = [4] expected['double'] = [8] expected['long double'] = [8, 12, 16] expected['Py_intptr_t'] = [4, 8] expected['PY_LONG_LONG'] = [8] expected['long long'] = [8] # Check we have the python header (-dev* packages on Linux) result = config_cmd.check_header('Python.h') if not result: raise SystemError( "Cannot compiler 'Python.h'. Perhaps you need to "\ "install python-dev|python-devel.") # Check basic types sizes for type in ('short', 'int', 'long'): res = config_cmd.check_decl("SIZEOF_%s" % sym2def(type), headers = ["Python.h"]) if res: public_defines.append(('NPY_SIZEOF_%s' % sym2def(type), "SIZEOF_%s" % sym2def(type))) else: res = config_cmd.check_type_size(type, expected=expected[type]) if res >= 0: public_defines.append(('NPY_SIZEOF_%s' % sym2def(type), '%d' % res)) else: raise SystemError("Checking sizeof (%s) failed !" % type) for type in ('float', 'double', 'long double'): already_declared = config_cmd.check_decl("SIZEOF_%s" % sym2def(type), headers = ["Python.h"]) res = config_cmd.check_type_size(type, expected=expected[type]) if res >= 0: public_defines.append(('NPY_SIZEOF_%s' % sym2def(type), '%d' % res)) if not already_declared and not type == 'long double': private_defines.append(('SIZEOF_%s' % sym2def(type), '%d' % res)) else: raise SystemError("Checking sizeof (%s) failed !" % type) # Compute size of corresponding complex type: used to check that our # definition is binary compatible with C99 complex type (check done at # build time in npy_common.h) complex_def = "struct {%s __x; %s __y;}" % (type, type) res = config_cmd.check_type_size(complex_def, expected=2*expected[type]) if res >= 0: public_defines.append(('NPY_SIZEOF_COMPLEX_%s' % sym2def(type), '%d' % res)) else: raise SystemError("Checking sizeof (%s) failed !" % complex_def) for type in ('Py_intptr_t',): res = config_cmd.check_type_size(type, headers=["Python.h"], library_dirs=[pythonlib_dir()], expected=expected[type]) if res >= 0: private_defines.append(('SIZEOF_%s' % sym2def(type), '%d' % res)) public_defines.append(('NPY_SIZEOF_%s' % sym2def(type), '%d' % res)) else: raise SystemError("Checking sizeof (%s) failed !" % type) # We check declaration AND type because that's how distutils does it. if config_cmd.check_decl('PY_LONG_LONG', headers=['Python.h']): res = config_cmd.check_type_size('PY_LONG_LONG', headers=['Python.h'], library_dirs=[pythonlib_dir()], expected=expected['PY_LONG_LONG']) if res >= 0: private_defines.append(('SIZEOF_%s' % sym2def('PY_LONG_LONG'), '%d' % res)) public_defines.append(('NPY_SIZEOF_%s' % sym2def('PY_LONG_LONG'), '%d' % res)) else: raise SystemError("Checking sizeof (%s) failed !" % 'PY_LONG_LONG') res = config_cmd.check_type_size('long long', expected=expected['long long']) if res >= 0: #private_defines.append(('SIZEOF_%s' % sym2def('long long'), '%d' % res)) public_defines.append(('NPY_SIZEOF_%s' % sym2def('long long'), '%d' % res)) else: raise SystemError("Checking sizeof (%s) failed !" % 'long long') if not config_cmd.check_decl('CHAR_BIT', headers=['Python.h']): raise RuntimeError( "Config wo CHAR_BIT is not supported"\ ", please contact the maintainers") return private_defines, public_defines | 5ab03fb56e40dd5ec20535297d1c2602e8ad17a1 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14925/5ab03fb56e40dd5ec20535297d1c2602e8ad17a1/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
2352,
12,
1425,
67,
4172,
16,
1110,
16,
1361,
67,
1214,
4672,
3238,
67,
536,
1465,
273,
5378,
1071,
67,
536,
1465,
273,
5378,
225,
468,
13219,
963,
261,
267,
1300,
434,
1731,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
2352,
12,
1425,
67,
4172,
16,
1110,
16,
1361,
67,
1214,
4672,
3238,
67,
536,
1465,
273,
5378,
1071,
67,
536,
1465,
273,
5378,
225,
468,
13219,
963,
261,
267,
1300,
434,
1731,
... |
testsrc=filter(is_chpl_source, dirlist) | testsrc=filter(IsChplSource, dirlist) | def kill_proc(p, timeout): k = subprocess.Popen(['kill',str(p.pid)]) k.wait() now = time.time() end_time = now + timeout # give it a little time while end_time > now: if p.poll(): return now = time.time() # use the big hammer (and don't bother waiting) subprocess.Popen(['kill','-9', str(p.pid)]) return | 6726fd039f7a0494d6243f612d9ade4ec88db45d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6895/6726fd039f7a0494d6243f612d9ade4ec88db45d/sub_test.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8673,
67,
9381,
12,
84,
16,
2021,
4672,
417,
273,
6652,
18,
52,
3190,
12,
3292,
16418,
2187,
701,
12,
84,
18,
6610,
13,
5717,
417,
18,
7048,
1435,
2037,
273,
813,
18,
957,
1435,
679,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8673,
67,
9381,
12,
84,
16,
2021,
4672,
417,
273,
6652,
18,
52,
3190,
12,
3292,
16418,
2187,
701,
12,
84,
18,
6610,
13,
5717,
417,
18,
7048,
1435,
2037,
273,
813,
18,
957,
1435,
679,... |
(isinstance(tree.childNodes[i+1], _Text) or | (isinstance(tree.childNodes[i+1], xml.dom.minidom.Text) or | def _dom_to_html_helper(self, tree, container, indent, seclevel): """ Helper function for L{_dom_to_html}, that does the real work of converting a DOM tree for an epytext string to HTML. @param tree: The DOM tree for an epytext string. @type tree: C{xml.dom.minidom.Element} @param container: The container in which to look up objects by name (e.g., for link directives). @type container: Python module @param indent: The number of characters indentation that should be used for HTML tags. @type indent: C{int} @param seclevel: The current section level. @type seclevel: int """ # This takes care of converting > to >, etc.: if isinstance(tree, _Text): return tree.toxml() if tree.tagName == 'epytext': indent -= 2 if tree.tagName == 'section': seclevel += 1 # Process the children first. children = [self._dom_to_html_helper(c, container, indent+2, seclevel) for c in tree.childNodes] # Get rid of unnecessary <P>...</P> tags; they introduce extra # space on most browsers that we don't want. for i in range(len(children)-1): if (not isinstance(tree.childNodes[i], _Text) and tree.childNodes[i].tagName == 'para' and (isinstance(tree.childNodes[i+1], _Text) or tree.childNodes[i+1].tagName != 'para')): children[i] = ' '*(indent+2)+children[i][5+indent:-5]+'\n' if (tree.hasChildNodes() and not isinstance(tree.childNodes[-1], _Text) and tree.childNodes[-1].tagName == 'para'): children[-1] = ' '*(indent+2)+children[-1][5+indent:-5]+'\n' # Construct the HTML string for the children. childstr = ''.join(children) # Perform the approriate action for the DOM tree type. if tree.tagName == 'para': return epytext.wordwrap('<p>%s</p>' % childstr, indent) elif tree.tagName == 'code': return '<code>%s</code>' % childstr elif tree.tagName == 'uri': return '<a href="%s">%s</a>' % (children[1], children[0]) elif tree.tagName == 'link': return self._dom_link_to_html(children[1], children[0], container) elif tree.tagName == 'italic': return '<i>%s</i>' % childstr elif tree.tagName == 'math': return '<i class="math">%s</i>' % childstr elif tree.tagName == 'index': return ('<a name="%s"></a>' % self._index_to_anchor(childstr) + '<i class="indexterm">%s</i>' % childstr) elif tree.tagName == 'bold': return '<b>%s</b>' % childstr elif tree.tagName == 'ulist': return '%s<ul>\n%s%s</ul>\n' % (indent*' ', childstr, indent*' ') elif tree.tagName == 'olist': startAttr = tree.getAttributeNode('start') if startAttr: start = ' start="%s"' % startAttr.value else: start = '' return ('%s<ol%s>\n%s%s</ol>\n' % (indent*' ', start, childstr, indent*' ')) elif tree.tagName == 'li': return indent*' '+'<li>\n%s%s</li>\n' % (childstr, indent*' ') elif tree.tagName == 'heading': return ('%s<h%s class="heading">%s</h%s>\n' % ((indent-2)*' ', seclevel, childstr, seclevel)) elif tree.tagName == 'literalblock': return '<pre class="literalblock">\n%s\n</pre>\n' % childstr elif tree.tagName == 'doctestblock': dtb = self._colorize_doctestblock(childstr.strip()) return '<pre class="doctestblock">\n%s</pre>\n' % dtb elif tree.tagName == 'fieldlist': raise AssertionError("There should not be any field lists left") elif tree.tagName in ('epytext', 'section', 'tag', 'arg', 'name', 'target'): return childstr else: raise ValueError('Warning: unknown epytext DOM element %r' % tree.tagName) | 36aef14a0edb4bd18076686105d0e4e6c7c10c23 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11420/36aef14a0edb4bd18076686105d0e4e6c7c10c23/html.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
9859,
67,
869,
67,
2620,
67,
4759,
12,
2890,
16,
2151,
16,
1478,
16,
3504,
16,
1428,
2815,
4672,
3536,
9705,
445,
364,
511,
95,
67,
9859,
67,
869,
67,
2620,
5779,
716,
1552,
326... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
9859,
67,
869,
67,
2620,
67,
4759,
12,
2890,
16,
2151,
16,
1478,
16,
3504,
16,
1428,
2815,
4672,
3536,
9705,
445,
364,
511,
95,
67,
9859,
67,
869,
67,
2620,
5779,
716,
1552,
326... |
replace_lines(os.path.basename(sys.argv[1])) | if dirname != '': os.chdir(dirname) replace_lines(basename) | def replace_lines(filename): try: file = open(filename) for line in file.readlines(): matched = re.compile("^%include\s+(.*)").match(line) if matched: replace_lines(matched.group(1)) else: sys.stdout.write(line) file.close() except IOError, detail: print detail sys.exit(2) | 848eaf483df9eb79342ae841a80f7724c85899f0 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1639/848eaf483df9eb79342ae841a80f7724c85899f0/ks_fold_include.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1453,
67,
3548,
12,
3459,
4672,
775,
30,
585,
273,
1696,
12,
3459,
13,
364,
980,
316,
585,
18,
896,
3548,
13332,
4847,
273,
283,
18,
11100,
2932,
66,
9,
6702,
64,
87,
15,
7444,
2225,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1453,
67,
3548,
12,
3459,
4672,
775,
30,
585,
273,
1696,
12,
3459,
13,
364,
980,
316,
585,
18,
896,
3548,
13332,
4847,
273,
283,
18,
11100,
2932,
66,
9,
6702,
64,
87,
15,
7444,
2225,... |
session.close() conn.close() | if __name__ == '__main__': unittest.main() | def onException(self, exc): print exc, ':', exc.message | 16f5191b239ef3879b9382d7f75c107fe1c814b3 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/4106/16f5191b239ef3879b9382d7f75c107fe1c814b3/test.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
503,
12,
2890,
16,
3533,
4672,
1172,
3533,
16,
4290,
2187,
3533,
18,
2150,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
503,
12,
2890,
16,
3533,
4672,
1172,
3533,
16,
4290,
2187,
3533,
18,
2150,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
if options['escape']: | if options['escape'] and not options['notty']: self.escapeMode = 1 | def channelOpen(self, foo): #global globalSession #globalSession = self # turn off local echo if options['agent']: d = self.conn.sendRequest(self, 'auth-agent-req@openssh.com', '', wantReply=1) d.addBoth(lambda x:log.msg(x)) if options['noshell']: return self.escapeMode = 1 fd = 0 #sys.stdin.fileno() try: new = tty.tcgetattr(fd) except: log.msg('not a typewriter!') else: new[3] = new[3] & ~tty.ICANON & ~tty.ECHO new[6][tty.VMIN] = 1 new[6][tty.VTIME] = 0 tty.tcsetattr(fd, tty.TCSANOW, new) tty.setraw(fd) c = session.SSHSessionClient() if options['escape']: c.dataReceived = self.handleInput else: c.dataReceived = self.write c.connectionLost = self.sendEOF self.stdio = stdio.StandardIO(c) if options['subsystem']: self.conn.sendRequest(self, 'subsystem', \ common.NS(options['command'])) elif options['command']: if options['tty']: term = os.environ['TERM'] winsz = fcntl.ioctl(fd, tty.TIOCGWINSZ, '12345678') winSize = struct.unpack('4H', winsz) ptyReqData = session.packRequest_pty_req(term, winSize, '') self.conn.sendRequest(self, 'pty-req', ptyReqData) signal.signal(signal.SIGWINCH, self._windowResized) self.conn.sendRequest(self, 'exec', \ common.NS(options['command'])) else: if not options['notty']: term = os.environ['TERM'] winsz = fcntl.ioctl(fd, tty.TIOCGWINSZ, '12345678') winSize = struct.unpack('4H', winsz) ptyReqData = session.packRequest_pty_req(term, winSize, '') self.conn.sendRequest(self, 'pty-req', ptyReqData) signal.signal(signal.SIGWINCH, self._windowResized) self.conn.sendRequest(self, 'shell', '') if hasattr(self.conn.transport, 'transport'): self.conn.transport.transport.setTcpNoDelay(1) | 0b278fb0f6b45a71f7e8a6ee05680d7b5b1f6d7b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12595/0b278fb0f6b45a71f7e8a6ee05680d7b5b1f6d7b/conch.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1904,
3678,
12,
2890,
16,
8431,
4672,
468,
6347,
2552,
2157,
468,
6347,
2157,
273,
365,
468,
7005,
3397,
1191,
3376,
309,
702,
3292,
5629,
3546,
30,
302,
273,
365,
18,
4646,
18,
4661,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1904,
3678,
12,
2890,
16,
8431,
4672,
468,
6347,
2552,
2157,
468,
6347,
2157,
273,
365,
468,
7005,
3397,
1191,
3376,
309,
702,
3292,
5629,
3546,
30,
302,
273,
365,
18,
4646,
18,
4661,
... |
rho = matrixmultiply(evec_L, crho) | rho = matrixmultiply(RLt, crho) | def calc_TLS_center_of_reaction(T_orig, L_orig, S_orig, origin): """Calculate new tensors based on the center for reaction. This method returns a dictionary of the calculations: T^: T tensor in the coordinate system of L L^: L tensor in the coordinate system of L S^: S tensor in the coordinate system of L COR: Center of Reaction T',S',L': T,L,S tensors in origonal coordinate system with the origin shifted to the center of reaction. """ ## LSMALL is the smallest magnitude of L before it is considered 0.0 LSMALL = 1e-6 rdict = {} ## set the L tensor eigenvalues and eigenvectors (eval_L, evec_L) = eigenvectors(L_orig) rdict["L1_eigen_val"] = eval_L[0] rdict["L2_eigen_val"] = eval_L[1] rdict["L3_eigen_val"] = eval_L[2] rdict["L1_eigen_vec"] = evec_L[0] rdict["L2_eigen_vec"] = evec_L[1] rdict["L3_eigen_vec"] = evec_L[2] ## transpose the original the evec_L so it can be used ## to rotate the other tensors evec_L = transpose(evec_L) ## carrot-L tensor (tensor WRT principal axes of L) cL = zeros([3,3], Float) cL[0,0] = eval_L[0] cL[1,1] = eval_L[1] cL[2,2] = eval_L[2] rdict["L^"] = cL ## carrot-T tensor (T tensor WRT principal axes of L) cT = matrixmultiply( matrixmultiply(transpose(evec_L), T_orig), evec_L) rdict["T^"] = cT ## carrot-S tensor (S tensor WRT principal axes of L) cS = matrixmultiply( matrixmultiply(transpose(evec_L), S_orig), evec_L) ## correct for left-handed libration eigenvectors det = determinant(evec_L) if int(det) != 1: cS = -cS rdict["S^"] = cS ## ^rho: the origin-shift vector in the coordinate system of L cL1122 = cL[1,1] + cL[2,2] cL2200 = cL[2,2] + cL[0,0] cL0011 = cL[0,0] + cL[1,1] if cL1122>LSMALL: crho0 = (cS[1,2]-cS[2,1]) / cL1122 else: crho0 = 0.0 if cL2200>LSMALL: crho1 = (cS[2,0]-cS[0,2]) / cL2200 else: crho1 = 0.0 if cL0011>LSMALL: crho2 = (cS[0,1]-cS[1,0]) / cL0011 else: crho2 = 0.0 crho = array([crho0, crho1, crho2], Float) rdict["RHO^"] = crho ## rho: the origin-shift vector in orthogonal coordinates rho = matrixmultiply(evec_L, crho) rdict["RHO"] = rho rdict["COR"] = origin + rho.copy() ## set up the origin shift matrix PRHO WRT orthogonal axes PRHO = array([ [ 0.0, rho[2], -rho[1]], [-rho[2], 0.0, rho[0]], [ rho[1], -rho[0], 0.0] ], Float) ## set up the origin shift matrix cPRHO WRT libration axes cPRHO = array([ [ 0.0, crho[2], -crho[1]], [-crho[2], 0.0, crho[0]], [ crho[1], -crho[0], 0.0] ], Float) ## calculate tranpose of cPRHO, ans cS cSt = transpose(cS) cPRHOt = transpose(cPRHO) ## calculate S'^ = S^ + L^*pRHOt cSp = cS + matrixmultiply(cL, cPRHOt) rdict["S'^"] = cSp ## L'^ = L^ = cL rdict["L'^"] = cL ## calculate T'^ = cT + cPRHO*S^ + cSt*cPRHOt + cPRHO*cL*cPRHOt * cTp = cT + \ matrixmultiply(cPRHO, cS) + \ matrixmultiply(cSt, cPRHOt) + \ matrixmultiply(matrixmultiply(cPRHO, cL), cPRHOt) rdict["T'^"] = cTp ## transpose of PRHO and S PRHOt = transpose(PRHO) St = transpose(S_orig) ## calculate S' = S + L*PRHOt Sp = S_orig + matrixmultiply(L_orig, PRHOt) rdict["S'"] = Sp ## calculate T' = T + PRHO*S + St*PRHOT + PRHO*L*PRHOt Tp = T_orig + \ matrixmultiply(PRHO, S_orig) + \ matrixmultiply(St, PRHOt) + \ matrixmultiply(matrixmultiply(PRHO, L_orig), PRHOt) rdict["T'"] = Tp ## L' is just L rdict["L'"] = L_orig.copy() ## now calculate the TLS motion description using 3 non ## intersecting screw axes, with one ## libration axis 1 shift in the L coordinate system if cL[0,0]>LSMALL: cL1rho = array([0.0, -cSp[0,2]/cL[0,0], cSp[0,1]/cL[0,0]], Float) else: cL1rho = zeros(3, Float) ## libration axis 2 shift in the L coordinate system if cL[1,1]>LSMALL: cL2rho = array([cSp[1,2]/cL[1,1], 0.0, -cSp[1,0]/cL[1,1]], Float) else: cL2rho = zeros(3, Float) ## libration axis 2 shift in the L coordinate system if cL[2,2]>LSMALL: cL3rho = array([-cSp[2,1]/cL[2,2], cSp[2,0]/cL[2,2], 0.0], Float) else: cL3rho = zeros(3, Float) ## libration axes shifts in the origional orthogonal ## coordinate system rdict["L1_rho"] = matrixmultiply(evec_L, cL1rho) rdict["L2_rho"] = matrixmultiply(evec_L, cL2rho) rdict["L3_rho"] = matrixmultiply(evec_L, cL3rho) ## calculate screw pitches (A*R / R*R) = (A/R) if cL[0,0]>LSMALL: rdict["L1_pitch"] = cS[0,0]/cL[0,0] else: rdict["L1_pitch"] = 0.0 if cL[1,1]>LSMALL: rdict["L2_pitch"] = cS[1,1]/cL[1,1] else: rdict["L2_pitch"] = 0.0 if cL[2,2]>LSMALL: rdict["L3_pitch"] = cS[2,2]/cL[2,2] else: rdict["L3_pitch"] = 0.0 ## now calculate the reduction in T for the screw rotation axes cTred = cT.copy() for i in (0, 1, 2): for k in (0, 1, 2): if i==k: continue if cL[k,k]>LSMALL: cTred[i,i] -= (cS[k,i]**2) / cL[k,k] for i in (0, 1, 2): for j in (0, 1, 2): for k in (0, 1, 2): if j==i: continue if cL[k,k]>LSMALL: cTred[i,j] -= (cS[k,i]*cS[k,j]) / cL[k,k] ## rotate the newly calculated reduced-T tensor from the carrot ## coordinate system (coordinate system of L) back to the structure ## coordinate system rdict["rT'"] = matrixmultiply( transpose(evec_L), matrixmultiply(cTred, evec_L)) return rdict | 742140185bc8b6b74df991525abcac33bdc82b5f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10674/742140185bc8b6b74df991525abcac33bdc82b5f/TLS.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7029,
67,
9905,
67,
5693,
67,
792,
67,
266,
1128,
12,
56,
67,
4949,
16,
511,
67,
4949,
16,
348,
67,
4949,
16,
4026,
4672,
3536,
8695,
394,
25122,
2511,
603,
326,
4617,
364,
12836,
18... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7029,
67,
9905,
67,
5693,
67,
792,
67,
266,
1128,
12,
56,
67,
4949,
16,
511,
67,
4949,
16,
348,
67,
4949,
16,
4026,
4672,
3536,
8695,
394,
25122,
2511,
603,
326,
4617,
364,
12836,
18... |
time_first_flower_open = 'DNF' | time_tuple = 'DNF' | def get_experiment_id2accession_id2reading_ls(self, curs, raw_phenotype_table): """ 2008-02-29 replicate gets converted to arabic and included in the returning results 2008-02-23 'long ago' now mapped to 'NA' the reading_error_dict applied before testing whether it's NA or not 2008-02-20 """ sys.stderr.write("Getting experiment_id2accession_id2reading_ls ...\n") curs.execute("select experiment, accession, replicate, reading from %s where measure='first flower open'"%raw_phenotype_table) rows = curs.fetchall() experiment_id2accession_id2reading_ls = {} reading_error_dict = {'1/14/2003+C41':'1/14/03', '110/20/03':'11/20/03', '1/12003':'1/12/03', #could also be 1/1/03 but after discussing with glenda and look around the table, 1/12/03 is more likely. '10//24/03':'10/24/03', 'long ago':'NA'} from pymodule.roman import fromRoman for row in rows: experiment_id, accession_id, replicate, reading = row if reading in reading_error_dict: #correct the typo reading = reading_error_dict[reading] if reading!='NA' and reading!='N': if experiment_id not in experiment_id2accession_id2reading_ls: experiment_id2accession_id2reading_ls[experiment_id] = {} if accession_id not in experiment_id2accession_id2reading_ls[experiment_id]: experiment_id2accession_id2reading_ls[experiment_id][accession_id] = [] try: if experiment_id==3 and reading[-2:]=='99': #wrong year input reading = reading[:-2]+'03' if experiment_id==4 and reading[-2:]=='00': #wrong year input reading = reading[:-2]+'02' if reading!='DNF': time_first_flower_open = time.strptime(reading, '%m/%d/%y') #transform it into a time tuple else: time_first_flower_open = 'DNF' except: print "reading:",reading traceback.print_exc() print sys.exc_info() sys.exit(2) experiment_id2accession_id2reading_ls[experiment_id][accession_id].append((time_first_flower_open, fromRoman(replicate))) sys.stderr.write("Done.\n") return experiment_id2accession_id2reading_ls | e6897ae45631832ceb92bfc9db5ab27edc82a829 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9645/e6897ae45631832ceb92bfc9db5ab27edc82a829/ProcessPhenotype.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
18142,
67,
350,
22,
3860,
285,
67,
350,
22,
21803,
67,
3251,
12,
2890,
16,
25326,
16,
1831,
67,
13819,
10570,
67,
2121,
4672,
3536,
4044,
28,
17,
3103,
17,
5540,
28476,
5571,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
18142,
67,
350,
22,
3860,
285,
67,
350,
22,
21803,
67,
3251,
12,
2890,
16,
25326,
16,
1831,
67,
13819,
10570,
67,
2121,
4672,
3536,
4044,
28,
17,
3103,
17,
5540,
28476,
5571,
... |
game.state.galaxy[w.x][w.y].romulans += 1 | game.state.galaxy[w.i][w.j].romulans += 1 | def setup(): "Prepare to play, set up cosmos." w = coord() # Decide how many of everything if choose(): return # frozen game # Prepare the Enterprise game.alldone = game.gamewon = False game.ship = IHE game.state.crew = FULLCREW game.energy = game.inenrg = 5000.0 game.shield = game.inshld = 2500.0 game.shldchg = False game.shldup = False game.inlsr = 4.0 game.lsupres = 4.0 game.quadrant = randplace(GALSIZE) game.sector = randplace(QUADSIZE) game.torps = game.intorps = 10 game.nprobes = randrange(2, 5) game.warpfac = 5.0 game.wfacsq = game.warpfac * game.warpfac for i in range(NDEVICES): game.damage[i] = 0.0 # Set up assorted game parameters game.battle = coord() game.state.date = game.indate = 100.0 * randreal(20, 51) game.nkinks = game.nhelp = game.casual = game.abandoned = 0 game.iscate = game.resting = game.imine = game.icrystl = game.icraft = False game.isatb = game.state.nplankl = 0 game.state.starkl = game.state.basekl = 0 game.iscraft = "onship" game.landed = False game.alive = True game.docfac = 0.25 # Starchart is functional but we've never seen it game.lastchart = FOREVER # Put stars in the galaxy game.instar = 0 for i in range(GALSIZE): for j in range(GALSIZE): k = randrange(1, QUADSIZE**2/10+1) game.instar += k game.state.galaxy[i][j].stars = k # Locate star bases in galaxy for i in range(game.inbase): while True: while True: w = randplace(GALSIZE) if not game.state.galaxy[w.x][w.y].starbase: break contflag = False # C version: for (j = i-1; j > 0; j--) # so it did them in the opposite order. for j in range(1, i): # Improved placement algorithm to spread out bases distq = (w - game.state.baseq[j]).distance() if distq < 6.0*(BASEMAX+1-game.inbase) and withprob(0.75): contflag = True if idebug: prout("=== Abandoning base #%d at %s" % (i, w)) break elif distq < 6.0 * (BASEMAX+1-game.inbase): if idebug: prout("=== Saving base #%d, close to #%d" % (i, j)) if not contflag: break game.state.baseq.append(w) game.state.galaxy[w.x][w.y].starbase = game.state.chart[w.x][w.y].starbase = True # Position ordinary Klingon Battle Cruisers krem = game.inkling klumper = 0.25*game.skill*(9.0-game.length)+1.0 if klumper > MAXKLQUAD: klumper = MAXKLQUAD while True: r = randreal() klump = (1.0 - r*r)*klumper if klump > krem: klump = krem krem -= klump while True: w = randplace(GALSIZE) if not game.state.galaxy[w.x][w.y].supernova and \ game.state.galaxy[w.x][w.y].klingons + klump <= MAXKLQUAD: break game.state.galaxy[w.x][w.y].klingons += int(klump) if krem <= 0: break # Position Klingon Commander Ships for i in range(game.incom): while True: w = randplace(GALSIZE) if not welcoming(w) or w in game.state.kcmdr: continue if (game.state.galaxy[w.x][w.y].klingons or withprob(0.25)): break game.state.galaxy[w.x][w.y].klingons += 1 game.state.kcmdr.append(w) # Locate planets in galaxy for i in range(game.inplan): while True: w = randplace(GALSIZE) if game.state.galaxy[w.x][w.y].planet == None: break new = planet() new.quadrant = w new.crystals = "absent" if (game.options & OPTION_WORLDS) and i < NINHAB: new.pclass = "M" # All inhabited planets are class M new.crystals = "absent" new.known = "known" new.name = systnames[i] new.inhabited = True else: new.pclass = ("M", "N", "O")[randrange(0, 3)] if withprob(0.33): new.crystals = "present" new.known = "unknown" new.inhabited = False game.state.galaxy[w.x][w.y].planet = new game.state.planets.append(new) # Locate Romulans for i in range(game.state.nromrem): w = randplace(GALSIZE) game.state.galaxy[w.x][w.y].romulans += 1 # Place the Super-Commander if needed if game.state.nscrem > 0: while True: w = randplace(GALSIZE) if welcoming(w): break game.state.kscmdr = w game.state.galaxy[w.x][w.y].klingons += 1 # Initialize times for extraneous events schedule(FSNOVA, expran(0.5 * game.intime)) schedule(FTBEAM, expran(1.5 * (game.intime / len(game.state.kcmdr)))) schedule(FSNAP, randreal(1.0, 2.0)) # Force an early snapshot schedule(FBATTAK, expran(0.3*game.intime)) unschedule(FCDBAS) if game.state.nscrem: schedule(FSCMOVE, 0.2777) else: unschedule(FSCMOVE) unschedule(FSCDBAS) unschedule(FDSPROB) if (game.options & OPTION_WORLDS) and game.skill >= SKILL_GOOD: schedule(FDISTR, expran(1.0 + game.intime)) else: unschedule(FDISTR) unschedule(FENSLV) unschedule(FREPRO) # Place thing (in tournament game, we don't want one!) # New in SST2K: never place the Thing near a starbase. # This makes sense and avoids a special case in the old code. global thing if game.tourn is None: while True: thing = randplace(GALSIZE) if thing not in game.state.baseq: break skip(2) game.state.snap = False if game.skill == SKILL_NOVICE: prout(_("It is stardate %d. The Federation is being attacked by") % int(game.state.date)) prout(_("a deadly Klingon invasion force. As captain of the United")) prout(_("Starship U.S.S. Enterprise, it is your mission to seek out")) prout(_("and destroy this invasion force of %d battle cruisers.") % ((game.inkling + game.incom + game.inscom))) prout(_("You have an initial allotment of %d stardates to complete") % int(game.intime)) prout(_("your mission. As you proceed you may be given more time.")) skip(1) prout(_("You will have %d supporting starbases.") % (game.inbase)) proutn(_("Starbase locations- ")) else: prout(_("Stardate %d.") % int(game.state.date)) skip(1) prout(_("%d Klingons.") % (game.inkling + game.incom + game.inscom)) prout(_("An unknown number of Romulans.")) if game.state.nscrem: prout(_("And one (GULP) Super-Commander.")) prout(_("%d stardates.") % int(game.intime)) proutn(_("%d starbases in ") % game.inbase) for i in range(game.inbase): proutn(`game.state.baseq[i]`) proutn(" ") skip(2) proutn(_("The Enterprise is currently in Quadrant %s") % game.quadrant) proutn(_(" Sector %s") % game.sector) skip(2) prout(_("Good Luck!")) if game.state.nscrem: prout(_(" YOU'LL NEED IT.")) waitfor() newqad() if len(game.enemies) - (thing == game.quadrant) - (game.tholian != None): game.shldup = True if game.neutz: # bad luck to start in a Romulan Neutral Zone attack(torps_ok=False) | e67cf36a789c4ab1cd905a9a08e91d219395f538 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3176/e67cf36a789c4ab1cd905a9a08e91d219395f538/sst.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
13332,
315,
7543,
358,
6599,
16,
444,
731,
4987,
26719,
1199,
341,
273,
2745,
1435,
468,
225,
3416,
831,
3661,
4906,
434,
7756,
309,
9876,
13332,
327,
468,
12810,
7920,
468,
7730,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
13332,
315,
7543,
358,
6599,
16,
444,
731,
4987,
26719,
1199,
341,
273,
2745,
1435,
468,
225,
3416,
831,
3661,
4906,
434,
7756,
309,
9876,
13332,
327,
468,
12810,
7920,
468,
7730,
... |
if firstStatement.type == "throw": node.set("isAbstract", True) | def handleFunction(funcItem, name, commentAttributes, classNode): if funcItem.type != "function": raise DocException("'funcItem' is no function", funcItem) node = tree.Node("method") node.set("name", name) # Read the parameters params = funcItem.getChild("params", False) if params and params.hasChildren(): for param in params.children: if param.type != "variable": continue paramNode = tree.Node("param") paramNode.set("name", param.getFirstChild().get("name")) node.addListChild("params", paramNode) # Check whether the function is abstract bodyBlockItem = funcItem.getChild("body").getFirstChild(); if bodyBlockItem.type == "block" and bodyBlockItem.hasChildren(): firstStatement = bodyBlockItem.children[0]; if firstStatement.type == "throw": # The first statement of the function is a throw statement # -> The function is abstract node.set("isAbstract", True) handleAccess(node, commentAttributes) if len(commentAttributes) == 0: addError(node, "Documentation is missing.", funcItem) return node handleDeprecated(node, commentAttributes) # Read all description, param and return attributes for attrib in commentAttributes: # Add description if attrib["category"] == "description": if attrib.has_key("text"): if "TODOC" in attrib["text"]: addError(node, "Documentation is missing.", funcItem) descNode = tree.Node("desc").set("text", attrib["text"]) node.addChild(descNode) elif attrib["category"] == "see": if not attrib.has_key("name"): raise DocException("Missing target for see.", funcItem) seeNode = tree.Node("see").set("name", attrib["name"]) node.addChild(seeNode) elif attrib["category"] == "param": if not attrib.has_key("name"): raise DocException("Missing name of parameter.", funcItem) # Find the matching param node paramName = attrib["name"] paramNode = node.getListChildByAttribute("params", "name", paramName, False) if not paramNode: addError(node, "Contains information for a non-existing parameter <code>%s</code>." % paramName, funcItem) continue addTypeInfo(paramNode, attrib, funcItem) elif attrib["category"] == "return": returnNode = tree.Node("return") node.addChild(returnNode) addTypeInfo(returnNode, attrib, funcItem) # Check for documentation errors # Check whether all parameters have been documented if node.hasChild("params"): paramsListNode = node.getChild("params"); for paramNode in paramsListNode.children: if not paramNode.getChild("desc", False): addError(node, "Parameter <code>%s</code> is not documented." % paramNode.get("name"), funcItem) if not node.hasChild("desc"): addError(node, "Documentation is missing.", funcItem) return node | 553eb07a7af2000fdf74c165c23d21c2b9e376e2 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/5718/553eb07a7af2000fdf74c165c23d21c2b9e376e2/api.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
2083,
12,
644,
1180,
16,
508,
16,
2879,
2498,
16,
25822,
4672,
309,
1326,
1180,
18,
723,
480,
315,
915,
6877,
1002,
3521,
503,
29823,
644,
1180,
11,
353,
1158,
445,
3113,
1326,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
2083,
12,
644,
1180,
16,
508,
16,
2879,
2498,
16,
25822,
4672,
309,
1326,
1180,
18,
723,
480,
315,
915,
6877,
1002,
3521,
503,
29823,
644,
1180,
11,
353,
1158,
445,
3113,
1326,
1... | |
col_width = [-1] + col_width | col_width = [None] + col_width | def datefmt_func(date): return date.strfmt(datefmt) | 49f7cf850b803093831ddb07ea689be861903eaa /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12971/49f7cf850b803093831ddb07ea689be861903eaa/reportlib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1509,
8666,
67,
644,
12,
712,
4672,
327,
1509,
18,
701,
8666,
12,
712,
8666,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1509,
8666,
67,
644,
12,
712,
4672,
327,
1509,
18,
701,
8666,
12,
712,
8666,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
elif self.distro_type == 'debian-based': | elif self.distro_type == 'debian-based': print 'Remember that you must have the following lsb compliance packages installed:' print 'lsb-core lsb-cxx lsb-graphics lsb-desktop lsb-qt4 lsb-languages lsb-multimedia lsb-printing' print 'Converting and installing LSB packages' self.dtk_manager_dpkg = package.convert(self.dtk_manager_pkg, 'dpkg') package.install(self.dtk_manager_dpkg) for self.lsb_rpm in self.lsb_pkg_list: | def install_lsb_packages(self, srcdir, cachedir, my_config): # Installing LSB DTK manager packages self.dtk_manager_arch = my_config.get('dtk-manager', 'arch-%s' % self.get_lsb_arch()) self.dtk_manager_url = my_config.get('dtk-manager', 'tarball_url') % self.dtk_manager_arch if not self.dtk_manager_url: raise TestError('Could not get DTK manager URL from configuration file') self.dtk_md5 = my_config.get('dtk-manager', 'md5-%s' % self.get_lsb_arch()) if self.dtk_md5: print 'Caching LSB DTK manager RPM' self.dtk_manager_pkg = unmap_url_cache(cachedir, self.dtk_manager_url, self.dtk_md5) else: raise TestError('Could not find DTK manager package md5, cannot cache DTK manager tarball') print 'Installing LSB DTK manager RPM' package.install(self.dtk_manager_pkg) | 74523bd453d8f2601557c66254590e04dd035f68 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/10349/74523bd453d8f2601557c66254590e04dd035f68/lsb_dtk.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3799,
67,
3251,
70,
67,
10308,
12,
2890,
16,
1705,
1214,
16,
3472,
481,
16,
3399,
67,
1425,
4672,
468,
10284,
310,
511,
14541,
10696,
47,
3301,
5907,
365,
18,
7510,
79,
67,
4181,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3799,
67,
3251,
70,
67,
10308,
12,
2890,
16,
1705,
1214,
16,
3472,
481,
16,
3399,
67,
1425,
4672,
468,
10284,
310,
511,
14541,
10696,
47,
3301,
5907,
365,
18,
7510,
79,
67,
4181,
67,
... |
numbers larger than 30. ( or even less) turn to 1. and numbers | numbers larger than 30. (or even less then that) turn to 1. and numbers | def get_reconstruction_cost(self, updates, pre_sigmoid_nv): """Approximation to the reconstruction error Note that this function requires the pre-sigmoid activation. To understand why this is so you need to understand a bit about how Theano works. Once you express a computational graph in Theano, it will apply to it several optimizations which will lead to a faster and more stable computational graph. One of these optimizations expresses terms of the form log(sigmoid(x)) in terms of softplus. We need this optimization for the cross-entropy since sigmoid of numbers larger than 30. ( or even less) turn to 1. and numbers smaller than -30. turn to 0 which in terms will force theano to compute log(0) and therefore we will get either -inf or NaN as cost. If the value is expressed in terms of softplus we do not get this undesirable behaviour. This optimization usually works fine, but here we have a special case. The sigmoid is applied inside the scan op, while the log is outside. Therefore Theano will only see log(scan(..)) instead of log(sigmoid(..)) and will not apply the wanted optimization. We can not go and replace the sigmoid in scan with something else also, because this only needs to be done on the last step. Therefore the easiest and more efficient way is to get also the pre-sigmoid activation as an output of scan, and apply bot the log and sigmoid outside scan such that Theano can catch and optimize the expression. """ | 509a1157ce281f27267369f92c339d7bd5791429 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/242/509a1157ce281f27267369f92c339d7bd5791429/rbm.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
266,
591,
4549,
67,
12398,
12,
2890,
16,
4533,
16,
675,
67,
7340,
30501,
67,
16588,
4672,
3536,
25789,
5115,
358,
326,
25792,
4549,
555,
225,
3609,
716,
333,
445,
4991,
326,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
266,
591,
4549,
67,
12398,
12,
2890,
16,
4533,
16,
675,
67,
7340,
30501,
67,
16588,
4672,
3536,
25789,
5115,
358,
326,
25792,
4549,
555,
225,
3609,
716,
333,
445,
4991,
326,
6... |
top[i] = new | if SCons.Util.is_List(new): top[i:i+1] = new else: top[i] = new | def Replace(self, old, new): for top, lists, nonlists in FileListWalk(self, topdown=False): try: i = top.index(old) except ValueError: pass else: top[i] = new | e0b3bb7143d8b6461777c55e2d6d3bb25e1a4720 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5060/e0b3bb7143d8b6461777c55e2d6d3bb25e1a4720/chromium_builders.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6910,
12,
2890,
16,
1592,
16,
394,
4672,
364,
1760,
16,
6035,
16,
1661,
9772,
316,
1387,
682,
9463,
12,
2890,
16,
1760,
2378,
33,
8381,
4672,
775,
30,
277,
273,
1760,
18,
1615,
12,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6910,
12,
2890,
16,
1592,
16,
394,
4672,
364,
1760,
16,
6035,
16,
1661,
9772,
316,
1387,
682,
9463,
12,
2890,
16,
1760,
2378,
33,
8381,
4672,
775,
30,
277,
273,
1760,
18,
1615,
12,
1... |
for attr in ["pnode", "snode", "iallocator", "hypervisor"]: | for attr in ["pnode", "snode", "iallocator", "hypervisor", "disk_template"]: | def CheckArguments(self): """Check arguments. | f276c4b5558338ba7832b5a92b822f0695a5ac6c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7542/f276c4b5558338ba7832b5a92b822f0695a5ac6c/cmdlib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2073,
4628,
12,
2890,
4672,
3536,
1564,
1775,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2073,
4628,
12,
2890,
4672,
3536,
1564,
1775,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
filename) | os.path.basename(filepath)) | def decode_filename(filename): """ Breaks filename into component parts based on regex for example: firefox-3.0b3pre.en-US.linux-i686.complete.mar Returns dict with keys product, version, locale, platform, type """ try: m = re.search( '(?P<product>\w+)(-)(?P<version>\w+\.\w+)(\.)(?P<locale>.+?)(\.)(?P<platform>.+?)(\.)(?P<type>\w+)(.mar)', filename) return m.groupdict() except Exception, exc: raise Exception("could not parse filename %s: %s" % (filename, exc)) | a58f0266a0bdadbbef96708435c6cd47901860ae /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/11102/a58f0266a0bdadbbef96708435c6cd47901860ae/make_incremental_updates.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2495,
67,
3459,
12,
3459,
4672,
3536,
17030,
87,
1544,
1368,
1794,
2140,
2511,
603,
3936,
364,
3454,
30,
4452,
13873,
17,
23,
18,
20,
70,
23,
1484,
18,
275,
17,
3378,
18,
20132,
17,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2495,
67,
3459,
12,
3459,
4672,
3536,
17030,
87,
1544,
1368,
1794,
2140,
2511,
603,
3936,
364,
3454,
30,
4452,
13873,
17,
23,
18,
20,
70,
23,
1484,
18,
275,
17,
3378,
18,
20132,
17,
... |
``associated'' in the literature) partition of the partition pi which is | ``associated'' in the literature) partition of the partition p which is | def conjugate(self): """ conjugate() returns the ``conjugate'' (also called ``associated'' in the literature) partition of the partition pi which is obtained by transposing the corresponding Ferrers diagram. | 7896fee27ebd1c280842dcd38fb5539dd2c77a15 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/7896fee27ebd1c280842dcd38fb5539dd2c77a15/partition.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10550,
31529,
12,
2890,
4672,
3536,
10550,
31529,
1435,
1135,
326,
12176,
591,
78,
31529,
6309,
261,
31144,
2566,
12176,
28441,
6309,
316,
326,
328,
2165,
1231,
13,
3590,
434,
326,
3590,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10550,
31529,
12,
2890,
4672,
3536,
10550,
31529,
1435,
1135,
326,
12176,
591,
78,
31529,
6309,
261,
31144,
2566,
12176,
28441,
6309,
316,
326,
328,
2165,
1231,
13,
3590,
434,
326,
3590,
2... |
import imp | def tmpl_import(name): if os.sep not in name: return __import__(name) else: dir = os.path.dirname(name) modname = os.path.basename(name) import imp modfile, pathname, description = imp.find_module(modname, [dir]) try: return imp.load_module('template_' + modname, modfile, pathname, description) finally: modfile.close() | 3b7c9db09b88b8b3aa775c397f0565bc369ed89d /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10695/3b7c9db09b88b8b3aa775c397f0565bc369ed89d/server.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10720,
67,
5666,
12,
529,
4672,
309,
1140,
18,
10814,
486,
316,
508,
30,
327,
1001,
5666,
972,
12,
529,
13,
469,
30,
1577,
273,
1140,
18,
803,
18,
12287,
12,
529,
13,
16037,
273,
114... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10720,
67,
5666,
12,
529,
4672,
309,
1140,
18,
10814,
486,
316,
508,
30,
327,
1001,
5666,
972,
12,
529,
13,
469,
30,
1577,
273,
1140,
18,
803,
18,
12287,
12,
529,
13,
16037,
273,
114... | |
logger.error('xmlrpc exception: %s', reduce_homedir(e.faultCode.strip())) | e_fc = str(e.faultCode) logger.error('xmlrpc exception: %s', reduce_homedir(e_fc.strip())) | def load_mod_info(mdir, module): """ :param module: The name of the module (sale, purchase, ...) """ for filename in ['__openerp__.py', '__terp__.py']: description_file = os.path.join(mdir, module, filename) # Broken for zip modules. if description_file and os.path.isfile(description_file): return eval(open(description_file,'rb').read()) return {} | ce61ee67edfad756fb002d0ff6bbdd868145fdc6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12355/ce61ee67edfad756fb002d0ff6bbdd868145fdc6/base_quality_interrogation.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
67,
1711,
67,
1376,
12,
81,
1214,
16,
1605,
4672,
3536,
294,
891,
1605,
30,
1021,
508,
434,
326,
1605,
261,
87,
5349,
16,
23701,
16,
1372,
13,
3536,
364,
1544,
316,
10228,
972,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
67,
1711,
67,
1376,
12,
81,
1214,
16,
1605,
4672,
3536,
294,
891,
1605,
30,
1021,
508,
434,
326,
1605,
261,
87,
5349,
16,
23701,
16,
1372,
13,
3536,
364,
1544,
316,
10228,
972,
... |
this = apply(_quickfix.new_MDEntrySpotRate, args) | this = _quickfix.new_MDEntrySpotRate(*args) | def __init__(self, *args): this = apply(_quickfix.new_MDEntrySpotRate, args) try: self.this.append(this) except: self.this = this | 7e632099fd421880c8c65fb0cf610d338d115ee9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8819/7e632099fd421880c8c65fb0cf610d338d115ee9/quickfix.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
380,
1968,
4672,
333,
273,
389,
19525,
904,
18,
2704,
67,
6188,
1622,
17292,
4727,
30857,
1968,
13,
775,
30,
365,
18,
2211,
18,
6923,
12,
2211,
13,
1335,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
380,
1968,
4672,
333,
273,
389,
19525,
904,
18,
2704,
67,
6188,
1622,
17292,
4727,
30857,
1968,
13,
775,
30,
365,
18,
2211,
18,
6923,
12,
2211,
13,
1335,... |
for header in resp.getheaders(): self.infostring += header[0] + ": " + header[1] + "\r\n" | if resp != None: for header in resp.getheaders(): self.infostring += header[0] + ": " + header[1] + "\r\n" | def __init__(self, url): self.infostring = "" self.url = url urlparts = urlparse.urlparse(url) self.scheme = urlparts.scheme if self.scheme == "http": # need to set default port here port = httplib.HTTP_PORT try: if urlparts.port != None: port = urlparts.port except ValueError: self.infostring += _("Response") + ": " + _("Bad URL") + "\r\n" return conn = download.HTTPConnection(urlparts.hostname, port) try: conn.request("HEAD", url) except socket.error, error: self.infostring += _("Response") + ": " + _("Connection Error") + "\r\n" return | f0fc3f3d4f260e51b53b5886e4aecb19001929df /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/6425/f0fc3f3d4f260e51b53b5886e4aecb19001929df/checker.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
880,
4672,
365,
18,
18227,
371,
273,
1408,
365,
18,
718,
273,
880,
880,
6019,
273,
15185,
18,
718,
2670,
12,
718,
13,
365,
18,
9068,
273,
880,
6019,
18... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
880,
4672,
365,
18,
18227,
371,
273,
1408,
365,
18,
718,
273,
880,
880,
6019,
273,
15185,
18,
718,
2670,
12,
718,
13,
365,
18,
9068,
273,
880,
6019,
18... |
return str(x0) else: return "%d,%d" % (x0, x1) | return "line " + str(x0) else: return "lines %d-%d" % (x0, x1) | def pair(x0, x1): x0 += 1 if x0 >= x1: return str(x0) else: return "%d,%d" % (x0, x1) | e35ebca574f26c382a6a5d825fcc932185e500cb /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/e35ebca574f26c382a6a5d825fcc932185e500cb/regrtest.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3082,
12,
92,
20,
16,
619,
21,
4672,
619,
20,
1011,
404,
309,
619,
20,
1545,
619,
21,
30,
327,
609,
12,
92,
20,
13,
469,
30,
327,
2213,
72,
15529,
72,
6,
738,
261,
92,
20,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3082,
12,
92,
20,
16,
619,
21,
4672,
619,
20,
1011,
404,
309,
619,
20,
1545,
619,
21,
30,
327,
609,
12,
92,
20,
13,
469,
30,
327,
2213,
72,
15529,
72,
6,
738,
261,
92,
20,
16,
... |
else | else: | def __init__(data = None) if data == None: quickfix.StringField.__init__(self, 527) else quickfix.StringField.__init__(self, 527, data) | 484890147d4b23aac4b9d0e85e84fceab7e137c3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8819/484890147d4b23aac4b9d0e85e84fceab7e137c3/quickfix_fields.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
892,
273,
599,
13,
309,
501,
422,
599,
30,
9549,
904,
18,
780,
974,
16186,
2738,
972,
12,
2890,
16,
1381,
5324,
13,
469,
30,
9549,
904,
18,
780,
974,
16186,
2738... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
892,
273,
599,
13,
309,
501,
422,
599,
30,
9549,
904,
18,
780,
974,
16186,
2738,
972,
12,
2890,
16,
1381,
5324,
13,
469,
30,
9549,
904,
18,
780,
974,
16186,
2738... |
if DEBUG and endchar: self.err_write( "subconvert(%s)\n line = %s\n" % (`endchar`, `line[:20]`)) | def subconvert(self, endchar=None, depth=0): stack = [] line = self.line if DEBUG and endchar: self.err_write( "subconvert(%s)\n line = %s\n" % (`endchar`, `line[:20]`)) while line: if line[0] == endchar and not stack: if DEBUG: self.err_write("subconvert() --> %s\n" % `line[1:21]`) self.line = line return line m = _comment_rx.match(line) if m: text = m.group(1) if text: self.write("(COMMENT\n- %s \n)COMMENT\n-\\n\n" % encode(text)) line = line[m.end():] continue m = _begin_env_rx.match(line) if m: # re-write to use the macro handler line = r"\%s %s" % (m.group(1), line[m.end():]) continue m = _end_env_rx.match(line) if m: # end of environment envname = m.group(1) if envname == "document": # special magic for n in stack[1:]: if n not in self.autoclosing: raise LaTeXFormatError( "open element on stack: " + `n`) # should be more careful, but this is easier to code: stack = [] self.write(")document\n") elif stack and envname == stack[-1]: self.write(")%s\n" % envname) del stack[-1] popping(envname, "a", len(stack) + depth) else: self.err_write("stack: %s\n" % `stack`) raise LaTeXFormatError( "environment close for %s doesn't match" % envname) line = line[m.end():] continue m = _begin_macro_rx.match(line) if m: # start of macro macroname = m.group(1) if macroname == "verbatim": # really magic case! pos = string.find(line, "\\end{verbatim}") text = line[m.end(1):pos] self.write("(verbatim\n") self.write("-%s\n" % encode(text)) self.write(")verbatim\n") line = line[pos + len("\\end{verbatim}"):] continue numbered = 1 opened = 0 if macroname[-1] == "*": macroname = macroname[:-1] numbered = 0 if macroname in self.autoclosing and macroname in stack: while stack[-1] != macroname: top = stack.pop() if top and top not in self.discards: self.write(")%s\n-\\n\n" % top) popping(top, "b", len(stack) + depth) if macroname not in self.discards: self.write("-\\n\n)%s\n-\\n\n" % macroname) popping(macroname, "c", len(stack) + depth - 1) del stack[-1] # if macroname in self.discards: self.push_output(StringIO.StringIO()) else: self.push_output(self.ofp) # params, optional, empty, environ = self.start_macro(macroname) if not numbered: self.write("Anumbered TOKEN no\n") # rip off the macroname if params: line = line[m.end(1):] elif empty: line = line[m.end(1):] else: line = line[m.end():] # # Very ugly special case to deal with \item[]. The catch # is that this needs to occur outside the for loop that # handles attribute parsing so we can 'continue' the outer # loop. # if optional and type(params[0]) is TupleType: # the attribute name isn't used in this special case pushing(macroname, "a", depth + len(stack)) stack.append(macroname) self.write("(%s\n" % macroname) m = _start_optional_rx.match(line) if m: self.line = line[m.end():] line = self.subconvert("]", depth + len(stack)) line = "}" + line continue # handle attribute mappings here: for attrname in params: if optional: optional = 0 if type(attrname) is StringType: m = _optional_rx.match(line) if m: line = line[m.end():] self.write("A%s TOKEN %s\n" % (attrname, encode(m.group(1)))) elif type(attrname) is TupleType: # This is a sub-element; but place the and attribute # we found on the stack (\section-like); the # content of the macro will become the content # of the attribute element, and the macro will # have to be closed some other way (such as # auto-closing). pushing(macroname, "b", len(stack) + depth) stack.append(macroname) self.write("(%s\n" % macroname) macroname = attrname[0] m = _start_group_rx.match(line) if m: line = line[m.end():] elif type(attrname) is ListType: # A normal subelement: <macroname><attrname>...</>... attrname = attrname[0] if not opened: opened = 1 self.write("(%s\n" % macroname) pushing(macroname, "c", len(stack) + depth) self.write("(%s\n" % attrname) pushing(attrname, "sub-elem", len(stack) + depth + 1) self.line = skip_white(line)[1:] line = self.subconvert("}", len(stack) + depth + 1)[1:] popping(attrname, "sub-elem", len(stack) + depth + 1) self.write(")%s\n" % attrname) else: m = _parameter_rx.match(line) if not m: raise LaTeXFormatError( "could not extract parameter %s for %s: %s" % (attrname, macroname, `line[:100]`)) value = m.group(1) if _token_rx.match(value): dtype = "TOKEN" else: dtype = "CDATA" self.write("A%s %s %s\n" % (attrname, dtype, encode(value))) line = line[m.end():] if params and type(params[-1]) is StringType \ and (not empty) and not environ: # attempt to strip off next '{' m = _start_group_rx.match(line) if not m: raise LaTeXFormatError( "non-empty element '%s' has no content: %s" % (macroname, line[:12])) line = line[m.end():] if not opened: self.write("(%s\n" % macroname) pushing(macroname, "d", len(stack) + depth) if empty: line = "}" + line stack.append(macroname) self.pop_output() continue if line[0] == endchar and not stack: if DEBUG: self.err_write("subconvert() --> %s\n" % `line[1:21]`) self.line = line[1:] return self.line if line[0] == "}": # end of macro or group macroname = stack[-1] conversion = self.table.get(macroname) if macroname \ and macroname not in self.discards \ and type(conversion) is not StringType: # otherwise, it was just a bare group self.write(")%s\n" % stack[-1]) popping(macroname, "d", len(stack) + depth - 1) del stack[-1] line = line[1:] continue if line[0] == "{": pushing("", "e", len(stack) + depth) stack.append("") line = line[1:] continue if line[0] == "\\" and line[1] in ESCAPED_CHARS: self.write("-%s\n" % encode(line[1])) line = line[2:] continue if line[:2] == r"\\": self.write("(BREAK\n)BREAK\n") line = line[2:] continue m = _text_rx.match(line) if m: text = encode(m.group()) self.write("-%s\n" % text) line = line[m.end():] continue # special case because of \item[] if line[0] == "]": self.write("-]\n") line = line[1:] continue # avoid infinite loops extra = "" if len(line) > 100: extra = "..." raise LaTeXFormatError("could not identify markup: %s%s" % (`line[:100]`, extra)) while stack and stack[-1] in self.autoclosing: self.write("-\\n\n") self.write(")%s\n" % stack[-1]) popping(stack.pop(), "e", len(stack) + depth - 1) if stack: raise LaTeXFormatError("elements remain on stack: " + string.join(stack, ", ")) # otherwise we just ran out of input here... | ab75840fa4cfb9d978e30ffcae071fbad52b75e3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/ab75840fa4cfb9d978e30ffcae071fbad52b75e3/latex2esis.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
720,
6283,
12,
2890,
16,
679,
3001,
33,
7036,
16,
3598,
33,
20,
4672,
2110,
273,
5378,
980,
273,
365,
18,
1369,
1323,
980,
30,
309,
980,
63,
20,
65,
422,
679,
3001,
471,
486,
2110,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
720,
6283,
12,
2890,
16,
679,
3001,
33,
7036,
16,
3598,
33,
20,
4672,
2110,
273,
5378,
980,
273,
365,
18,
1369,
1323,
980,
30,
309,
980,
63,
20,
65,
422,
679,
3001,
471,
486,
2110,
... | |
if (ProxyUtil.findHTTPProxy()): DebugPrint(0, 'WARNING: http_proxy is set but not supported') __connection = httplib.HTTPConnection(Config.get_SOAPHost()) | if (ProxyUtil.findHTTPProxy()): DebugPrint(0, 'WARNING: http_proxy is set but not supported') __connection = httplib.HTTPConnection(Config.get_SOAPHost()) | def __connect(): global __connection global __connected global __connectionError global __certificateRejected global __connectionRetries global __retryDelay global __last_retry_time #__connectionError = True #return __connected if __connectionError: __disconnect() __connectionError = False if __connectionRetries > MaxConnectionRetries: current_time = time.time() if not __last_retry_time: # Set time but do not reset failures __last_retry_time = current_time return __connected; if (current_time - __last_retry_time) > __retryDelay: __last_retry_time = current_time DebugPrint(1, "Retry connection after ", __retryDelay, "s") __retryDelay = __retryDelay * __backoff_factor if __retryDelay > __maximumDelay: __retryDelay = __maximumDelay __connectionRetries = 0 __connectionRetries = __connectionRetries + 1 if (not __connected) and (__connectionRetries <= MaxConnectionRetries): if Config.get_UseSSL() == 0 and Config.get_UseSoapProtocol() == 1: if (ProxyUtil.findHTTPProxy()): DebugPrint(0, 'WARNING: http_proxy is set but not supported') __connection = httplib.HTTP(Config.get_SOAPHost()) DebugPrint(1, 'Connected via HTTP to: ' + Config.get_SOAPHost()) #print "Using SOAP protocol" elif Config.get_UseSSL() == 0 and Config.get_UseSoapProtocol() == 0: try: if (ProxyUtil.findHTTPProxy()): DebugPrint(0, 'WARNING: http_proxy is set but not supported') #__connection = ProxyUtil.HTTPConnection(Config.get_SOAPHost(), # http_proxy = ProxyUtil.findHTTPProxy()) __connection = httplib.HTTPConnection(Config.get_SOAPHost()) except Exception, e: DebugPrint(0, "ERROR: could not initialize HTTP connection") DebugPrintTraceback() __connectionError = True return __connected try: DebugPrint(4, "DEBUG: Connect") __connection.connect() DebugPrint(4, "DEBUG: Connect: OK") except Exception, e: DebugPrint(4, "DEBUG: Connect: FAILED") DebugPrint(0, "Error: While trying to connect to HTTP, caught exception " + str(e)) DebugPrintTraceback() __connectionError = True return __connected DebugPrint(1,"Connection via HTTP to: " + Config.get_SOAPHost()) #print "Using POST protocol" else: # assert(Config.get_UseSSL() == 1) if Config.get_UseGratiaCertificates() == 0: pr_cert_file = Config.get_CertificateFile() pr_key_file = Config.get_KeyFile() else: pr_cert_file = Config.get_GratiaCertificateFile() pr_key_file = Config.get_GratiaKeyFile() if (pr_cert_file == None): DebugPrint(0, "Error: While trying to connect to HTTPS, no valid local certificate.") __connectionError = True return __connected DebugPrint(4, "DEBUG: Attempting to connect to HTTPS") try: if (ProxyUtil.findHTTPSProxy()): DebugPrint(0, 'WARNING: http_proxy is set but not supported when UseSoapProtocol is set to 1') #__connection = ProxyUtil.HTTPSConnection(Config.get_SSLHost(), # cert_file = pr_cert_file, # key_file = pr_key_file, # http_proxy = ProxyUtil.findHTTPSProxy()) __connection = httplib.HTTPSConnection(Config.get_SSLHost(), cert_file = pr_cert_file, key_file = pr_key_file) except Exception, e: DebugPrint(0, "ERROR: could not initialize HTTPS connection") DebugPrintTraceback() __connectionError = True return __connected try: DebugPrint(4, "DEBUG: Connect") __connection.connect() DebugPrint(4, "DEBUG: Connect: OK") except Exception, e: DebugPrint(4, "DEBUG: Connect: FAILED") DebugPrint(0, "Error: While trying to connect to HTTPS, caught exception " + str(e)) DebugPrintTraceback() __connectionError = True return __connected DebugPrint(1, "Connected via HTTPS to: " + Config.get_SSLHost()) #print "Using SSL protocol" # Successful DebugPrint(4, "DEBUG: Connection SUCCESS") __connected = True # Reset connection retry count to 0 and the retry delay to its initial value __connectionRetries = 0 __retryDelay = __initialDelay return __connected | 637869f92a8322c20ea53836dc7393fc485f077a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/637/637869f92a8322c20ea53836dc7393fc485f077a/Gratia.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
3612,
13332,
2552,
1001,
4071,
2552,
1001,
8537,
2552,
1001,
4071,
668,
2552,
1001,
14108,
19902,
2552,
1001,
4071,
13656,
2552,
1001,
9620,
6763,
2552,
1001,
2722,
67,
9620,
67,
957,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
3612,
13332,
2552,
1001,
4071,
2552,
1001,
8537,
2552,
1001,
4071,
668,
2552,
1001,
14108,
19902,
2552,
1001,
4071,
13656,
2552,
1001,
9620,
6763,
2552,
1001,
2722,
67,
9620,
67,
957,
... |
if excepted != actual: | if expected != actual: | def element_text_should_be(self, locator, excepted, message=''): """Verifies element identified by `locator` exactly contains text `expected`. In contrast to `Element Should Contain`, this keyword does not try a substring match but an exact match on the element identified by `locator`. | 7e5a3533b01b02f3592aed251c78e2c3aac0db38 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/2148/7e5a3533b01b02f3592aed251c78e2c3aac0db38/assertion.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
930,
67,
955,
67,
13139,
67,
2196,
12,
2890,
16,
8871,
16,
1335,
329,
16,
883,
2218,
11,
4672,
3536,
19802,
930,
9283,
635,
1375,
20048,
68,
8950,
1914,
977,
1375,
3825,
8338,
225,
657... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
930,
67,
955,
67,
13139,
67,
2196,
12,
2890,
16,
8871,
16,
1335,
329,
16,
883,
2218,
11,
4672,
3536,
19802,
930,
9283,
635,
1375,
20048,
68,
8950,
1914,
977,
1375,
3825,
8338,
225,
657... |
def purge(base_dir, gigs, ignore, dry_run=False): """Delete directories under `base_dir` until `gigs` GB are free | def purge(base_dirs, gigs, ignore, dry_run=False): """Delete directories under `base_dirs` until `gigs` GB are free | def purge(base_dir, gigs, ignore, dry_run=False): """Delete directories under `base_dir` until `gigs` GB are free Will not delete directories listed in the ignore list.""" gigs *= 1024 * 1024 * 1024 if freespace(base_dir) >= gigs: return dirs = [os.path.join(base_dir, d) for d in os.listdir(base_dir) if os.path.isdir(os.path.join(base_dir, d)) and d not in ignore] dirs.sort(mtime_sort) while dirs and freespace(base_dir) < gigs: d = dirs.pop(0) print "Deleting", d if not dry_run: try: clobber_path=d+clobber_suffix if os.path.exists(clobber_path): rmdirRecursive(clobber_path) # Prevent repeated moving. if d.endswith(clobber_suffix): rmdirRecursive(d) else: shutil.move(d, clobber_path) rmdirRecursive(clobber_path) except: print >>sys.stderr, "Couldn't purge %s properly. Skipping." % d | c12b0ed3642211354e16f3ad24c1b0c1d06b35bc /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6206/c12b0ed3642211354e16f3ad24c1b0c1d06b35bc/purge_builds.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11668,
12,
1969,
67,
8291,
16,
314,
360,
87,
16,
2305,
16,
10299,
67,
2681,
33,
8381,
4672,
3536,
2613,
6402,
3613,
1375,
1969,
67,
8291,
68,
3180,
1375,
75,
360,
87,
68,
25069,
854,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11668,
12,
1969,
67,
8291,
16,
314,
360,
87,
16,
2305,
16,
10299,
67,
2681,
33,
8381,
4672,
3536,
2613,
6402,
3613,
1375,
1969,
67,
8291,
68,
3180,
1375,
75,
360,
87,
68,
25069,
854,
... |
def flimage_color_to_pixel(p1, p2, p3, p4, p5): """ flimage_color_to_pixel(p1, p2, p3, p4, p5) -> num. """ retval = _flimage_color_to_pixel(p1, p2, p3, p4, p5) | def flimage_color_to_pixel(pImage, p2, p3, p4, p5): """ flimage_color_to_pixel(pImage, p2, p3, p4, p5) -> num. """ retval = _flimage_color_to_pixel(pImage, p2, p3, p4, p5) | def flimage_color_to_pixel(p1, p2, p3, p4, p5): """ flimage_color_to_pixel(p1, p2, p3, p4, p5) -> num. """ retval = _flimage_color_to_pixel(p1, p2, p3, p4, p5) return retval | 8765c710f695de392f6fc7c664c746ec98668b1d /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/2429/8765c710f695de392f6fc7c664c746ec98668b1d/xformslib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1183,
2730,
67,
3266,
67,
869,
67,
11743,
12,
84,
2040,
16,
293,
22,
16,
293,
23,
16,
293,
24,
16,
293,
25,
4672,
3536,
1183,
2730,
67,
3266,
67,
869,
67,
11743,
12,
84,
2040,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1183,
2730,
67,
3266,
67,
869,
67,
11743,
12,
84,
2040,
16,
293,
22,
16,
293,
23,
16,
293,
24,
16,
293,
25,
4672,
3536,
1183,
2730,
67,
3266,
67,
869,
67,
11743,
12,
84,
2040,
16,
... |
v, maxlinelen=self.__maxheaderlen, | v, maxlinelen=self._maxheaderlen, | def _write_headers(self, msg): for h, v in msg.items(): print >> self._fp, '%s:' % h, if self.__maxheaderlen == 0: # Explicit no-wrapping print >> self._fp, v elif isinstance(v, Header): # Header instances know what to do print >> self._fp, v.encode() elif _is8bitstring(v): # If we have raw 8bit data in a byte string, we have no idea # what the encoding is. There is no safe way to split this # string. If it's ascii-subset, then we could do a normal # ascii split, but if it's multibyte then we could break the # string. There's no way to know so the least harm seems to # be to not split the string and risk it being too long. print >> self._fp, v else: # Header's got lots of smarts, so use it. print >> self._fp, Header( v, maxlinelen=self.__maxheaderlen, header_name=h, continuation_ws='\t').encode() # A blank line always separates headers from body print >> self._fp | 36112f2d34253c36784302cb703977ec78ccc7b2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/36112f2d34253c36784302cb703977ec78ccc7b2/Generator.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2626,
67,
2485,
12,
2890,
16,
1234,
4672,
364,
366,
16,
331,
316,
1234,
18,
3319,
13332,
1172,
1671,
365,
6315,
7944,
16,
1995,
87,
2497,
738,
366,
16,
309,
365,
16186,
1896,
3374... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2626,
67,
2485,
12,
2890,
16,
1234,
4672,
364,
366,
16,
331,
316,
1234,
18,
3319,
13332,
1172,
1671,
365,
6315,
7944,
16,
1995,
87,
2497,
738,
366,
16,
309,
365,
16186,
1896,
3374... |
config_file: location of the config file | config_file (string): location of the config file map_areas_prefix (string): map_areas table name prefix | def __init__(self, config_file=None, city_name=None, boundingbox=None, osmid=None, language=None): """Creates a new OCitySMap renderer instance for the given city. | 6179e7aca483e9d7ad5ce52cd02fb358c6e761ef /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9326/6179e7aca483e9d7ad5ce52cd02fb358c6e761ef/street_index.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
642,
67,
768,
33,
7036,
16,
12797,
67,
529,
33,
7036,
16,
12638,
33,
7036,
16,
1140,
13138,
33,
7036,
16,
2653,
33,
7036,
4672,
3536,
2729,
279,
394,
5... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
642,
67,
768,
33,
7036,
16,
12797,
67,
529,
33,
7036,
16,
12638,
33,
7036,
16,
1140,
13138,
33,
7036,
16,
2653,
33,
7036,
4672,
3536,
2729,
279,
394,
5... |
return " ".join([x for x in self.name, self.value, self.description if isinstance(x, basestring) and len(x) > 0 and not x.isspace()]) | basicText=" ".join([x for x in self.name, self.value, self.description if isinstance(x, basestring) and len(x) > 0 and not x.isspace()]) if len(basicText)==0: basicText="\n" return basicText | def _get_basicText(self): return " ".join([x for x in self.name, self.value, self.description if isinstance(x, basestring) and len(x) > 0 and not x.isspace()]) | 7bbcb26df7bcaa33983ac9ecef49ddd633221275 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9340/7bbcb26df7bcaa33983ac9ecef49ddd633221275/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
13240,
1528,
12,
2890,
4672,
327,
315,
3552,
5701,
3816,
92,
364,
619,
316,
365,
18,
529,
16,
365,
18,
1132,
16,
365,
18,
3384,
309,
1549,
12,
92,
16,
10699,
13,
471,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
13240,
1528,
12,
2890,
4672,
327,
315,
3552,
5701,
3816,
92,
364,
619,
316,
365,
18,
529,
16,
365,
18,
1132,
16,
365,
18,
3384,
309,
1549,
12,
92,
16,
10699,
13,
471,
... |
"Pythonwin": "Pythonwin/pywin"}, | "Pythonwin": "Pythonwin"}, | def swig_sources (self, sources): new_sources = [] swig_sources = [] swig_targets = {} # XXX this drops generated C/C++ files into the source tree, which # is fine for developers who want to distribute the generated # source -- but there should be an option to put SWIG output in # the temp dir. target_ext = '.cpp' for source in sources: (base, ext) = os.path.splitext(source) if ext == ".i": # SWIG interface file # Seems the files SWIG creates are not compiled separately, # they are #included somewhere else. So we don't include # the generated wrapper in the new_sources list. swig_sources.append(source) # and win32all has it's own naming convention for the wrappers: if base.endswith("win32pipe") or base.endswith("win32security"): swig_targets[source] = base + 'module' + target_ext else: swig_targets[source] = base + 'module_win32' + target_ext else: new_sources.append(source) | 6bdda3ba83a5f691e3af1da556fd0e9b2bc1a7e2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/677/6bdda3ba83a5f691e3af1da556fd0e9b2bc1a7e2/setup_win32all_core.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1352,
360,
67,
10141,
261,
2890,
16,
5550,
4672,
394,
67,
10141,
273,
5378,
1352,
360,
67,
10141,
273,
5378,
1352,
360,
67,
11358,
273,
2618,
468,
11329,
333,
29535,
4374,
385,
19,
39,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1352,
360,
67,
10141,
261,
2890,
16,
5550,
4672,
394,
67,
10141,
273,
5378,
1352,
360,
67,
10141,
273,
5378,
1352,
360,
67,
11358,
273,
2618,
468,
11329,
333,
29535,
4374,
385,
19,
39,
... |
self.log.info('submitTasks: Failed to submit %d tasks to RMS.' % (failed,time.time()-startTime)) | self.log.info('submitTasks: Failed to submit %d tasks to RMS.' % (failed)) | def submitTasks(self,taskDict): submitted = 0 failed = 0 startTime = time.time() for taskID in sortLit(taskDict.keys()): taskDict[taskID] res = self.submitToExternal(taskDict[taskID]['TaskObject']) if res['OK']: taskDict[taskID]['ExternalID'] = res['Value'] taskDict[taskID]['Success'] = True submitted +=1 else: self.log.warn("Failed to submit task to WMS",res['Message']) taskDict[taskID]['Success'] = False failed += 1 self.log.info('submitTasks: Submitted %d tasks to RMS in %.1f seconds' % (submitted,time.time()-startTime)) self.log.info('submitTasks: Failed to submit %d tasks to RMS.' % (failed,time.time()-startTime)) return S_OK(taskDict) | 7e361d624cd9374c5dec904aad18c8eb176604b6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/7e361d624cd9374c5dec904aad18c8eb176604b6/TaskManager.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4879,
6685,
12,
2890,
16,
4146,
5014,
4672,
9638,
273,
374,
2535,
273,
374,
8657,
273,
813,
18,
957,
1435,
364,
29772,
316,
1524,
23707,
12,
4146,
5014,
18,
2452,
1435,
4672,
1562,
5014,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4879,
6685,
12,
2890,
16,
4146,
5014,
4672,
9638,
273,
374,
2535,
273,
374,
8657,
273,
813,
18,
957,
1435,
364,
29772,
316,
1524,
23707,
12,
4146,
5014,
18,
2452,
1435,
4672,
1562,
5014,... |
import warnings, traceback, StringIO | import io, warnings, traceback | def _warn_unhandled_exception(): # There are a few catch-all except: statements in this module, for # catching input that's bad in unexpected ways. Warn if any # exceptions are caught there. import warnings, traceback, StringIO f = io.StringIO() traceback.print_exc(None, f) msg = f.getvalue() warnings.warn("cookielib bug!\n%s" % msg, stacklevel=2) | aeb23c0340a22e2d737b490b94e874840e709ebd /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8125/aeb23c0340a22e2d737b490b94e874840e709ebd/cookielib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8935,
67,
318,
14387,
67,
4064,
13332,
468,
6149,
854,
279,
11315,
1044,
17,
454,
1335,
30,
6317,
316,
333,
1605,
16,
364,
468,
1044,
310,
810,
716,
1807,
5570,
316,
9733,
16226,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8935,
67,
318,
14387,
67,
4064,
13332,
468,
6149,
854,
279,
11315,
1044,
17,
454,
1335,
30,
6317,
316,
333,
1605,
16,
364,
468,
1044,
310,
810,
716,
1807,
5570,
316,
9733,
16226,
... |
ig, r2 = self.fit_item_in_rectangle( head_val.rectangle, col_size, head_h ) | ig, r2 = self.fit_item_in_rectangle( head_val.rectangle, col_size, head_h, head_h ) | def update_content(self): """ update the listing area """ | ec09a89abffc006a16c2262aef73f463346519b4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11399/ec09a89abffc006a16c2262aef73f463346519b4/tvlisting_area.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
67,
1745,
12,
2890,
4672,
3536,
1089,
326,
11591,
5091,
3536,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
67,
1745,
12,
2890,
4672,
3536,
1089,
326,
11591,
5091,
3536,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
@param shrink if True child can be made smaller than its minimum size request | @param shrink if True child can be made smaller than its minimum size request | def bind_session(self, ptype, persp_klass, show_pers=True, resize=False, shrink=True): """ Bind the perspective 'pers_klass' to Session 'ptype' | 05f68a255d3ea85d6e22c14339e6ebf36badb224 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11468/05f68a255d3ea85d6e22c14339e6ebf36badb224/MainWindow.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1993,
67,
3184,
12,
2890,
16,
24203,
16,
13508,
84,
67,
22626,
16,
2405,
67,
10422,
33,
5510,
16,
7041,
33,
8381,
16,
17697,
33,
5510,
4672,
3536,
6936,
326,
26651,
296,
10422,
67,
226... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1993,
67,
3184,
12,
2890,
16,
24203,
16,
13508,
84,
67,
22626,
16,
2405,
67,
10422,
33,
5510,
16,
7041,
33,
8381,
16,
17697,
33,
5510,
4672,
3536,
6936,
326,
26651,
296,
10422,
67,
226... |
matrix_integer_dense, matrix_rational_dense, matrix_rational_sparse, matrix_cyclo_dense, matrix_integer_2x2, matrix_integer_sparse, matrix_double_dense, matrix_real_double_dense, matrix_change_ring, matrix_complex_double_dense, linbox, matrix_modn_dense, matrix_modn_sparse, matrix_mod2_dense, matrix_mpolynomial_dense, \ matrix_symbolic_dense, \ cremona_mat, \ cremona_homspace, \ cremona_newforms, \ finite_field_givaro, \ finite_field_ntl_gf2e, \ libsingular, \ fplll, \ | matrix_integer_dense, matrix_rational_dense, matrix_rational_sparse, matrix_cyclo_dense, matrix_integer_2x2, matrix_double_dense, matrix_integer_sparse, matrix_real_double_dense, matrix_change_ring, matrix_complex_double_dense, linbox, matrix_modn_dense, matrix_modn_sparse, matrix_mod2_dense, matrix_mpolynomial_dense, matrix_symbolic_dense, cremona_mat, cremona_homspace, cremona_newforms, finite_field_givaro, finite_field_ntl_gf2e, libsingular, fplll, | def is_newer(file1, file2): """ Return True if either file2 does not exist or is older than file1. If file1 does not exist, always return False. """ if not os.path.exists(file1): return False if not os.path.exists(file2): return True if os.path.getmtime(file2) < os.path.getmtime(file1): return True return False | 12aff705aa1e2b2dc6e401b763fe2e9e51bdf02d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/12aff705aa1e2b2dc6e401b763fe2e9e51bdf02d/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
353,
67,
2704,
264,
12,
768,
21,
16,
585,
22,
4672,
3536,
2000,
1053,
309,
3344,
585,
22,
1552,
486,
1005,
578,
353,
12156,
2353,
585,
21,
18,
225,
971,
585,
21,
1552,
486,
1005,
16,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
353,
67,
2704,
264,
12,
768,
21,
16,
585,
22,
4672,
3536,
2000,
1053,
309,
3344,
585,
22,
1552,
486,
1005,
578,
353,
12156,
2353,
585,
21,
18,
225,
971,
585,
21,
1552,
486,
1005,
16,... |
if not single_article_per_section: navPointByLetterTag['id'] = "%sTitles-ID" % (title_letters[i].upper()) | navPointByLetterTag['id'] = "%sTitles-ID" % (title_letters[i].upper()) | def generateNCXByTitle(self, tocTitle, single_article_per_section=True): | 073710f405f291eedc7992694a78db2be7983d10 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9125/073710f405f291eedc7992694a78db2be7983d10/catalog.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2103,
10346,
60,
858,
4247,
12,
2890,
16,
17919,
4247,
16,
2202,
67,
11480,
67,
457,
67,
3464,
33,
5510,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2103,
10346,
60,
858,
4247,
12,
2890,
16,
17919,
4247,
16,
2202,
67,
11480,
67,
457,
67,
3464,
33,
5510,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
self.found_type = mime_type if not self.found_type: | self.sound_file.mime_type = mime_type if not self.sound_file.mime_type: | def have_type(self, typefind, probability, caps): mime_type = caps.to_string() #debug("have_type:", mime_type, self.sound_file.get_filename_for_display()) self.found_type = None for t in mime_whitelist: if t in mime_type: self.found_type = mime_type if not self.found_type: log("Mime type skipped: %s (mail us if this is an error)" % mime_type) | a18793b0ad63f711dbec0cbd8e8e8a81457c6811 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2369/a18793b0ad63f711dbec0cbd8e8e8a81457c6811/soundconverter.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1240,
67,
723,
12,
2890,
16,
618,
4720,
16,
11331,
16,
15788,
4672,
4892,
67,
723,
273,
15788,
18,
869,
67,
1080,
1435,
468,
4148,
2932,
21516,
67,
723,
2773,
16,
4892,
67,
723,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1240,
67,
723,
12,
2890,
16,
618,
4720,
16,
11331,
16,
15788,
4672,
4892,
67,
723,
273,
15788,
18,
869,
67,
1080,
1435,
468,
4148,
2932,
21516,
67,
723,
2773,
16,
4892,
67,
723,
16,
... |
for k, v in os.environ.iteritems(): | for k in os.environ.keys(): | def setUp(self): # Records changes to env vars self.env = test_support.EnvironmentVarGuard() # Delete all proxy related env vars for k, v in os.environ.iteritems(): if 'proxy' in k.lower(): self.env.unset(k) | fd3889e27269537ad7d93a585493599047812a1a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8125/fd3889e27269537ad7d93a585493599047812a1a/test_urllib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
24292,
12,
2890,
4672,
468,
27390,
3478,
358,
1550,
4153,
365,
18,
3074,
273,
1842,
67,
13261,
18,
5494,
1537,
16709,
1435,
468,
2504,
777,
2889,
3746,
1550,
4153,
364,
417,
316,
1140,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
24292,
12,
2890,
4672,
468,
27390,
3478,
358,
1550,
4153,
365,
18,
3074,
273,
1842,
67,
13261,
18,
5494,
1537,
16709,
1435,
468,
2504,
777,
2889,
3746,
1550,
4153,
364,
417,
316,
1140,
1... |
def fit(self, method=None, effects='oneway', *opts): """ method : LSDV, demeaned, MLE, GLS, BE, FE effects : 'invidividual', 'oneway', or 'twoway' """ method = method.lower() if method not in ["lsdv", "demeaned", "mle", "gls", "be", "fe"]: | def fit(self, model=None, method=None, effects='oneway'): """ method : LSDV, demeaned, MLE, GLS, BE, FE, optional model : BTWNP : Between panels BTWNT : Between time periods FIXONE : One-way fixed effects FIXONETIME : One-way time fixed effects GMM : Dynamic panel estimator POOLED : Pooled RANONE : One-way random effects RANTWO : Two-way random effects between fixed random pooled [gmm] effects : oneway time twoway Notes ------ This is unfinished. None of the method arguments work yet. Only oneway effects should work. """ if method: method = method.lower() model = model.lower() if method and method not in ["lsdv", "demeaned", "mle", "gls", "be", "fe"]: | def fit(self, method=None, effects='oneway', *opts): """ method : LSDV, demeaned, MLE, GLS, BE, FE effects : 'invidividual', 'oneway', or 'twoway' """ method = method.lower() if method not in ["lsdv", "demeaned", "mle", "gls", "be", "fe"]: raise ValueError, "%s not a valid method" % method if method == "lsdv": self.fit_lsdv(opts) | 7de62ef8a76e8d78e08a9767ab1f536f7f5b28f1 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12658/7de62ef8a76e8d78e08a9767ab1f536f7f5b28f1/panelmod.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4845,
12,
2890,
16,
938,
33,
7036,
16,
707,
33,
7036,
16,
16605,
2218,
265,
359,
528,
11,
4672,
3536,
707,
294,
511,
9903,
58,
16,
302,
4698,
304,
329,
16,
490,
900,
16,
611,
3045,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4845,
12,
2890,
16,
938,
33,
7036,
16,
707,
33,
7036,
16,
16605,
2218,
265,
359,
528,
11,
4672,
3536,
707,
294,
511,
9903,
58,
16,
302,
4698,
304,
329,
16,
490,
900,
16,
611,
3045,
... |
self.plist.CFBundleExecutable = self.name | def preProcess(self): self.plist.CFBundleExecutable = self.name resdir = pathjoin("Contents", "Resources") if self.executable is not None: if self.mainprogram is None: execpath = pathjoin(self.execdir, self.name) else: execpath = pathjoin(resdir, os.path.basename(self.executable)) self.files.append((self.executable, execpath)) # For execve wrapper setexecutable = setExecutableTemplate % os.path.basename(self.executable) else: setexecutable = "" # XXX for locals() call | 42fffc1d3be00aca96bdd2fc2006ce5494ca3fea /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/42fffc1d3be00aca96bdd2fc2006ce5494ca3fea/bundlebuilder.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
675,
2227,
12,
2890,
4672,
400,
1214,
273,
589,
5701,
2932,
6323,
3113,
315,
3805,
7923,
309,
365,
18,
17751,
353,
486,
599,
30,
309,
365,
18,
5254,
12890,
353,
599,
30,
1196,
803,
273... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
675,
2227,
12,
2890,
4672,
400,
1214,
273,
589,
5701,
2932,
6323,
3113,
315,
3805,
7923,
309,
365,
18,
17751,
353,
486,
599,
30,
309,
365,
18,
5254,
12890,
353,
599,
30,
1196,
803,
273... | |
else: | if self.packages: | def run (self): | 1ac604e51b928235dd31e18a1f796baf4f8db869 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/1ac604e51b928235dd31e18a1f796baf4f8db869/build_py.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
261,
2890,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
261,
2890,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
checkProg('a EPS -> PDF converter', ['epstopdf'], | checkProg('an EPS -> PDF converter', ['epstopdf'], | def checkConverterEntries(): ''' Check all converters (\converter entries) ''' checkProg('the pdflatex program', ['pdflatex $$i'], rc_entry = [ r'\converter pdflatex pdf2 "%%" "latex"' ]) ''' If we're running LyX in-place then tex2lyx will be found in ../src/tex2lyx. Add this directory to the PATH temporarily and search for tex2lyx. Use PATH to avoid any problems with paths-with-spaces. ''' path_orig = os.environ["PATH"] os.environ["PATH"] = os.path.join('..', 'src', 'tex2lyx') + \ os.pathsep + path_orig checkProg('a LaTeX/Noweb -> LyX converter', ['tex2lyx', 'tex2lyx' + version_suffix], rc_entry = [r'''\converter latex lyx "%% -f $$i $$o" "" | 53e186297e30359c13977bfd83730f30df754470 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7514/53e186297e30359c13977bfd83730f30df754470/configure.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
5072,
5400,
13332,
9163,
2073,
777,
19414,
17938,
15747,
3222,
13,
9163,
866,
626,
75,
2668,
5787,
8169,
26264,
5402,
2187,
10228,
7699,
26264,
5366,
77,
17337,
4519,
67,
4099,
273,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
5072,
5400,
13332,
9163,
2073,
777,
19414,
17938,
15747,
3222,
13,
9163,
866,
626,
75,
2668,
5787,
8169,
26264,
5402,
2187,
10228,
7699,
26264,
5366,
77,
17337,
4519,
67,
4099,
273,
3... |
if w_bigint2.sign > 0: | if other.sign > 0: | def lt(w_bigint1, w_bigint2): if w_bigint1.sign > w_bigint2.sign: return False if w_bigint1.sign < w_bigint2.sign: return True ld1 = len(w_bigint1.digits) ld2 = len(w_bigint2.digits) if ld1 > ld2: if w_bigint2.sign > 0: return False else: return True elif ld1 < ld2: if w_bigint2.sign > 0: return True else: return False i = ld1 - 1 while i >= 0: d1 = w_bigint1.digits[i] d2 = w_bigint2.digits[i] if d1 < d2: if w_bigint2.sign > 0: return True else: return False elif d1 > d2: if w_bigint2.sign > 0: return False else: return True i -= 1 return False | 87da397eba037ce53e8c5847c7c4b29acfda97b2 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/6934/87da397eba037ce53e8c5847c7c4b29acfda97b2/rbigint.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13489,
12,
91,
67,
14002,
474,
21,
16,
341,
67,
14002,
474,
22,
4672,
309,
341,
67,
14002,
474,
21,
18,
2977,
405,
341,
67,
14002,
474,
22,
18,
2977,
30,
327,
1083,
309,
341,
67,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13489,
12,
91,
67,
14002,
474,
21,
16,
341,
67,
14002,
474,
22,
4672,
309,
341,
67,
14002,
474,
21,
18,
2977,
405,
341,
67,
14002,
474,
22,
18,
2977,
30,
327,
1083,
309,
341,
67,
1... |
overlays = ResolveOverlays(buildconfig['overlays']) | overlays = ResolveOverlays(buildroot, buildconfig['overlays']) | def main(): # Parse options usage = "usage: %prog [options] cbuildbot_config" parser = optparse.OptionParser(usage=usage) parser.add_option('-r', '--buildroot', help='root directory where build occurs', default=".") parser.add_option('-n', '--buildnumber', help='build number', type='int', default=0) parser.add_option('-f', '--revisionfile', help='file where new revisions are stored') parser.add_option('--clobber', action='store_true', dest='clobber', default=False, help='Clobbers an old checkout before syncing') parser.add_option('--debug', action='store_true', dest='debug', default=False, help='Override some options to run as a developer.') parser.add_option('-t', '--tracking-branch', dest='tracking_branch', default='cros/master', help='Run the buildbot on a branch') parser.add_option('-u', '--url', dest='url', default='http://git.chromium.org/git/manifest', help='Run the buildbot on internal manifest') (options, args) = parser.parse_args() buildroot = options.buildroot revisionfile = options.revisionfile tracking_branch = options.tracking_branch if len(args) >= 1: buildconfig = _GetConfig(args[-1]) else: Warning('Missing configuration description') parser.print_usage() sys.exit(1) # Calculate list of overlay directories. overlays = ResolveOverlays(buildconfig['overlays']) try: _PreFlightRinse(buildroot, buildconfig['board'], tracking_branch, overlays) if options.clobber or not os.path.isdir(buildroot): _FullCheckout(buildroot, tracking_branch, url=options.url) else: _IncrementalCheckout(buildroot) chroot_path = os.path.join(buildroot, 'chroot') if not os.path.isdir(chroot_path): _MakeChroot(buildroot) boardpath = os.path.join(chroot_path, 'build', buildconfig['board']) if not os.path.isdir(boardpath): _SetupBoard(buildroot, board=buildconfig['board']) if buildconfig['uprev']: _UprevPackages(buildroot, tracking_branch, revisionfile, buildconfig['board'], overlays) _EnableLocalAccount(buildroot) _Build(buildroot) if buildconfig['unittests']: _RunUnitTests(buildroot) _BuildImage(buildroot) if buildconfig['smoke_bvt']: _BuildVMImageForTesting(buildroot) test_results_dir = '/tmp/run_remote_tests.%s' % options.buildnumber try: _RunSmokeSuite(buildroot, test_results_dir) finally: _ArchiveTestResults(buildroot, buildconfig['board'], archive_dir=options.buildnumber, test_results_dir=test_results_dir) if buildconfig['uprev']: # Don't push changes for developers. if not options.debug: if buildconfig['master']: # Master bot needs to check if the other slaves completed. if cbuildbot_comm.HaveSlavesCompleted(config): _UprevPush(buildroot, tracking_branch, buildconfig['board'], overlays) else: Die('CBUILDBOT - One of the slaves has failed!!!') else: # Publish my status to the master if its expecting it. if buildconfig['important']: cbuildbot_comm.PublishStatus(cbuildbot_comm.STATUS_BUILD_COMPLETE) except: # Send failure to master bot. if not buildconfig['master'] and buildconfig['important']: cbuildbot_comm.PublishStatus(cbuildbot_comm.STATUS_BUILD_FAILED) raise | f9b84d61a13673957c404cda0703386a659431bc /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9626/f9b84d61a13673957c404cda0703386a659431bc/cbuildbot.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
468,
2884,
702,
4084,
273,
315,
9167,
30,
738,
14654,
306,
2116,
65,
2875,
680,
4819,
67,
1425,
6,
2082,
273,
2153,
2670,
18,
1895,
2678,
12,
9167,
33,
9167,
13,
2082,
18,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
468,
2884,
702,
4084,
273,
315,
9167,
30,
738,
14654,
306,
2116,
65,
2875,
680,
4819,
67,
1425,
6,
2082,
273,
2153,
2670,
18,
1895,
2678,
12,
9167,
33,
9167,
13,
2082,
18,... |
def __init__(data = None) | def __init__(data = None): | def __init__(data = None) if data == None: quickfix.DoubleField.__init__(self, 151) else quickfix.DoubleField.__init__(self, 151, data) | 484890147d4b23aac4b9d0e85e84fceab7e137c3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8819/484890147d4b23aac4b9d0e85e84fceab7e137c3/quickfix_fields.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
892,
273,
599,
4672,
309,
501,
422,
599,
30,
9549,
904,
18,
5265,
974,
16186,
2738,
972,
12,
2890,
16,
4711,
21,
13,
469,
9549,
904,
18,
5265,
974,
16186,
2738,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
892,
273,
599,
4672,
309,
501,
422,
599,
30,
9549,
904,
18,
5265,
974,
16186,
2738,
972,
12,
2890,
16,
4711,
21,
13,
469,
9549,
904,
18,
5265,
974,
16186,
2738,
... |
w1 = eval("w["+repr(tuple(r))+"]*ones(ash, float)") n = add.reduce(a*w1, axis) d = add.reduce(w1, axis) | r = tuple(r) n = add.reduce(a*w[r], axis) d = add.reduce(w, axis) | def average(a, axis=None, weights=None, returned=False): """Average the array over the given axis. If the axis is None, average over all dimensions of the array. Equivalent to a.mean(axis) and to a.sum(axis) / size(a, axis) If weights are given, result is: sum(a * weights,axis) / sum(weights,axis), where the weights must have a's shape or be 1D with length the size of a in the given axis. Integer weights are converted to Float. Not specifying weights is equivalent to specifying weights that are all 1. If 'returned' is True, return a tuple: the result and the sum of the weights or count of values. The shape of these two results will be the same. Raises ZeroDivisionError if appropriate. (The version in MA does not -- it returns masked values). """ if axis is None: a = array(a).ravel() if weights is None: n = add.reduce(a) d = len(a) * 1.0 else: w = array(weights).ravel() * 1.0 n = add.reduce(multiply(a, w)) d = add.reduce(w) else: a = array(a) ash = a.shape if ash == (): a.shape = (1,) if weights is None: n = add.reduce(a, axis) d = ash[axis] * 1.0 if returned: d = ones(n.shape) * d else: w = array(weights, copy=False) * 1.0 wsh = w.shape if wsh == (): wsh = (1,) if wsh == ash: n = add.reduce(a*w, axis) d = add.reduce(w, axis) elif wsh == (ash[axis],): ni = len(ash) r = [newaxis]*ni r[axis] = slice(None, None, 1) w1 = eval("w["+repr(tuple(r))+"]*ones(ash, float)") n = add.reduce(a*w1, axis) d = add.reduce(w1, axis) else: raise ValueError, 'averaging weights have wrong shape' if not isinstance(d, ndarray): if d == 0.0: raise ZeroDivisionError, 'zero denominator in average()' if returned: return n/d, d else: return n/d | 0abaf57d5c305044148d7688ce248ad050963a7f /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/14925/0abaf57d5c305044148d7688ce248ad050963a7f/function_base.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8164,
12,
69,
16,
2654,
33,
7036,
16,
5376,
33,
7036,
16,
2106,
33,
8381,
4672,
3536,
17115,
326,
526,
1879,
326,
864,
2654,
18,
225,
971,
326,
2654,
353,
599,
16,
8164,
1879,
777,
5... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8164,
12,
69,
16,
2654,
33,
7036,
16,
5376,
33,
7036,
16,
2106,
33,
8381,
4672,
3536,
17115,
326,
526,
1879,
326,
864,
2654,
18,
225,
971,
326,
2654,
353,
599,
16,
8164,
1879,
777,
5... |
elif UName == "osf1": | elif UName == "osf1": | def GetVersion(Key): # # Only read the file if an environment variable is "missing" (they # usually all are, ok), and only read it once. # #print("WriteVariablesIntoEnvironment:3") Value = Versions.get(Key) if Value: return Value # # CM3VERSION d5.7.1 # CM3VERSIONNUM 050701 # CM3LASTCHANGED 2009-01-21 # RegExp = re.compile("(" + "|".join(Versions.keys()) + ") (.+)$", re.IGNORECASE) ShFilePath = os.path.join(os.path.dirname(os.path.dirname(os.path.abspath(__file__))), "version") for Line in open(ShFilePath): Match = RegExp.match(Line) if Match: MatchKey = Match.group(1) # # We are here because one of them wasn't found, but we should be # sure only to overwrite what we don't have. # if not Versions[MatchKey]: Value = Match.group(2) Versions[MatchKey] = Value exec("%s = \"%s\"" % (MatchKey, Value), locals(), globals()) # # Make sure we found every key in the file (at least those # not defined in the environment) # MissingKey = None for Item in Versions.iteritems(): #print(Item) if Item[1] is None: MissingKey = Item[0] File = __file__ sys.stderr.write("%(File)s: %(MissingKey)s not found in %(ShFilePath)s\n" % vars()) if MissingKey: sys.exit(1) return Versions.get(Key) | 1783e8b3d0c9afbb454634de474d7f6aa0025cf0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9328/1783e8b3d0c9afbb454634de474d7f6aa0025cf0/pylib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
1444,
12,
653,
4672,
468,
468,
5098,
855,
326,
585,
309,
392,
3330,
2190,
353,
315,
7337,
6,
261,
451,
402,
468,
11234,
777,
854,
16,
1529,
3631,
471,
1338,
855,
518,
3647,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
1444,
12,
653,
4672,
468,
468,
5098,
855,
326,
585,
309,
392,
3330,
2190,
353,
315,
7337,
6,
261,
451,
402,
468,
11234,
777,
854,
16,
1529,
3631,
471,
1338,
855,
518,
3647,
18,
... |
import textwrap | def display(self, add_comments=False): '''Display options in a config file form.''' import textwrap output = StringIO.StringIO() keys = self._options.keys() keys.sort() currentSection = None for sect, opt in keys: if sect != currentSection: if currentSection is not None: output.write('\n') output.write('[') output.write(sect) output.write("]\n") currentSection = sect if add_comments: doc = self._options[sect, opt].doc() if not doc: doc = "No information available, sorry." doc = re.sub(r"\s+", " ", doc) output.write("\n# %s\n" % ("\n# ".join(textwrap.wrap(doc)),)) self._options[sect, opt].write_config(output) return output.getvalue() | 86b6046afea2cb3ae7886681cffdbe2cf51eeb3c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6126/86b6046afea2cb3ae7886681cffdbe2cf51eeb3c/OptionsClass.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2562,
12,
2890,
16,
527,
67,
9231,
33,
8381,
4672,
9163,
4236,
702,
316,
279,
642,
585,
646,
1093,
6309,
876,
273,
15777,
18,
780,
4294,
1435,
1311,
273,
365,
6315,
2116,
18,
2452,
143... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2562,
12,
2890,
16,
527,
67,
9231,
33,
8381,
4672,
9163,
4236,
702,
316,
279,
642,
585,
646,
1093,
6309,
876,
273,
15777,
18,
780,
4294,
1435,
1311,
273,
365,
6315,
2116,
18,
2452,
143... | |
urllib.urlretrieve( "http://gforge.inria.fr/plugins/scmsvn/viewcvs.php/*checkout*/trunk/deploygui/src/openalea/deploygui/auth.py?root=openalea", | urllib.urlretrieve( "http://gforge.inria.fr/scm/viewcvs.php/*checkout*/trunk/deploygui/src/openalea/deploygui/auth.py?root=openalea", | def _validate_md5(egg_name, data): if egg_name in md5_data: digest = md5(data).hexdigest() if digest != md5_data[egg_name]: print >>sys.stderr, ( "md5 validation of %s failed! (Possible download problem?)" % egg_name ) sys.exit(2) return data | 9d2751722052e7a7f806b8c3f54ae122e6e9b38b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4914/9d2751722052e7a7f806b8c3f54ae122e6e9b38b/ez_alea_setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
5662,
67,
1264,
25,
12,
23171,
67,
529,
16,
501,
4672,
309,
25144,
67,
529,
316,
3481,
25,
67,
892,
30,
5403,
273,
3481,
25,
12,
892,
2934,
7118,
10171,
1435,
309,
5403,
480,
34... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
5662,
67,
1264,
25,
12,
23171,
67,
529,
16,
501,
4672,
309,
25144,
67,
529,
316,
3481,
25,
67,
892,
30,
5403,
273,
3481,
25,
12,
892,
2934,
7118,
10171,
1435,
309,
5403,
480,
34... |
print " ".join(texts) | self.stdout.write("%s\n"%str(" ".join(texts))) | def columnize(self, list, displaywidth=80): """Display a list of strings as a compact set of columns. | 5cfb0c082eb272eba0c1782c2d88426ad04a99e0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/5cfb0c082eb272eba0c1782c2d88426ad04a99e0/cmd.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1057,
554,
12,
2890,
16,
666,
16,
2562,
2819,
33,
3672,
4672,
3536,
4236,
279,
666,
434,
2064,
487,
279,
6739,
444,
434,
2168,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1057,
554,
12,
2890,
16,
666,
16,
2562,
2819,
33,
3672,
4672,
3536,
4236,
279,
666,
434,
2064,
487,
279,
6739,
444,
434,
2168,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
def desperate_fixer(g): return ' ' html = re.sub('<.*?>', desperate_fixer, html) | html = re.sub('<.*?>', ' ', html) | def html2plaintext(html, body_id=None, encoding='utf-8'): ## (c) Fry-IT, www.fry-it.com, 2007 ## <peter@fry-it.com> ## download here: http://www.peterbe.com/plog/html2plaintext """ from an HTML text, convert the HTML to plain text. If @body_id is provided then this is the tag where the body (not necessarily <body>) starts. """ try: from BeautifulSoup import BeautifulSoup, SoupStrainer, Comment except: return html urls = [] if body_id is not None: strainer = SoupStrainer(id=body_id) else: strainer = SoupStrainer('body') soup = BeautifulSoup(html, parseOnlyThese=strainer, fromEncoding=encoding) for link in soup.findAll('a'): title = unicode(link) for url in [x[1] for x in link.attrs if x[0]=='href']: urls.append(dict(url=url, tag=unicode(link), title=title)) html = unicode(soup) url_index = [] i = 0 for d in urls: if d['title'] == d['url'] or 'http://'+d['title'] == d['url']: html = html.replace(d['tag'], d['url']) else: i += 1 html = html.replace(d['tag'], '%s [%s]' % (d['title'], i)) url_index.append(d['url']) html = html.replace('<strong>', '*').replace('</strong>', '*') html = html.replace('<b>', '*').replace('</b>', '*') html = html.replace('<h3>', '*').replace('</h3>', '*') html = html.replace('<h2>', '**').replace('</h2>', '**') html = html.replace('<h1>', '**').replace('</h1>', '**') html = html.replace('<em>', '/').replace('</em>', '/') # the only line breaks we respect is those of ending tags and # breaks html = html.replace('\n', ' ') html = html.replace('<br>', '\n') html = html.replace('<tr>', '\n') html = html.replace('</p>', '\n\n') html = re.sub('<br\s*/>', '\n', html) html = html.replace(' ' * 2, ' ') # for all other tags we failed to clean up, just remove then and # complain about them on the stderr def desperate_fixer(g): #print >>sys.stderr, "failed to clean up %s" % str(g.group()) return ' ' html = re.sub('<.*?>', desperate_fixer, html) # lstrip all lines html = '\n'.join([x.lstrip() for x in html.splitlines()]) for i, url in enumerate(url_index): if i == 0: html += '\n\n' html += '[%s] %s\n' % (i+1, url) return html | 58e9a51264038e97e4ac99a3a64ac10d68e4c26a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/58e9a51264038e97e4ac99a3a64ac10d68e4c26a/openerp_mailgate.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1729,
22,
412,
10133,
12,
2620,
16,
1417,
67,
350,
33,
7036,
16,
2688,
2218,
3158,
17,
28,
11,
4672,
7541,
261,
71,
13,
478,
1176,
17,
1285,
16,
8765,
18,
74,
1176,
17,
305,
18,
83... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1729,
22,
412,
10133,
12,
2620,
16,
1417,
67,
350,
33,
7036,
16,
2688,
2218,
3158,
17,
28,
11,
4672,
7541,
261,
71,
13,
478,
1176,
17,
1285,
16,
8765,
18,
74,
1176,
17,
305,
18,
83... |
.selectmany() \ .tolist() | .select_many() \ .to_list() | def selectmany_4(self): return Query(self.W1) \ .select(lambda s: s.split(' ')) \ .selectmany() \ .tolist() | f134b825431ac82f3dbba76e7a945a0e195d27cf /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14315/f134b825431ac82f3dbba76e7a945a0e195d27cf/pyquery.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2027,
9353,
67,
24,
12,
2890,
4672,
327,
2770,
12,
2890,
18,
59,
21,
13,
521,
263,
4025,
12,
14661,
272,
30,
272,
18,
4939,
2668,
296,
3719,
521,
263,
4025,
9353,
1435,
521,
263,
349... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2027,
9353,
67,
24,
12,
2890,
4672,
327,
2770,
12,
2890,
18,
59,
21,
13,
521,
263,
4025,
12,
14661,
272,
30,
272,
18,
4939,
2668,
296,
3719,
521,
263,
4025,
9353,
1435,
521,
263,
349... |
return self.snippet.snippet + (' ' + self.snippet.variant) if self.snippet.variant else '' | return self.snippet.snippet + ((' ' + self.snippet.variant) if self.snippet.variant else '') | def do_get_text(self): return self.snippet.snippet + (' ' + self.snippet.variant) if self.snippet.variant else '' | f2c51d599e9d60e88de102dcba8e4e5ba864a9e3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14800/f2c51d599e9d60e88de102dcba8e4e5ba864a9e3/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
588,
67,
955,
12,
2890,
4672,
327,
365,
18,
25788,
18,
25788,
397,
7707,
296,
397,
365,
18,
25788,
18,
8688,
13,
309,
365,
18,
25788,
18,
8688,
469,
875,
2,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
588,
67,
955,
12,
2890,
4672,
327,
365,
18,
25788,
18,
25788,
397,
7707,
296,
397,
365,
18,
25788,
18,
8688,
13,
309,
365,
18,
25788,
18,
8688,
469,
875,
2,
-100,
-100,
-100... |
u.removeCapability('editfactoids') | try: u.removeCapability('editfactoids') except: irc.error('User %s is not an editor' % name) else: irc.replySuccess() | def removeeditor(self, irc, msg, args, name): self._precheck(irc, msg, capability='addeditors') try: u = ircdb.users.getUser(name) except: irc.error('User %s is not registered' % name) else: u.removeCapability('editfactoids') | b6b1c9c4b641e4f6024353e96ec3ef816fd54091 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3105/b6b1c9c4b641e4f6024353e96ec3ef816fd54091/plugin.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1206,
9177,
12,
2890,
16,
277,
1310,
16,
1234,
16,
833,
16,
508,
4672,
365,
6315,
1484,
1893,
12,
481,
71,
16,
1234,
16,
12593,
2218,
9665,
13704,
6134,
775,
30,
582,
273,
277,
1310,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1206,
9177,
12,
2890,
16,
277,
1310,
16,
1234,
16,
833,
16,
508,
4672,
365,
6315,
1484,
1893,
12,
481,
71,
16,
1234,
16,
12593,
2218,
9665,
13704,
6134,
775,
30,
582,
273,
277,
1310,
... |
checkin, checkout = self.checkin, self.checkout | self.checkin_mutex.acquire() self.waiting = self.waiting + 1 if self.waiting == self.num_threads: self.waiting = self.num_threads - 1 self.checkout_mutex.release() return self.checkin_mutex.release() | def enter(self): checkin, checkout = self.checkin, self.checkout | 12c238c50046dc378380207f15830cf4af619e32 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8125/12c238c50046dc378380207f15830cf4af619e32/test_thread.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6103,
12,
2890,
4672,
866,
267,
16,
13926,
273,
365,
18,
1893,
267,
16,
365,
18,
17300,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6103,
12,
2890,
4672,
866,
267,
16,
13926,
273,
365,
18,
1893,
267,
16,
365,
18,
17300,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
pass | if self.completekey: try: import readline readline.set_completer(self.old_completer) except ImportError: pass | def postloop(self): """Hook method executed once when the cmdloop() method is about to return. | 54d798a04cfa882fd8593592d8a63bba2cf11dc5 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/54d798a04cfa882fd8593592d8a63bba2cf11dc5/cmd.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1603,
6498,
12,
2890,
4672,
3536,
5394,
707,
7120,
3647,
1347,
326,
1797,
6498,
1435,
707,
353,
2973,
358,
327,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1603,
6498,
12,
2890,
4672,
3536,
5394,
707,
7120,
3647,
1347,
326,
1797,
6498,
1435,
707,
353,
2973,
358,
327,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
self._agent = qmf.Agent(self) | self._agent = qmf.Agent(self, "agent_test_label") | def main(self): ''' Agent application's main processing loop. ''' # Connect to the broker self._settings = qmf.ConnectionSettings() self._settings.sendUserId = True if len(sys.argv) > 1: self._settings.host = str(sys.argv[1]) if len(sys.argv) > 2: self._settings.port = int(sys.argv[2]) self._connection = qmf.Connection(self._settings) | 48fdca8b89f5e3babf7c90335e84cef4d0644a58 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/198/48fdca8b89f5e3babf7c90335e84cef4d0644a58/python_agent.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
2890,
4672,
9163,
8669,
2521,
1807,
2774,
4929,
2798,
18,
9163,
468,
8289,
358,
326,
8625,
365,
6315,
4272,
273,
1043,
16126,
18,
1952,
2628,
1435,
365,
6315,
4272,
18,
4661,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
2890,
4672,
9163,
8669,
2521,
1807,
2774,
4929,
2798,
18,
9163,
468,
8289,
358,
326,
8625,
365,
6315,
4272,
273,
1043,
16126,
18,
1952,
2628,
1435,
365,
6315,
4272,
18,
4661,
1... |
label = gtk.Label("\n".join(["M" * 36] * 3)) | label = gtk.Label("\n".join(["M" * 34] * 3)) | def _init_sizes(self): """Initialize the widget sizes.""" | 20b3d053861fa68852ef6d9ac1afa2ae8e9e3c39 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12852/20b3d053861fa68852ef6d9ac1afa2ae8e9e3c39/posadjust.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2738,
67,
11914,
12,
2890,
4672,
3536,
7520,
326,
3604,
8453,
12123,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2738,
67,
11914,
12,
2890,
4672,
3536,
7520,
326,
3604,
8453,
12123,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
items = map(lambda (k, v), type = type, simp = SIMPLE_TYPES, indent = indent: (k, v, not type(v) in simp, indent), items) | items = [(k, v, not isinstance(v, SIMPLE_TYPES), indent) for k, v in items] | def pack_items(items, indent = 0): items = map(lambda (k, v), type = type, simp = SIMPLE_TYPES, indent = indent: (k, v, not type(v) in simp, indent), items) return tuple_caselesssort(items) | 8b1e775b54a15ebaaa7e03a0d0e105c7f635388d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/8b1e775b54a15ebaaa7e03a0d0e105c7f635388d/PyBrowser.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2298,
67,
3319,
12,
3319,
16,
3504,
273,
374,
4672,
1516,
273,
306,
12,
79,
16,
331,
16,
486,
1549,
12,
90,
16,
19098,
67,
10564,
3631,
3504,
13,
364,
417,
16,
331,
316,
1516,
65,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2298,
67,
3319,
12,
3319,
16,
3504,
273,
374,
4672,
1516,
273,
306,
12,
79,
16,
331,
16,
486,
1549,
12,
90,
16,
19098,
67,
10564,
3631,
3504,
13,
364,
417,
16,
331,
316,
1516,
65,
... |
md_fh.write(json.write(_sanitize_dbus_dict(entry))+'\n') | var = json.write(_sanitize_dbus_dict(entry))+'\n' md_fh.write(var.encode('utf-8')) | def write_metadata(ds_path): # setup datastore connection bus = dbus.SessionBus() obj = bus.get_object(DS_DBUS_SERVICE, DS_DBUS_PATH) datastore = dbus.Interface(obj, DS_DBUS_INTERFACE) # name the backup file # and open a tmpfile in the same dir # to ensure an atomic replace md_path = os.path.join(ds_path, 'metadata.json') (md_fd, md_tmppath) = tempfile.mkstemp(suffix='.json', prefix='.metadata-', dir=ds_path) md_fh = os.fdopen(md_fd, 'w') # preview contains binary data we # don't actually want... drop_properties = ['preview'] query = {} entries, count = datastore.find(query, [], byte_arrays=True) print 'Writing metadata and file indexes for %d entries.' % len(entries) for entry in entries: for prop in drop_properties: if prop in entry: del entry[prop] md_fh.write(json.write(_sanitize_dbus_dict(entry))+'\n') md_fh.close() os.rename(md_tmppath, md_path) cleanup_stale_metadata(ds_path) return md_path | 30b59aa85950b6f894ebe11a5ada0f788709f5c7 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9380/30b59aa85950b6f894ebe11a5ada0f788709f5c7/ds_backup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
67,
4165,
12,
2377,
67,
803,
4672,
225,
468,
3875,
9290,
1459,
5766,
273,
21866,
18,
2157,
7086,
1435,
1081,
273,
5766,
18,
588,
67,
1612,
12,
3948,
67,
2290,
3378,
67,
12426,
16... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
67,
4165,
12,
2377,
67,
803,
4672,
225,
468,
3875,
9290,
1459,
5766,
273,
21866,
18,
2157,
7086,
1435,
1081,
273,
5766,
18,
588,
67,
1612,
12,
3948,
67,
2290,
3378,
67,
12426,
16... |
self.wait_visibility() self.grab_set() | def __init__(self, parent, title = None): | 3c4bad7c8b322ab646d6971dae2a208eee8eeadd /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12029/3c4bad7c8b322ab646d6971dae2a208eee8eeadd/tkSimpleDialog.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
982,
16,
2077,
273,
599,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
982,
16,
2077,
273,
599,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... | |
Inf = Decimal('Inf') negInf = Decimal('-Inf') NaN = Decimal('NaN') Dec_0 = Decimal(0) Dec_p1 = Decimal(1) Dec_n1 = Decimal(-1) Infsign = (Inf, negInf) | _Inf = Decimal('Inf') _negInf = Decimal('-Inf') _NaN = Decimal('NaN') _Dec_0 = Decimal(0) _Dec_p1 = Decimal(1) _Dec_n1 = Decimal(-1) _Infsign = (_Inf, _negInf) | def _format_align(body, spec_dict): """Given an unpadded, non-aligned numeric string, add padding and aligment to conform with the given format specifier dictionary (as output from parse_format_specifier). It's assumed that if body is negative then it starts with '-'. Any leading sign ('-' or '+') is stripped from the body before applying the alignment and padding rules, and replaced in the appropriate position. """ # figure out the sign; we only examine the first character, so if # body has leading whitespace the results may be surprising. if len(body) > 0 and body[0] in '-+': sign = body[0] body = body[1:] else: sign = '' if sign != '-': if spec_dict['sign'] in ' +': sign = spec_dict['sign'] else: sign = '' # how much extra space do we have to play with? minimumwidth = spec_dict['minimumwidth'] fill = spec_dict['fill'] padding = fill*(max(minimumwidth - (len(sign+body)), 0)) align = spec_dict['align'] if align == '<': result = padding + sign + body elif align == '>': result = sign + body + padding elif align == '=': result = sign + padding + body else: #align == '^' half = len(padding)//2 result = padding[:half] + sign + body + padding[half:] # make sure that result is unicode if necessary if spec_dict['unicode']: result = unicode(result) return result | c5de0969ca2a82c66efd30bb675f03c2324a3b27 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8546/c5de0969ca2a82c66efd30bb675f03c2324a3b27/decimal.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2139,
67,
7989,
12,
3432,
16,
857,
67,
1576,
4672,
3536,
6083,
392,
640,
6982,
785,
16,
1661,
17,
20677,
6389,
533,
16,
527,
4992,
471,
524,
360,
475,
358,
20156,
598,
326,
864,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2139,
67,
7989,
12,
3432,
16,
857,
67,
1576,
4672,
3536,
6083,
392,
640,
6982,
785,
16,
1661,
17,
20677,
6389,
533,
16,
527,
4992,
471,
524,
360,
475,
358,
20156,
598,
326,
864,
... |
if len(member.params) <= 3: mayTrace = True | def addStubMember(memberId, member, traceable): # Check that the member is ok. mayTrace = False if member.kind not in ('method', 'attribute'): raise UserError("Member %s is %r, not a method or attribute." % (memberId, member.kind)) if member.noscript: raise UserError("%s %s is noscript." % (member.kind.capitalize(), memberId)) if member.notxpcom: raise UserError( "%s %s: notxpcom methods are not supported." % (member.kind.capitalize(), memberId)) if (member.kind == 'attribute' and not member.readonly and isSpecificInterfaceType(member.realtype, 'nsIVariant')): raise UserError( "Attribute %s: Non-readonly attributes of type nsIVariant " "are not supported." % memberId) # Check for unknown properties. for attrname, value in vars(member).items(): if value is True and attrname not in ('readonly',): raise UserError("%s %s: unrecognized property %r" % (member.kind.capitalize(), memberId, attrname)) if member.kind == 'method': for param in member.params: for attrname, value in vars(param).items(): if value is True and attrname not in ('optional',): raise UserError("Method %s, parameter %s: " "unrecognized property %r" % (memberId, param.name, attrname)) if len(member.params) <= 3: mayTrace = True member.traceable = traceable and mayTrace # Add this member to the list. member.iface.stubMembers.append(member) | 547c9137fe7f5061409005e05ec145bbf18f5ab6 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11102/547c9137fe7f5061409005e05ec145bbf18f5ab6/qsgen.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
527,
11974,
4419,
12,
5990,
548,
16,
3140,
16,
2606,
429,
4672,
468,
2073,
716,
326,
3140,
353,
1529,
18,
2026,
3448,
273,
1083,
309,
3140,
18,
9224,
486,
316,
7707,
2039,
2187,
296,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
527,
11974,
4419,
12,
5990,
548,
16,
3140,
16,
2606,
429,
4672,
468,
2073,
716,
326,
3140,
353,
1529,
18,
2026,
3448,
273,
1083,
309,
3140,
18,
9224,
486,
316,
7707,
2039,
2187,
296,
4... | |
self.assertEqual(p.stdout.read(), | self.assertEqual(p.communicate()[0], | def test_envvar_and_command_line(self): newenv = os.environ.copy() newenv["PYTHONWARNINGS"] = "ignore::DeprecationWarning" p = subprocess.Popen([sys.executable, "-W" "ignore::UnicodeWarning", "-c", "import sys; sys.stdout.write(str(sys.warnoptions))"], stdout=subprocess.PIPE, env=newenv) self.assertEqual(p.stdout.read(), "['ignore::UnicodeWarning', 'ignore::DeprecationWarning']") | 3ef293ada8c4f36295d1aeba2899a2ef7d251064 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12029/3ef293ada8c4f36295d1aeba2899a2ef7d251064/test_warnings.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
3074,
1401,
67,
464,
67,
3076,
67,
1369,
12,
2890,
4672,
394,
3074,
273,
1140,
18,
28684,
18,
3530,
1435,
394,
3074,
9614,
16235,
20131,
9511,
55,
11929,
273,
315,
6185,
2866,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
3074,
1401,
67,
464,
67,
3076,
67,
1369,
12,
2890,
4672,
394,
3074,
273,
1140,
18,
28684,
18,
3530,
1435,
394,
3074,
9614,
16235,
20131,
9511,
55,
11929,
273,
315,
6185,
2866,
... |
string = string.replace('<p>', '') string = string.replace('</p>', '') string = string.replace('\n', '') string = re.sub('> ', '>', string) | replacements = {'<p>': '', '</p>': '', '\n': '', '> ': '>'} for src, dst in replacements.items(): string = string.replace(src, dst) | def rst_write(self, string): if self.strip_paragraph: string = string.replace('<p>', '') string = string.replace('</p>', '') string = string.replace('\n', '') string = re.sub('> ', '>', string) # Everything seems to have a space ending the text block. We want to # get rid of this if len(string) and string[-1] == ' ': string = string[:-1] self.body.append(string) | c9655705a93f1df5d7b80440a5e3fd18916989bc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5620/c9655705a93f1df5d7b80440a5e3fd18916989bc/rst.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13978,
67,
2626,
12,
2890,
16,
533,
4672,
309,
365,
18,
6406,
67,
22445,
30,
11413,
273,
13666,
32,
84,
1870,
30,
10226,
4357,
84,
1870,
30,
10226,
2337,
82,
4278,
10226,
11279,
4290,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13978,
67,
2626,
12,
2890,
16,
533,
4672,
309,
365,
18,
6406,
67,
22445,
30,
11413,
273,
13666,
32,
84,
1870,
30,
10226,
4357,
84,
1870,
30,
10226,
2337,
82,
4278,
10226,
11279,
4290,
... |
self.tools_path = self.tools_path.rstrip('\\') | self.tools_path = self.tools_path.rstrip('\\') os.environ['PATH'] = os.environ['PATH'] + ';' + self.tools_path | def SetUp(self, binaries): """Set up the platform specific environment and instrument the binaries for coverage. | 1581ad170e9c9d3a92c66b377c35a5abc747dc4c /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9392/1581ad170e9c9d3a92c66b377c35a5abc747dc4c/coverage.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1000,
1211,
12,
2890,
16,
4158,
5646,
4672,
3536,
694,
731,
326,
4072,
2923,
3330,
471,
10353,
326,
4158,
5646,
364,
11196,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1000,
1211,
12,
2890,
16,
4158,
5646,
4672,
3536,
694,
731,
326,
4072,
2923,
3330,
471,
10353,
326,
4158,
5646,
364,
11196,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
security.declareProtected(CMFCorePermissions.View, 'CreationDate') | security.declareProtected(permissions.View, 'CreationDate') | def languages(self): """Vocabulary method for the language field """ # XXX document me # use a list of languages from PLT? available_langs = getattr(self, 'availableLanguages', None) if available_langs is None: return DisplayList( (('en','English'), ('fr','French'), ('es','Spanish'), ('pt','Portuguese'), ('ru','Russian'))) if callable(available_langs): available_langs = available_langs() return DisplayList(available_langs) | bffbf06ccb0e344dcf82f7addf15838da46c2ba5 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12165/bffbf06ccb0e344dcf82f7addf15838da46c2ba5/ExtensibleMetadata.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8191,
12,
2890,
4672,
3536,
30747,
707,
364,
326,
2653,
652,
3536,
468,
11329,
1668,
1791,
468,
999,
279,
666,
434,
8191,
628,
453,
12050,
35,
2319,
67,
4936,
87,
273,
3869,
12,
2890,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8191,
12,
2890,
4672,
3536,
30747,
707,
364,
326,
2653,
652,
3536,
468,
11329,
1668,
1791,
468,
999,
279,
666,
434,
8191,
628,
453,
12050,
35,
2319,
67,
4936,
87,
273,
3869,
12,
2890,
... |
if release: | if release_flag: | def fail(out): status, stdout = out if status != 0: print "#" * 24 print stdout print "#" * 24 clean() sys.exit() return out | 0f68352daacce38707c3c193d134c6ede97f8784 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10995/0f68352daacce38707c3c193d134c6ede97f8784/quodlibet.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2321,
12,
659,
4672,
1267,
16,
3909,
273,
596,
309,
1267,
480,
374,
30,
1172,
13092,
380,
4248,
1172,
3909,
1172,
13092,
380,
4248,
2721,
1435,
2589,
18,
8593,
1435,
327,
596,
225,
2,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2321,
12,
659,
4672,
1267,
16,
3909,
273,
596,
309,
1267,
480,
374,
30,
1172,
13092,
380,
4248,
1172,
3909,
1172,
13092,
380,
4248,
2721,
1435,
2589,
18,
8593,
1435,
327,
596,
225,
2,
... |
if table.domain.classVar.varType == orange.VarTypes.Discrete: | if testTable.domain.classVar.varType == orange.VarTypes.Discrete: | def kNNComputeAccuracy(self, table): # select a subset of the data if necessary percentDataUsed = int(str(self.percentDataUsedCombo.currentText())) if percentDataUsed != 100: indices = orange.MakeRandomIndices2(table, 1.0-float(percentDataUsed)/100.0) testTable = table.select(indices) else: testTable = table #qApp.processEvents() # allow processing of other events | 1a26c0137e0f96ec4a8edbab5d6eb19b08c69a1f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6366/1a26c0137e0f96ec4a8edbab5d6eb19b08c69a1f/OWkNNOptimization.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
417,
12225,
7018,
37,
10988,
12,
2890,
16,
1014,
4672,
468,
2027,
279,
7931,
434,
326,
501,
309,
4573,
5551,
751,
6668,
273,
509,
12,
701,
12,
2890,
18,
8849,
751,
6668,
16156,
18,
297... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
417,
12225,
7018,
37,
10988,
12,
2890,
16,
1014,
4672,
468,
2027,
279,
7931,
434,
326,
501,
309,
4573,
5551,
751,
6668,
273,
509,
12,
701,
12,
2890,
18,
8849,
751,
6668,
16156,
18,
297... |
radius == err | radius = err radius *= (1.1/0.77) * 1.0 | dispdef = molecule_dispdef(self.basemol, self.o) | 831de45168607289ff2c0e720c2e1f831471f13b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/831de45168607289ff2c0e720c2e1f831471f13b/extrudeMode.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
16232,
536,
273,
13661,
67,
25453,
536,
12,
2890,
18,
1969,
21260,
16,
365,
18,
83,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
16232,
536,
273,
13661,
67,
25453,
536,
12,
2890,
18,
1969,
21260,
16,
365,
18,
83,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
def test_debug_undefined(): | def test_debug_undefined(self): | def test_debug_undefined(): env = Environment(undefined=DebugUndefined) self.assert_equal(env.from_string('{{ missing }}').render(), '{{ missing }}') self.assert_raises(UndefinedError, env.from_string('{{ missing.attribute }}').render()) self.assert_equal(env.from_string('{{ missing|list }}').render(), '[]') u'[]' self.assert_equal(env.from_string('{{ missing is not defined }}').render, 'True') self.assert_equal(env.from_string('{{ foo.missing }}').render(foo=42), u"{{ no such element: int['missing'] }}") self.assert_equal(env.from_string('{{ not missing }}').render(), 'True') | 82a2225ac939318e7d812cb4ae537be9fb6362cf /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14446/82a2225ac939318e7d812cb4ae537be9fb6362cf/api.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
4148,
67,
5978,
12,
2890,
4672,
1550,
273,
7518,
12,
5978,
33,
2829,
10317,
13,
365,
18,
11231,
67,
9729,
12,
3074,
18,
2080,
67,
1080,
2668,
12187,
3315,
29893,
2934,
5902,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
4148,
67,
5978,
12,
2890,
4672,
1550,
273,
7518,
12,
5978,
33,
2829,
10317,
13,
365,
18,
11231,
67,
9729,
12,
3074,
18,
2080,
67,
1080,
2668,
12187,
3315,
29893,
2934,
5902,
... |
gain = float(l[1]) | gain = 1.333 * float(l[1]) | def gain(self): '''Estimate gain for this image.''' | f0994dc59eebf98dd27a811673a91c82c22df132 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/3290/f0994dc59eebf98dd27a811673a91c82c22df132/Diffdump.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
17527,
12,
2890,
4672,
9163,
13638,
17527,
364,
333,
1316,
1093,
6309,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
17527,
12,
2890,
4672,
9163,
13638,
17527,
364,
333,
1316,
1093,
6309,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
def sendto(self, data, addr, flags=0): | def sendto(self, *args): | def sendto(self, data, addr, flags=0): if self._sslobj: raise ValueError("sendto not allowed on instances of %s" % self.__class__) else: return socket.sendto(self, data, addr, flags) | 13aa09070ef42915521c73d7f0f5217129e42c60 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10423/13aa09070ef42915521c73d7f0f5217129e42c60/ssl.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1366,
869,
12,
2890,
16,
380,
1968,
4672,
309,
365,
6315,
1049,
383,
441,
30,
1002,
2068,
2932,
4661,
869,
486,
2935,
603,
3884,
434,
738,
87,
6,
738,
365,
16186,
1106,
972,
13,
469,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1366,
869,
12,
2890,
16,
380,
1968,
4672,
309,
365,
6315,
1049,
383,
441,
30,
1002,
2068,
2932,
4661,
869,
486,
2935,
603,
3884,
434,
738,
87,
6,
738,
365,
16186,
1106,
972,
13,
469,
... |
print content | def sychronize_with_vcs(self, vcs_sync_url): """ Synchronizes the Mantis issues and the VCS history. """ try: file_object = urllib2.urlopen(vcs_sync_url) content = file_object.read() print content self.logger.debug(content) file_object.close() except IOError, error: self.logger.error("Cannot open URL. Reason: '%s'" % error.message) | dc88ab05d63fd02c59a29b9c58e61119fc3d5b9c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5396/dc88ab05d63fd02c59a29b9c58e61119fc3d5b9c/mantis.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1393,
29115,
554,
67,
1918,
67,
18982,
12,
2890,
16,
20255,
67,
8389,
67,
718,
4672,
3536,
26535,
3128,
326,
490,
970,
291,
8296,
471,
326,
776,
4596,
4927,
18,
3536,
225,
775,
30,
585... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1393,
29115,
554,
67,
1918,
67,
18982,
12,
2890,
16,
20255,
67,
8389,
67,
718,
4672,
3536,
26535,
3128,
326,
490,
970,
291,
8296,
471,
326,
776,
4596,
4927,
18,
3536,
225,
775,
30,
585... | |
finally: cr.commit() cr.close() | def _poolJobs(self, db_name, check=False): try: db, pool = pooler.get_db_and_pool(db_name) except: return False if pool._init: # retry in a few minutes next_call = 600 else: now = DateTime.now() try: cr = db.cursor() cr.execute('select * from ir_cron where numbercall<>0 and active and nextcall<=now() order by priority') for job in cr.dictfetchall(): nextcall = DateTime.strptime(job['nextcall'], '%Y-%m-%d %H:%M:%S') numbercall = job['numbercall'] ok = False while nextcall < now and numbercall: if numbercall > 0: numbercall -= 1 if not ok or job['doall']: self._callback(cr, job['user_id'], job['model'], job['function'], job['args']) if numbercall: nextcall += _intervalTypes[job['interval_type']](job['interval_number']) ok = True addsql='' if not numbercall: addsql = ', active=False' cr.execute("update ir_cron set nextcall=%s, numbercall=%s"+addsql+" where id=%s", (nextcall.strftime('%Y-%m-%d %H:%M:%S'), numbercall, job['id'])) cr.commit() finally: cr.commit() cr.close() | 1b607549a7281cba356e71492737284ff9da15bb /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7397/1b607549a7281cba356e71492737284ff9da15bb/ir_cron.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
6011,
7276,
12,
2890,
16,
1319,
67,
529,
16,
866,
33,
8381,
4672,
775,
30,
1319,
16,
2845,
273,
2845,
264,
18,
588,
67,
1966,
67,
464,
67,
6011,
12,
1966,
67,
529,
13,
1335,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
6011,
7276,
12,
2890,
16,
1319,
67,
529,
16,
866,
33,
8381,
4672,
775,
30,
1319,
16,
2845,
273,
2845,
264,
18,
588,
67,
1966,
67,
464,
67,
6011,
12,
1966,
67,
529,
13,
1335,
3... | |
if not hasattr(plist, "CFBundleIdentifier"): | if self.bundle_id: plist.CFBundleIdentifier = self.bundle_id elif not hasattr(plist, "CFBundleIdentifier"): | def setup(self): # XXX rethink self.name munging, this is brittle. self.name, ext = os.path.splitext(self.name) if not ext: ext = ".bundle" bundleextension = ext # misc (derived) attributes self.bundlepath = pathjoin(self.builddir, self.name + bundleextension) | 9e74465ef8769cd60189563e76048b0b27365c67 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/97/9e74465ef8769cd60189563e76048b0b27365c67/bundlebuilder.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
12,
2890,
4672,
468,
11329,
19911,
754,
365,
18,
529,
312,
318,
1998,
16,
333,
353,
324,
583,
5929,
18,
365,
18,
529,
16,
1110,
273,
1140,
18,
803,
18,
4939,
408,
12,
2890,
18,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
12,
2890,
4672,
468,
11329,
19911,
754,
365,
18,
529,
312,
318,
1998,
16,
333,
353,
324,
583,
5929,
18,
365,
18,
529,
16,
1110,
273,
1140,
18,
803,
18,
4939,
408,
12,
2890,
18,... |
tout = '' if test.timeout.secs > 0: | if test.timeout.secs < 0: tout = '' else: | def _write_test_metadata(self, test): self._start_suite_or_test_metadata(test) tout = '' if test.timeout.secs > 0: tout = utils.secs_to_timestr(test.timeout.secs) if test.timeout.message: tout += ' | ' + test.timeout.message self._write_metadata_row('Timeout', tout) self._write_metadata_row('Tags', ', '.join(test.tags)) self._writer.end_element('table') | ef62a52c35d76a494de63689fb20166fc91ca595 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7408/ef62a52c35d76a494de63689fb20166fc91ca595/testdoc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2626,
67,
3813,
67,
4165,
12,
2890,
16,
1842,
4672,
365,
6315,
1937,
67,
30676,
67,
280,
67,
3813,
67,
4165,
12,
3813,
13,
309,
1842,
18,
4538,
18,
21317,
411,
374,
30,
358,
322... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2626,
67,
3813,
67,
4165,
12,
2890,
16,
1842,
4672,
365,
6315,
1937,
67,
30676,
67,
280,
67,
3813,
67,
4165,
12,
3813,
13,
309,
1842,
18,
4538,
18,
21317,
411,
374,
30,
358,
322... |
if summary: statuses = ({path: statuses[path]}, check_summary) if callback: callback(path, statuses) | if statuses: if summary: statuses = ({path: statuses[path]}, check_summary) if callback: callback(path, statuses) | def _update_path_status(self, path, recurse=False, invalidate=False, summary=False, callback=None): """ Update the cached information for the given path, notifying the callback upon completion. This function will check the cache first, just in case a previous call has populated the path in question and we are not invalidating. The parameters are as per check_status, but instead of a return type there is the callback. """ statuses = {} | a90225b00e8fae8335109d10a6784f3a4d112ddb /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5119/a90225b00e8fae8335109d10a6784f3a4d112ddb/statuscache.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2725,
67,
803,
67,
2327,
12,
2890,
16,
589,
16,
11502,
33,
8381,
16,
11587,
33,
8381,
16,
4916,
33,
8381,
16,
1348,
33,
7036,
4672,
3536,
2315,
326,
3472,
1779,
364,
326,
864,
5... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2725,
67,
803,
67,
2327,
12,
2890,
16,
589,
16,
11502,
33,
8381,
16,
11587,
33,
8381,
16,
4916,
33,
8381,
16,
1348,
33,
7036,
4672,
3536,
2315,
326,
3472,
1779,
364,
326,
864,
5... |
vList.append( Violation(self, line.number) ) | vList.append( Violation(self, line) ) | def apply(self, lines): vList = [] if (self.getFiletype() in self.getTypeList()): for line in lines: if ( re.search("^\s*using", line.stripped) and not re.search("\s*std\s*;\s*$", line.stripped) ): vList.append( Violation(self, line.number) ) return vList | cb6d1c1b9da3a15f55b42e7ffc55f274aee325d2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6529/cb6d1c1b9da3a15f55b42e7ffc55f274aee325d2/style.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2230,
12,
2890,
16,
2362,
4672,
331,
682,
273,
5378,
309,
261,
2890,
18,
588,
25579,
5872,
1435,
316,
365,
18,
588,
559,
682,
1435,
4672,
364,
980,
316,
2362,
30,
309,
261,
283,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2230,
12,
2890,
16,
2362,
4672,
331,
682,
273,
5378,
309,
261,
2890,
18,
588,
25579,
5872,
1435,
316,
365,
18,
588,
559,
682,
1435,
4672,
364,
980,
316,
2362,
30,
309,
261,
283,
18,
... |
print _(" Sync Successful. Repo [ %s ] now has a total of [ %s ] packages" % (self.options.id, pkg_count)) | print _(" Sync Successful. Repo [ %s ] now has a total of [ %s ] packages") % \ (self.options.id, pkg_count) | def _sync(self): if not self.options.id: print("repo id required. Try --help") sys.exit(0) try: task_object = self.pconn.sync(self.options.id, self.options.timeout) state = "waiting" print "Task created with ID::", task_object['id'] while state not in ["finished", "error", 'timed out', 'canceled']: time.sleep(5) status = self.pconn.sync_status(task_object['status_path']) if status is None: raise SyncError(_('No sync for repository [%s] found') % self.options.id) state = status['state'] print "Sync Status::", state packages = self.pconn.packages(self.options.id) pkg_count = 0 if packages: pkg_count = len(packages) if state == "error": raise SyncError(status['traceback'][-1]) else: print _(" Sync Successful. Repo [ %s ] now has a total of [ %s ] packages" % (self.options.id, pkg_count)) except RestlibException, re: log.info("REST Error.", exc_info=True) systemExit(re.code, re.msg) except SyncError, se: log.info("Sync Error: ", exc_info=True) systemExit("Error : %s" % se) except Exception, e: log.error("General Error: %s" % e) raise | 3031d7f1572d6916f4f024e22264c3de668d641f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10692/3031d7f1572d6916f4f024e22264c3de668d641f/core_repo.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8389,
12,
2890,
4672,
309,
486,
365,
18,
2116,
18,
350,
30,
1172,
2932,
7422,
612,
1931,
18,
6161,
1493,
5201,
7923,
2589,
18,
8593,
12,
20,
13,
775,
30,
1562,
67,
1612,
273,
36... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8389,
12,
2890,
4672,
309,
486,
365,
18,
2116,
18,
350,
30,
1172,
2932,
7422,
612,
1931,
18,
6161,
1493,
5201,
7923,
2589,
18,
8593,
12,
20,
13,
775,
30,
1562,
67,
1612,
273,
36... |
"autotest_lib.server.site_job", "get_site_job_data", | "autotest_lib.server.site_server_job", "get_site_job_data", | def _get_site_job_data_dummy(job): return {} | aed44eb7eef1ffa3a2e5b76bcb3347198861a224 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12268/aed44eb7eef1ffa3a2e5b76bcb3347198861a224/server_job.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
4256,
67,
4688,
67,
892,
67,
21050,
12,
4688,
4672,
327,
2618,
282,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
4256,
67,
4688,
67,
892,
67,
21050,
12,
4688,
4672,
327,
2618,
282,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
SimpleDialog(_('Error setting the new engine, please check your engine settings.'), _('Warning'), 'warning').run() | SimpleDialog(_('Error setting the new engine; please check your engine settings.'), _('Warning'), 'warning').run() | def setEngine(self): try: engine = self.engine_manager.newEngine(self.name_entry.get_text(), self.engine_path_entry.get_text(), self.arguments_entry.get_text(), self.image_format_entry.get_text(), self.failure_string_entry.get_text() ) if self.engine: self.engine_manager.replaceEngine(self.engine, engine) self.engine = engine else: self.engine_manager.addNewEngine(engine) return True except: SimpleDialog(_('Error setting the new engine, please check your engine settings.'), _('Warning'), 'warning').run() print sys.exc_info() return False | 1baf639b7c0b78357342a021c71b70a4eb7e2399 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3764/1baf639b7c0b78357342a021c71b70a4eb7e2399/widgetPresenter.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
4410,
12,
2890,
4672,
775,
30,
4073,
273,
365,
18,
8944,
67,
4181,
18,
2704,
4410,
12,
2890,
18,
529,
67,
4099,
18,
588,
67,
955,
9334,
365,
18,
8944,
67,
803,
67,
4099,
18,
5... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
4410,
12,
2890,
4672,
775,
30,
4073,
273,
365,
18,
8944,
67,
4181,
18,
2704,
4410,
12,
2890,
18,
529,
67,
4099,
18,
588,
67,
955,
9334,
365,
18,
8944,
67,
803,
67,
4099,
18,
5... |
print v | def loadrc(): """Use the shlex module to make a simple lexer for the settings, it also attempts to convert any integers to Python ints, otherwise leaves them as strings and handles hopefully all the sane ways of representing a boolean.""" if len(sys.argv) > 2: path = sys.argv[2] else: path = os.path.expanduser( '~/.bpythonrc' ) if not os.path.isfile( path ): return f = open( path ) parser = shlex.shlex( f ) bools = { 'true': True, 'yes': True, 'on': True, 'false': False, 'no': False, 'off': False } config = {} while True: k = parser.get_token() v = None if not k: break k = k.lower() if parser.get_token() == '=': v = parser.get_token() or None print v if v is not None: try: v = int(v) except ValueError: if v.lower() in bools: v = bools[v.lower()] config[k] = v f.close() for k, v in config.iteritems(): if hasattr( OPTS, k ): setattr( OPTS, k, v ) if not "pyparsing" in sys.modules: OPTS.arg_spec = False | 65f5a5ee359b6ec0ea1579c114276cf69961c80d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/6932/65f5a5ee359b6ec0ea1579c114276cf69961c80d/bpython.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
1310,
13332,
3536,
3727,
326,
27180,
1605,
358,
1221,
279,
4143,
8512,
364,
326,
1947,
16,
518,
2546,
7531,
358,
1765,
1281,
12321,
358,
6600,
15542,
16,
3541,
15559,
2182,
487,
2064... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
1310,
13332,
3536,
3727,
326,
27180,
1605,
358,
1221,
279,
4143,
8512,
364,
326,
1947,
16,
518,
2546,
7531,
358,
1765,
1281,
12321,
358,
6600,
15542,
16,
3541,
15559,
2182,
487,
2064... | |
except: | except TypeError: | def span(self, gens, base_ring=None, check=True, already_echelonized=False): """ Return the K-span of the given list of gens, where K is the base field of self. Note that this span is a subspace of the ambient vector space, but need not be a subspace of self. | 6875d7c71609de038c71080542943969764ed340 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/6875d7c71609de038c71080542943969764ed340/free_module.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4548,
12,
2890,
16,
314,
773,
16,
1026,
67,
8022,
33,
7036,
16,
866,
33,
5510,
16,
1818,
67,
8987,
292,
265,
1235,
33,
8381,
4672,
3536,
2000,
326,
1475,
17,
3969,
434,
326,
864,
666... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4548,
12,
2890,
16,
314,
773,
16,
1026,
67,
8022,
33,
7036,
16,
866,
33,
5510,
16,
1818,
67,
8987,
292,
265,
1235,
33,
8381,
4672,
3536,
2000,
326,
1475,
17,
3969,
434,
326,
864,
666... |
self._ClassLoader__classes = {} cdoc = self.fetch_document('<list-ops/>') cde = cdoc.documentElement for cn in cde.childNodes: if cn.nodeName == 'op': classname = str(os.path.join(cn.getAttribute('path'), cn.getAttribute('name'))) for vn in cn.childNodes: if vn.nodeName == 'version': version = int(vn.getAttribute('id')) c = self._ClassLoader__classes.setdefault(classname, {'versions':[], 'imports':{}}) c['versions'].append(version) for c in self._ClassLoader__classes.values(): c["versions"].sort() for k, v in defaultVersions.items(): if k in self._ClassLoader__classes and v: self.setDefaultVersion(k, v) | self._ClassLoader__classes = {} cdoc = self.fetch_document('<list-ops/>') cde = cdoc.documentElement for cn in cde.childNodes: if cn.nodeName == 'op': classname = str(os.path.join(cn.getAttribute('path'), cn.getAttribute('name'))) for vn in cn.childNodes: if vn.nodeName == 'version': version = int(vn.getAttribute('id')) c = self._ClassLoader__classes.setdefault(classname, {'versions':[], 'imports':{}}) c['versions'].append(version) for c in self._ClassLoader__classes.values(): c["versions"].sort() for k, v in defaultVersions.items(): if k in self._ClassLoader__classes and v: self.setDefaultVersion(k, v) | def refresh(self): | 00fa4caca3741bca4c0129528b7cd2fe2d2376e0 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9042/00fa4caca3741bca4c0129528b7cd2fe2d2376e0/RemoteOpLoader.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4460,
12,
2890,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4460,
12,
2890,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
control.loop(restart, fromFrame, toFrame) | if toFrame == None: control.loop(restart, fromFrame, control.getNumFrames() - 1) else: control.loop(restart, fromFrame, toFrame) | def loop(self, animName, restart=1, partName=None, fromFrame=None, toFrame=None): """loop(self, string, int=1, string=None) Loop the given animation on the given part of the actor, restarting at zero frame if requested. If no part name is given then try to loop on all parts. NOTE: loops on all LOD's""" if fromFrame == None: for control in self.getAnimControls(animName, partName): control.loop(restart) else: for control in self.getAnimControls(animName, partName): control.loop(restart, fromFrame, toFrame) | b1ae01f7db05d7da5e711e0a0a11b8866f240e95 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8543/b1ae01f7db05d7da5e711e0a0a11b8866f240e95/Actor.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2798,
12,
2890,
16,
10536,
461,
16,
7870,
33,
21,
16,
1087,
461,
33,
7036,
16,
628,
3219,
33,
7036,
16,
358,
3219,
33,
7036,
4672,
3536,
6498,
12,
2890,
16,
533,
16,
509,
33,
21,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2798,
12,
2890,
16,
10536,
461,
16,
7870,
33,
21,
16,
1087,
461,
33,
7036,
16,
628,
3219,
33,
7036,
16,
358,
3219,
33,
7036,
4672,
3536,
6498,
12,
2890,
16,
533,
16,
509,
33,
21,
1... |
state = XMLElementChild('state', type=State, required=False, test_equal=True) | state = XMLElementChild('state', type=State, required=True, test_equal=True) | def __new__(cls, value): value = int.__new__(cls, value) if value <= 0: raise ValueError("illegal value form refresh element") return value | 6fb985e1ba33642f6f71d96be5535477bccbd89b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5703/6fb985e1ba33642f6f71d96be5535477bccbd89b/iscomposing.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2704,
972,
12,
6429,
16,
460,
4672,
460,
273,
509,
16186,
2704,
972,
12,
6429,
16,
460,
13,
309,
460,
1648,
374,
30,
1002,
2068,
2932,
31751,
460,
646,
4460,
930,
7923,
327,
460,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2704,
972,
12,
6429,
16,
460,
4672,
460,
273,
509,
16186,
2704,
972,
12,
6429,
16,
460,
13,
309,
460,
1648,
374,
30,
1002,
2068,
2932,
31751,
460,
646,
4460,
930,
7923,
327,
460,... |
infos.append(CFG_WEBSESSION_INFO_MESSAGES["JOIN_GROUP"]) | infos.append(_(CFG_WEBSESSION_INFO_MESSAGES["JOIN_GROUP"])) | def perform_request_join_group(uid, grpID, group_name, search, ln=cdslang): """Join group @param grpID: list of the groups the user wants to join, only one value must be selected among the two group lists (default group list, group list resulting from the search) @param group_name: name of the group entered if search on name performed @param search=1 if search performed else 0 @param ln: language @return a (body, errors[], warnings[]) formed tuple warnings != [] if 0 or more than one group is selected """ _ = gettext_set_language(ln) body = "" warnings = [] errors = [] infos = [] if "-1" in grpID: grpID.remove("-1") if len(grpID)==1 : grpID = grpID[0] """test if user is already member or pending""" status = db.get_user_status(uid, grpID) if status: warnings.append('WRN_WEBSESSION_ALREADY_MEMBER') (body, errors, warnings) = perform_request_group_display(uid, infos=infos, errors=errors, warnings=warnings, ln=cdslang) """insert new user of group""" else: group_infos = db.get_group_infos(grpID) group_type = group_infos[0][3] if group_type == CFG_WEBSESSION_GROUP_JOIN_POLICY["VISIBLEMAIL"]: db.insert_new_member(uid, grpID, CFG_WEBSESSION_USERGROUP_STATUS["PENDING"]) admin = db.get_users_by_status(grpID, CFG_WEBSESSION_USERGROUP_STATUS["ADMIN"])[0][1] group_name = group_infos[0][1] msg_subjet, msg_body = websession_templates.tmpl_admin_msg(group_name=group_name, grpID=grpID, ln=ln) (body, errors, warnings, title, navtrail) = perform_request_send(uid, msg_to_user=admin, msg_to_group="", msg_subject=msg_subjet, msg_body=msg_body, ln=ln) infos.append(CFG_WEBSESSION_INFO_MESSAGES["JOIN_REQUEST"]) elif group_type == CFG_WEBSESSION_GROUP_JOIN_POLICY["VISIBLEOPEN"]: db.insert_new_member(uid, grpID, CFG_WEBSESSION_USERGROUP_STATUS["MEMBER"]) infos.append(CFG_WEBSESSION_INFO_MESSAGES["JOIN_GROUP"]) (body, errors, warnings) = perform_request_group_display(uid, infos=infos, errors=errors, warnings=warnings, ln=cdslang) else: warnings.append('WRN_WEBSESSION_MULTIPLE_GROUPS') (body, errors, warnings) = perform_request_input_join_group(uid, group_name, search, warnings, ln) return (body, errors, warnings) | 5b6e928b4222c0527298f19150e237b65b55803a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3933/5b6e928b4222c0527298f19150e237b65b55803a/webgroup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3073,
67,
2293,
67,
5701,
67,
1655,
12,
1911,
16,
14295,
734,
16,
1041,
67,
529,
16,
1623,
16,
7211,
33,
4315,
2069,
539,
4672,
3536,
4572,
1041,
632,
891,
14295,
734,
30,
666,
434,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3073,
67,
2293,
67,
5701,
67,
1655,
12,
1911,
16,
14295,
734,
16,
1041,
67,
529,
16,
1623,
16,
7211,
33,
4315,
2069,
539,
4672,
3536,
4572,
1041,
632,
891,
14295,
734,
30,
666,
434,
... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.