rem stringlengths 2 226k | add stringlengths 0 227k | context stringlengths 8 228k | meta stringlengths 156 215 | input_ids list | attention_mask list | labels list |
|---|---|---|---|---|---|---|
unsupportedError(rendererName) | unsupportedError("BASE") | def load(self, file): """ Loads ps image data from file. | b2a38de6fe098f2955ca5efe49f28bae0ffe39ac /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8476/b2a38de6fe098f2955ca5efe49f28bae0ffe39ac/image.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
12,
2890,
16,
585,
4672,
3536,
4444,
87,
4250,
1316,
501,
628,
585,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
12,
2890,
16,
585,
4672,
3536,
4444,
87,
4250,
1316,
501,
628,
585,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
assert len(topo)==0 | assert len(topo)==1 topo[0].op==theano.compile.function_module.deep_copy_op | def test_multiple_case(self): """ test those case take from the comment in Canonizer x / x -> 1 (x * y) / x -> y x / y / x -> 1 / y x / y / z -> x / (y * z) x / (y / z) -> (x * z) / y (a / b) * (b / c) * (c / d) -> a / d (2.0 * x) / (4.0 * y) -> (0.5 * x) / y 2 * x / 2 -> x with and without DimShuffle TODO: with DimShuffle """ import theano.tensor, theano.compile | 934708245b47acebab846b29b02c6dbed5d87b7d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12438/934708245b47acebab846b29b02c6dbed5d87b7d/test_opt.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
9622,
67,
3593,
12,
2890,
4672,
3536,
1842,
5348,
648,
4862,
628,
326,
2879,
316,
4480,
265,
1824,
619,
342,
619,
317,
404,
261,
92,
380,
677,
13,
342,
619,
317,
677,
619,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
9622,
67,
3593,
12,
2890,
4672,
3536,
1842,
5348,
648,
4862,
628,
326,
2879,
316,
4480,
265,
1824,
619,
342,
619,
317,
404,
261,
92,
380,
677,
13,
342,
619,
317,
677,
619,
... |
self.assertEqual(func, os.remove) self.assertEqual(arg, self.childpath) | if func is os.remove: self.assertEqual(arg, self.childpath) else: self.assertIs(func, os.listdir, "func must be either os.remove or os.listdir") self.assertEqual(arg, TESTFN) | def check_args_to_onerror(self, func, arg, exc): if self.errorState == 0: self.assertEqual(func, os.remove) self.assertEqual(arg, self.childpath) self.failUnless(issubclass(exc[0], OSError)) self.errorState = 1 else: self.assertEqual(func, os.rmdir) self.assertEqual(arg, TESTFN) self.failUnless(issubclass(exc[0], OSError)) self.errorState = 2 | 6234f70d7f634d5c84c8aea9364ffc2b8f3b2ed8 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12029/6234f70d7f634d5c84c8aea9364ffc2b8f3b2ed8/test_shutil.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
1968,
67,
869,
67,
265,
1636,
12,
2890,
16,
1326,
16,
1501,
16,
3533,
4672,
309,
365,
18,
1636,
1119,
422,
374,
30,
309,
1326,
353,
1140,
18,
4479,
30,
365,
18,
11231,
5812,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
1968,
67,
869,
67,
265,
1636,
12,
2890,
16,
1326,
16,
1501,
16,
3533,
4672,
309,
365,
18,
1636,
1119,
422,
374,
30,
309,
1326,
353,
1140,
18,
4479,
30,
365,
18,
11231,
5812,... |
raise TestFailed, 'zip() - no args, expected TypeError, got', e | raise TestFailed, 'zip() - no args, expected TypeError, got %s' % e | def __getitem__(self, i): if i < 0 or i > 2: raise IndexError return i + 4 | 205b98d73507f9524785c43fda7b256370bee4dd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/205b98d73507f9524785c43fda7b256370bee4dd/test_b2.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
31571,
972,
12,
2890,
16,
277,
4672,
309,
277,
411,
374,
578,
277,
405,
576,
30,
1002,
10195,
327,
277,
397,
1059,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
31571,
972,
12,
2890,
16,
277,
4672,
309,
277,
411,
374,
578,
277,
405,
576,
30,
1002,
10195,
327,
277,
397,
1059,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
if folder and self.folder_id.has_key(folder): folder_id = self.folder_id[folder] uri = (gdata.docs.client.FOLDERS_FEED_TEMPLATE % urllib.quote_plus(folder_id)) else: | uri = '' if folder: folder_feed = self.GetDocList(uri='/feeds/default/private/full/-/folder') folder_entry = None for f in folder_feed.entry: if f.title.text == folder: folder_entry = f break if not folder_entry: create_folder = raw_input('Folder ' + folder + ' not found. Create it? (y/N): ') if create_folder.lower() == 'y': folder_entry = self.create(gdata.docs.data.FOLDER_LABEL, folder) if folder_entry: uri = (gdata.docs.client.FOLDERS_FEED_TEMPLATE % urllib.quote_plus(folder_entry.resource_id.text)) if not uri: | def upload_docs(self, paths, title=None, folder=None, convert=True): """Upload a document. Keyword arguments: paths: Paths of files to upload. title: Title to give the files once uploaded. (Defaults to the names of the files). folder: Folder to put the files in. (Defaults to the root folder). convert: If True, converts the files to the native Google Docs format. Otherwise, leaves as arbitrary file type. Only Google Apps Premier users can specify a value other than True. (Default True) | f9669929e25f37072516ef2b51890b79ec08e3e3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/2899/f9669929e25f37072516ef2b51890b79ec08e3e3/service.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3617,
67,
8532,
12,
2890,
16,
2953,
16,
2077,
33,
7036,
16,
3009,
33,
7036,
16,
1765,
33,
5510,
4672,
3536,
4777,
279,
1668,
18,
225,
18317,
1775,
30,
2953,
30,
16643,
434,
1390,
358,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3617,
67,
8532,
12,
2890,
16,
2953,
16,
2077,
33,
7036,
16,
3009,
33,
7036,
16,
1765,
33,
5510,
4672,
3536,
4777,
279,
1668,
18,
225,
18317,
1775,
30,
2953,
30,
16643,
434,
1390,
358,
... |
-sum(sm.product_qty) as qty | -sum(sm.product_qty /uo.factor) as qty | def init(self, cr): cr.execute(""" create or replace view stock_report_prodlots as ( select max(id) as id, location_id, product_id, prodlot_id, sum(qty) as name from ( select -max(sm.id) as id, sm.location_id, sm.product_id, sm.prodlot_id, -sum(sm.product_qty) as qty from stock_move as sm left join stock_location sl on (sl.id = sm.location_id) where state = 'done' group by sm.location_id, sm.product_id, sm.product_uom, sm.prodlot_id union all select max(sm.id) as id, sm.location_dest_id as location_id, sm.product_id, sm.prodlot_id, sum(sm.product_qty) as qty from stock_move as sm left join stock_location sl on (sl.id = sm.location_dest_id) where sm.state = 'done' group by sm.location_dest_id, sm.product_id, sm.product_uom, sm.prodlot_id ) as report group by location_id, product_id, prodlot_id )""") | 181b1388a1ee18efd56178b66c38e63df2f51a3d /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7397/181b1388a1ee18efd56178b66c38e63df2f51a3d/report_stock.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1208,
12,
2890,
16,
4422,
4672,
4422,
18,
8837,
2932,
3660,
752,
578,
1453,
1476,
12480,
67,
6006,
67,
685,
5761,
6968,
487,
261,
2027,
943,
12,
350,
13,
487,
612,
16,
2117,
67,
350,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1208,
12,
2890,
16,
4422,
4672,
4422,
18,
8837,
2932,
3660,
752,
578,
1453,
1476,
12480,
67,
6006,
67,
685,
5761,
6968,
487,
261,
2027,
943,
12,
350,
13,
487,
612,
16,
2117,
67,
350,
... |
release = '2000' | if min == 0: release = '2000' elif min == 1: release = 'XP' elif min == 2: release = '2003Server' else: release = 'post2003' | def win32_ver(release='',version='',csd='',ptype=''): """ Get additional version information from the Windows Registry and return a tuple (version,csd,ptype) referring to version number, CSD level and OS type (multi/single processor). As a hint: ptype returns 'Uniprocessor Free' on single processor NT machines and 'Multiprocessor Free' on multi processor machines. The 'Free' refers to the OS version being free of debugging code. It could also state 'Checked' which means the OS version uses debugging code, i.e. code that checks arguments, ranges, etc. (Thomas Heller). Note: this function only works if Mark Hammond's win32 package is installed and obviously only runs on Win32 compatible platforms. """ # XXX Is there any way to find out the processor type on WinXX ? # XXX Is win32 available on Windows CE ? # Adapted from code posted by Karl Putland to comp.lang.python. # Import the needed APIs try: import win32api except ImportError: return release,version,csd,ptype from win32api import RegQueryValueEx,RegOpenKeyEx,RegCloseKey,GetVersionEx from win32con import HKEY_LOCAL_MACHINE,VER_PLATFORM_WIN32_NT,\ VER_PLATFORM_WIN32_WINDOWS # Find out the registry key and some general version infos maj,min,buildno,plat,csd = GetVersionEx() version = '%i.%i.%i' % (maj,min,buildno & 0xFFFF) if csd[:13] == 'Service Pack ': csd = 'SP' + csd[13:] if plat == VER_PLATFORM_WIN32_WINDOWS: regkey = 'SOFTWARE\\Microsoft\\Windows\\CurrentVersion' # Try to guess the release name if maj == 4: if min == 0: release = '95' else: release = '98' elif maj == 5: release = '2000' elif plat == VER_PLATFORM_WIN32_NT: regkey = 'SOFTWARE\\Microsoft\\Windows NT\\CurrentVersion' if maj <= 4: release = 'NT' elif maj == 5: release = '2000' else: if not release: # E.g. Win3.1 with win32s release = '%i.%i' % (maj,min) return release,version,csd,ptype # Open the registry key try: keyCurVer = RegOpenKeyEx(HKEY_LOCAL_MACHINE,regkey) # Get a value to make sure the key exists... RegQueryValueEx(keyCurVer,'SystemRoot') except: return release,version,csd,ptype # Parse values #subversion = _win32_getvalue(keyCurVer, # 'SubVersionNumber', # ('',1))[0] #if subversion: # release = release + subversion # 95a, 95b, etc. build = _win32_getvalue(keyCurVer, 'CurrentBuildNumber', ('',1))[0] ptype = _win32_getvalue(keyCurVer, 'CurrentType', (ptype,1))[0] # Normalize version version = _norm_version(version,build) # Close key RegCloseKey(keyCurVer) return release,version,csd,ptype | 91e83e209d66d4b7a892c9360f9e524d6e282191 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/91e83e209d66d4b7a892c9360f9e524d6e282191/platform.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5657,
1578,
67,
502,
12,
9340,
2218,
2187,
1589,
2218,
2187,
2143,
72,
2218,
2187,
27672,
2218,
11,
4672,
225,
3536,
968,
3312,
1177,
1779,
628,
326,
8202,
5438,
471,
327,
279,
3193,
261... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5657,
1578,
67,
502,
12,
9340,
2218,
2187,
1589,
2218,
2187,
2143,
72,
2218,
2187,
27672,
2218,
11,
4672,
225,
3536,
968,
3312,
1177,
1779,
628,
326,
8202,
5438,
471,
327,
279,
3193,
261... |
if (PkgSkip(ver)==0 and SDK.has_key(ver)==0): | if (PkgSkip(ver)==0 and ver not in SDK): | def SdkLocateMaya(): for (ver,key) in MAYAVERSIONINFO: if (PkgSkip(ver)==0 and SDK.has_key(ver)==0): GetSdkDir(ver.lower().replace("x",""), ver) if (not SDK.has_key(ver)): if (sys.platform == "win32"): for dev in ["Alias|Wavefront","Alias","Autodesk"]: fullkey="SOFTWARE\\"+dev+"\\Maya\\"+key+"\\Setup\\InstallPath" res = GetRegistryKey(fullkey, "MAYA_INSTALL_LOCATION") if (res != 0): res = res.replace("\\", "/").rstrip("/") SDK[ver] = res elif (sys.platform == "darwin"): ddir = "/Applications/Autodesk/maya"+key+"/Maya.app/Contents" if (os.path.isdir(ddir)): SDK[ver] = ddir else: if (platform.architecture()[0] == "64bit"): ddir1 = "/usr/autodesk/maya"+key+"-x64" ddir2 = "/usr/aw/maya"+key+"-x64" else: ddir1 = "/usr/autodesk/maya"+key ddir2 = "/usr/aw/maya"+key if (os.path.isdir(ddir1)): SDK[ver] = ddir1 elif (os.path.isdir(ddir2)): SDK[ver] = ddir2 | e0e3ede644b14dcbdc2136d9a8ff5730ebaa32a7 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8543/e0e3ede644b14dcbdc2136d9a8ff5730ebaa32a7/makepandacore.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3426,
1333,
340,
49,
528,
69,
13332,
364,
261,
502,
16,
856,
13,
316,
490,
5255,
37,
5757,
5923,
30,
309,
261,
11264,
6368,
12,
502,
13,
631,
20,
471,
1924,
486,
316,
3881,
4672,
968... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3426,
1333,
340,
49,
528,
69,
13332,
364,
261,
502,
16,
856,
13,
316,
490,
5255,
37,
5757,
5923,
30,
309,
261,
11264,
6368,
12,
502,
13,
631,
20,
471,
1924,
486,
316,
3881,
4672,
968... |
if self.user_version == 1: LibraryDatabase.upgrade_version1(self.conn) if self.user_version == 2: LibraryDatabase.upgrade_version2(self.conn) if self.user_version == 3: LibraryDatabase.upgrade_version3(self.conn) if self.user_version == 4: LibraryDatabase.upgrade_version4(self.conn) if self.user_version == 5: LibraryDatabase.upgrade_version5(self.conn) if self.user_version == 6: LibraryDatabase.upgrade_version6(self.conn) if self.user_version == 7: LibraryDatabase.upgrade_version7(self.conn) if self.user_version == 8: LibraryDatabase.upgrade_version8(self.conn) | i = 0 while True: i += 1 func = getattr(LibraryDatabase, 'upgrade_version%d'%i, None) if func is None: break if self.user_version == i: print 'Upgrading database from version: %d'%i func(self.conn) | def __init__(self, dbpath): self.dbpath = dbpath self.conn = _connect(dbpath) self.cache = [] self.data = [] if self.user_version == 0: # No tables have been created LibraryDatabase.create_version1(self.conn) if self.user_version == 1: # Upgrade to 2 LibraryDatabase.upgrade_version1(self.conn) if self.user_version == 2: # Upgrade to 3 LibraryDatabase.upgrade_version2(self.conn) if self.user_version == 3: # Upgrade to 4 LibraryDatabase.upgrade_version3(self.conn) if self.user_version == 4: # Upgrade to 5 LibraryDatabase.upgrade_version4(self.conn) if self.user_version == 5: # Upgrade to 6 LibraryDatabase.upgrade_version5(self.conn) if self.user_version == 6: # Upgrade to 7 LibraryDatabase.upgrade_version6(self.conn) if self.user_version == 7: # Upgrade to 8 LibraryDatabase.upgrade_version7(self.conn) if self.user_version == 8: # Upgrade to 9 LibraryDatabase.upgrade_version8(self.conn) | 7ec93d9f4349050b5ba784486d01dfaff5d17779 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9125/7ec93d9f4349050b5ba784486d01dfaff5d17779/database.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1319,
803,
4672,
365,
18,
1966,
803,
273,
1319,
803,
365,
18,
4646,
273,
389,
3612,
12,
1966,
803,
13,
365,
18,
2493,
273,
5378,
365,
18,
892,
225,
273... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1319,
803,
4672,
365,
18,
1966,
803,
273,
1319,
803,
365,
18,
4646,
273,
389,
3612,
12,
1966,
803,
13,
365,
18,
2493,
273,
5378,
365,
18,
892,
225,
273... |
c=[i for i in range(len(maps)) for j in maps[i]] | def commitData(self): self.settingsChanged = False self.selection=[] selection=self.selectionList maps=[self.rootCluster.mapping[c.first:c.last] for c in [e.rootCluster for e in selection]] self.selection=[self.matrix.items[k] for k in [j for i in range(len(maps)) for j in maps[i]]] if not self.selection: self.send("Selected Examples",None) self.send("Unselected Examples",None) self.send("Structured Data Files", None) return items = getattr(self.matrix, "items") if self.matrixSource == "Example Distance": unselected = [item for item in items if item not in self.selection] if self.ClassifySelected: clustVar=orange.EnumVariable(str(self.ClassifyName) , values=["Cluster " + str(i) for i in range(len(maps))] + ["Other"]) c=[i for i in range(len(maps)) for j in maps[i]] origDomain = items.domain if self.addIdAs == 0: domain=orange.Domain(origDomain.attributes,clustVar) if origDomain.classVar: domain.addmeta(orange.newmetaid(), origDomain.classVar) aid = -1 elif self.addIdAs == 1: domain=orange.Domain(origDomain.attributes+[clustVar], origDomain.classVar) aid = len(origDomain.attributes) else: domain=orange.Domain(origDomain.attributes, origDomain.classVar) aid=orange.newmetaid() domain.addmeta(aid, clustVar) | 8831277dc381559aa4de7aa848bba2c91b0d2e4c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6366/8831277dc381559aa4de7aa848bba2c91b0d2e4c/OWHierarchicalClustering.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3294,
751,
12,
2890,
4672,
365,
18,
4272,
5033,
273,
1083,
365,
18,
10705,
33,
8526,
4421,
33,
2890,
18,
10705,
682,
7565,
22850,
2890,
18,
3085,
3629,
18,
6770,
63,
71,
18,
3645,
30,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3294,
751,
12,
2890,
4672,
365,
18,
4272,
5033,
273,
1083,
365,
18,
10705,
33,
8526,
4421,
33,
2890,
18,
10705,
682,
7565,
22850,
2890,
18,
3085,
3629,
18,
6770,
63,
71,
18,
3645,
30,
... | |
self.current_table.attach(col,i,i+1,self.current_table_row-1,self.current_table_row,gtk.FILL,gtk.FILL,0,0) | right_attach = i+1 if options.has_key('colspan'): right_attach = i+options['colspan'] self.current_table.attach(col, i, right_attach, self.current_table_row-1, self.current_table_row, gtk.FILL|gtk.EXPAND, gtk.FILL|gtk.EXPAND, 0, 0) | def row(self,*columns, **options): if not self.current_table: self.start_table() | c2382404387dc87cc861620b92f02d5025c3db16 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5568/c2382404387dc87cc861620b92f02d5025c3db16/GTKUI.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1027,
12,
2890,
16,
14,
5112,
16,
2826,
2116,
4672,
309,
486,
365,
18,
2972,
67,
2121,
30,
365,
18,
1937,
67,
2121,
1435,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1027,
12,
2890,
16,
14,
5112,
16,
2826,
2116,
4672,
309,
486,
365,
18,
2972,
67,
2121,
30,
365,
18,
1937,
67,
2121,
1435,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
'now': datetime.datetime.now().isoformat(' '), | 'now': datetime.datetime.now().isoformat(' ')[0:-7] + '+0000', | def write_infos(self, modules): import release self.buffer.write("# Translation of %(project)s.\n" \ "# This file contains the translation of the following modules:\n" \ "%(modules)s" \ "#\n" \ "msgid \"\"\n" \ "msgstr \"\"\n" \ '''"Project-Id-Version: %(project)s %(version)s\\n"\n''' \ '''"Report-Msgid-Bugs-To: %(bugmail)s\\n"\n''' \ '''"POT-Creation-Date: %(now)s\\n"\n''' \ '''"PO-Revision-Date: %(now)s\\n"\n''' \ '''"Last-Translator: <>\\n"\n''' \ '''"Language-Team: \\n"\n''' \ '''"MIME-Version: 1.0\\n"\n''' \ '''"Content-Type: text/plain; charset=UTF-8\\n"\n''' \ '''"Content-Transfer-Encoding: \\n"\n''' \ '''"Plural-Forms: \\n"\n''' \ "\n" | 15b965fb6f19abed1edebdf07a59c2efa0ac2fbf /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7397/15b965fb6f19abed1edebdf07a59c2efa0ac2fbf/translate.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
67,
18227,
12,
2890,
16,
4381,
4672,
1930,
3992,
365,
18,
4106,
18,
2626,
2932,
7,
17427,
434,
8975,
4406,
13,
87,
8403,
82,
6,
521,
6619,
1220,
585,
1914,
326,
4794,
434,
326,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
67,
18227,
12,
2890,
16,
4381,
4672,
1930,
3992,
365,
18,
4106,
18,
2626,
2932,
7,
17427,
434,
8975,
4406,
13,
87,
8403,
82,
6,
521,
6619,
1220,
585,
1914,
326,
4794,
434,
326,
... |
'UNICODE': ('ref/unicode', 'encodings unicode TYPES STRING'), | 'UNICODE': ('ref/strings', 'encodings unicode TYPES STRING'), | def writedocs(dir, pkgpath='', done=None): """Write out HTML documentation for all modules in a directory tree.""" if done is None: done = {} for file in os.listdir(dir): path = os.path.join(dir, file) if ispackage(path): writedocs(path, pkgpath + file + '.', done) elif os.path.isfile(path): modname = inspect.getmodulename(path) if modname: modname = pkgpath + modname if not modname in done: done[modname] = 1 writedoc(modname) | df9dfa85d260060b89dc4c838c3bb4d39715243d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/df9dfa85d260060b89dc4c838c3bb4d39715243d/pydoc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2518,
329,
22280,
12,
1214,
16,
3475,
803,
2218,
2187,
2731,
33,
7036,
4672,
3536,
3067,
596,
3982,
7323,
364,
777,
4381,
316,
279,
1867,
2151,
12123,
309,
2731,
353,
599,
30,
2731,
273,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2518,
329,
22280,
12,
1214,
16,
3475,
803,
2218,
2187,
2731,
33,
7036,
4672,
3536,
3067,
596,
3982,
7323,
364,
777,
4381,
316,
279,
1867,
2151,
12123,
309,
2731,
353,
599,
30,
2731,
273,... |
self.perms="0%s"%(self.perms) elif d['paranoid']: | self.perms = "0%s" % (self.perms) elif mgd['paranoid']: | def GetInfo(self, filename): for line in open(filename).readlines(): m = self.info.match(line) if not m: continue else: d = m.groupdict() if d['owner']: self.owner=d['owner'] elif d['group']: self.group=d['group'] elif d['encoding']: self.encoding=d['encoding'] elif d['perms']: self.perms=d['perms'] if len(self.perms) == 3: self.perms="0%s"%(self.perms) elif d['paranoid']: self.paranoid = True | 853220b7567cb814567d38bf7d0924fd4ebd2d1c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11867/853220b7567cb814567d38bf7d0924fd4ebd2d1c/cfg.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
966,
12,
2890,
16,
1544,
4672,
364,
980,
316,
1696,
12,
3459,
2934,
896,
3548,
13332,
312,
273,
365,
18,
1376,
18,
1916,
12,
1369,
13,
309,
486,
312,
30,
1324,
469,
30,
302,
273... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
966,
12,
2890,
16,
1544,
4672,
364,
980,
316,
1696,
12,
3459,
2934,
896,
3548,
13332,
312,
273,
365,
18,
1376,
18,
1916,
12,
1369,
13,
309,
486,
312,
30,
1324,
469,
30,
302,
273... |
self.w.date_column.emit ("clicked") self.w.date_column.emit ("clicked") self.critical_section_leave () def critical_section_enter (self): if not self._stopevent.isSet (): gtk.gdk.threads_enter() return True return False def critical_section_leave (self): if not self._stopevent.isSet (): gtk.gdk.threads_leave() return True return False | gobject.idle_add(self.w.date_column.emit, "clicked") gobject.idle_add(self.w.date_column.emit, "clicked") | def run(self): l = self.w.snap_path.split (".zfs") path_before_snap = l[0] l = self.w.filename.split (path_before_snap) path_after_snap = l[1] snap_path = "%s.zfs/snapshot/" % path_before_snap; dirs = os.listdir(snap_path) | 76dfaec72b6d234901dd0cafefd54ce8da26001e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11404/76dfaec72b6d234901dd0cafefd54ce8da26001e/fileversion.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
328,
273,
365,
18,
91,
18,
28449,
67,
803,
18,
4939,
7566,
18,
94,
2556,
7923,
589,
67,
5771,
67,
28449,
273,
328,
63,
20,
65,
328,
273,
365,
18,
91,
18,
3459... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
328,
273,
365,
18,
91,
18,
28449,
67,
803,
18,
4939,
7566,
18,
94,
2556,
7923,
589,
67,
5771,
67,
28449,
273,
328,
63,
20,
65,
328,
273,
365,
18,
91,
18,
3459... |
d = A.solve_right(v, check_rank=False).denominator() | d = A._solve_right_nonsingular_square(v, check_rank=False).denominator() | def det_padic(A, proof=True, stabilize=2): """ Return the determinant of A, computed using a p-adic/multimodular algorithm. INPUTS: A -- a square matrix proof -- boolean stabilize (default: 2) -- if proof False, number of successive primes so that CRT det must stabilize. EXAMPLES: sage: import sage.matrix.matrix_integer_dense_hnf as h sage: a = matrix(ZZ, 3, [1..9]) sage: h.det_padic(a) 0 sage: a = matrix(ZZ, 3, [1,2,5,-7,8,10,192,5,18]) sage: h.det_padic(a) -3669 sage: a.determinant(algorithm='ntl') -3669 """ if not A.is_square(): raise ValueError, "A must be a square matrix" r = A.rank() if r < A.nrows(): return ZZ(0) v = random_matrix(ZZ, A.nrows(), 1) d = A.solve_right(v, check_rank=False).denominator() return det_given_divisor(A, d, proof=proof, stabilize=stabilize) | 6ba9dd423dd321b782176073bd278099084ce246 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/6ba9dd423dd321b782176073bd278099084ce246/matrix_integer_dense_hnf.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3023,
67,
6982,
335,
12,
37,
16,
14601,
33,
5510,
16,
384,
22681,
554,
33,
22,
4672,
3536,
2000,
326,
6328,
970,
434,
432,
16,
8470,
1450,
279,
293,
17,
20333,
19,
5421,
381,
369,
24... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3023,
67,
6982,
335,
12,
37,
16,
14601,
33,
5510,
16,
384,
22681,
554,
33,
22,
4672,
3536,
2000,
326,
6328,
970,
434,
432,
16,
8470,
1450,
279,
293,
17,
20333,
19,
5421,
381,
369,
24... |
{'allowXULXBL': ['localhost', '<file>']}) | {'allowXULXBL': [('localhost', True), ('<file>', True)]}) | def createReftestProfile(self, options, profileDir): "Sets up a profile for reftest." | 0c9d14a0434a02b39c32f3fd5df4cd797b26c567 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11102/0c9d14a0434a02b39c32f3fd5df4cd797b26c567/runreftest.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
426,
1222,
395,
4029,
12,
2890,
16,
702,
16,
3042,
1621,
4672,
315,
2785,
731,
279,
3042,
364,
283,
1222,
395,
1199,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
426,
1222,
395,
4029,
12,
2890,
16,
702,
16,
3042,
1621,
4672,
315,
2785,
731,
279,
3042,
364,
283,
1222,
395,
1199,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
self.assertEquals(2, obj.pageStart) self.assertEquals(4, obj.pageEnd) | def testFormat(self): portal = self.layer['portal'] request = self.layer['request'] setRoles(portal, TEST_USER_NAME, ['Manager']) reviews = portal.reviews reviews.invokeFactory('Publication', id='pub', title='pub') publication = reviews.pub publication.invokeFactory('Volume', id='vol', title='vol') vol = publication.vol vol.invokeFactory('Issue', id='issue', title='issue') issue = vol.issue pt = getToolByName(portal, 'portal_types') issueType = pt.getTypeInfo(issue) # We are lazy here, allowing an issue to contain all documents issueType.filter_content_types = False | 210449271edd39a39aa8b4fde2b3d84539c4b5a5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10562/210449271edd39a39aa8b4fde2b3d84539c4b5a5/testExcelImport.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
1630,
12,
2890,
4672,
11899,
273,
365,
18,
6363,
3292,
24386,
3546,
590,
273,
365,
18,
6363,
3292,
2293,
3546,
444,
6898,
12,
24386,
16,
22130,
67,
4714,
67,
1985,
16,
10228,
1318,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
1630,
12,
2890,
4672,
11899,
273,
365,
18,
6363,
3292,
24386,
3546,
590,
273,
365,
18,
6363,
3292,
2293,
3546,
444,
6898,
12,
24386,
16,
22130,
67,
4714,
67,
1985,
16,
10228,
1318,... | |
for hour in range(24): if (hour > 0): if (hour == 1): hourString = "am 1" elif (hour == 12): hourString = "pm 12" elif (hour > 12): hourString = str(hour - 12) else: hourString = str(hour) | for hour,hourString in self.GetLocaleHourStrings(range(24)): if hour > 0: | def DrawBackground(self, dc): styles = self.parent self._doDrawingCalculations() | 590a7421d4a56497721ad83a36a6f3eb4226407e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/590a7421d4a56497721ad83a36a6f3eb4226407e/CalendarCanvas.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10184,
8199,
12,
2890,
16,
6744,
4672,
5687,
273,
365,
18,
2938,
365,
6315,
2896,
26885,
4844,
1012,
1435,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10184,
8199,
12,
2890,
16,
6744,
4672,
5687,
273,
365,
18,
2938,
365,
6315,
2896,
26885,
4844,
1012,
1435,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
b = re.sub("(Linux 2.4\..+$", "Linux2.4", b) | b = re.sub("Linux 2.4\..+$", "Linux2.4", b) | def FormInstallRoot(PackageSetName): AltConfig = {"NT386":"x86"}.get(Config, Config) a = os.path.join(GetStage(), "cm3-" + PackageSetName + "-" + AltConfig + "-" + CM3VERSION) if Config == "NT386" or Config == "I386_NT": a = a + "-VC" + GetVisualCPlusPlusVersion() else: b = os.popen("uname -sr").read() b = re.sub("(Linux 2.4\..+$", "Linux2.4", b) b = re.sub("Linux 2.6\..+$", "", b) b = re.sub("-+$", "", b) b = re.sub("SunOS", "Solaris", b) b = re.sub(" ", "", b) a = a + "-" + b a = a + "-" + time.strftime("%Y%m%d") return a | aa730c9fadab02017215590dadfa2f0ad126598d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9328/aa730c9fadab02017215590dadfa2f0ad126598d/pylib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2748,
6410,
2375,
12,
2261,
11924,
4672,
27116,
809,
273,
12528,
1784,
23,
5292,
15563,
92,
5292,
6,
5496,
588,
12,
809,
16,
1903,
13,
279,
273,
1140,
18,
803,
18,
5701,
12,
967,
8755,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2748,
6410,
2375,
12,
2261,
11924,
4672,
27116,
809,
273,
12528,
1784,
23,
5292,
15563,
92,
5292,
6,
5496,
588,
12,
809,
16,
1903,
13,
279,
273,
1140,
18,
803,
18,
5701,
12,
967,
8755,... |
vals = isposinf(array((-1.,0,1))/0.) seterr(**olderr) | try: vals = isposinf(array((-1.,0,1))/0.) finally: seterr(**olderr) | def test_generic(self): olderr = seterr(divide='ignore', invalid='ignore') vals = isposinf(array((-1.,0,1))/0.) seterr(**olderr) assert(vals[0] == 0) assert(vals[1] == 0) assert(vals[2] == 1) | 5b22c87bed10fd5c3e028d23590e9b6dd5e8d3af /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14925/5b22c87bed10fd5c3e028d23590e9b6dd5e8d3af/test_type_check.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
13540,
12,
2890,
4672,
1592,
370,
273,
444,
370,
12,
2892,
831,
2218,
6185,
2187,
2057,
2218,
6185,
6134,
775,
30,
5773,
273,
353,
917,
10625,
12,
1126,
12443,
17,
21,
12990,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
13540,
12,
2890,
4672,
1592,
370,
273,
444,
370,
12,
2892,
831,
2218,
6185,
2187,
2057,
2218,
6185,
6134,
775,
30,
5773,
273,
353,
917,
10625,
12,
1126,
12443,
17,
21,
12990,
... |
availble, and can be used to control the aspect of the colorbar:: | available, and can be used to control the aspect of the colorbar:: | def add_colorbar(self, *args, **kwargs): ''' Add a colorbar to the current figure | cb4598af3ee0e5348da308837ea07d2cf3f29a0d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/13232/cb4598af3ee0e5348da308837ea07d2cf3f29a0d/aplpy.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
527,
67,
3266,
3215,
12,
2890,
16,
380,
1968,
16,
2826,
4333,
4672,
9163,
1436,
279,
20644,
358,
326,
783,
7837,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
527,
67,
3266,
3215,
12,
2890,
16,
380,
1968,
16,
2826,
4333,
4672,
9163,
1436,
279,
20644,
358,
326,
783,
7837,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
parent.window.set_cursor(None) | if parent.window: parent.window.set_cursor(None) | def __disconnect(self, sig, parent): parent.window.set_cursor(None) parent.disconnect(sig) | 9809de1ef07ab604ce6fad8163ae8806c3ec2e0f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4764/9809de1ef07ab604ce6fad8163ae8806c3ec2e0f/wlw.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
20177,
12,
2890,
16,
3553,
16,
982,
4672,
309,
982,
18,
5668,
30,
982,
18,
5668,
18,
542,
67,
9216,
12,
7036,
13,
982,
18,
20177,
12,
7340,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
20177,
12,
2890,
16,
3553,
16,
982,
4672,
309,
982,
18,
5668,
30,
982,
18,
5668,
18,
542,
67,
9216,
12,
7036,
13,
982,
18,
20177,
12,
7340,
13,
2,
-100,
-100,
-100,
-100,
-100,... |
Returns a Sylow p-subgroups of the finite group G, where p is a prime. This is a p-subgroup of G whose index in G is coprime to p. Wraps the GAP function SylowSubgroup. | Returns a Sylow `p`-subgroup of the finite group `G`, where `p` is a prime. This is a `p`-subgroup of `G` whose index in `G` is coprime to `p`. Wraps the GAP function ``SylowSubgroup``. | def sylow_subgroup(self, p): """ Returns a Sylow p-subgroups of the finite group G, where p is a prime. This is a p-subgroup of G whose index in G is coprime to p. Wraps the GAP function SylowSubgroup. EXAMPLES:: sage: G = PermutationGroup(['(1,2,3)', '(2,3)']) sage: G.sylow_subgroup(2) Permutation Group with generators [(2,3)] sage: G.sylow_subgroup(5) Permutation Group with generators [()] """ G = self gap.eval("G := %s"%G._gap_init_()) gap.eval("Ssgp := SylowSubgroup(G, %s);"%p) gap.eval("gens := GeneratorsOfGroup( Ssgp );") N = Integer(gap.eval("N := Length(gens);")) if N>0: gens = [PermutationGroupElement(gap.eval("gens[%s];"%j)) for j in range(1,N+1)] H = PermutationGroup(gens) else: H = PermutationGroup([()]) return H | 1a05983ebb3835bbd44e7ffd0287f6bbaee2ac36 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/1a05983ebb3835bbd44e7ffd0287f6bbaee2ac36/permgroup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1393,
821,
67,
1717,
1655,
12,
2890,
16,
293,
4672,
3536,
2860,
279,
16455,
821,
1375,
84,
68,
17,
1717,
1655,
434,
326,
25922,
1041,
1375,
43,
9191,
1625,
1375,
84,
68,
353,
279,
1701... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1393,
821,
67,
1717,
1655,
12,
2890,
16,
293,
4672,
3536,
2860,
279,
16455,
821,
1375,
84,
68,
17,
1717,
1655,
434,
326,
25922,
1041,
1375,
43,
9191,
1625,
1375,
84,
68,
353,
279,
1701... |
if n <= 0: return str | if n <= 0: return str | def _center(str, width): """Center a string in a field.""" n = width - len(str) if n <= 0: return str return ' '*((n+1)/2) + str + ' '*((n)/2) | ad3bc44d52b9b9c1f1c7449d0991138fed5fe978 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/ad3bc44d52b9b9c1f1c7449d0991138fed5fe978/calendar.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
5693,
12,
701,
16,
1835,
4672,
3536,
8449,
279,
533,
316,
279,
652,
12123,
290,
273,
1835,
300,
562,
12,
701,
13,
309,
290,
1648,
374,
30,
327,
609,
327,
296,
14609,
12443,
82,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
5693,
12,
701,
16,
1835,
4672,
3536,
8449,
279,
533,
316,
279,
652,
12123,
290,
273,
1835,
300,
562,
12,
701,
13,
309,
290,
1648,
374,
30,
327,
609,
327,
296,
14609,
12443,
82,
... |
if self.port != None and self.curpose != "": print "Capturing pose..." errors = "could not read servos: " errCount = 0.0 dlg = wx.ProgressDialog("capturing pose","this may take a few seconds, please wait...",self.parent.project.count + 1) dlg.Update(1) for servo in range(self.parent.project.count): pos = self.port.getReg(servo+1,P_PRESENT_POSITION_L, 2) if pos != -1 and len(pos) > 1: self.servos[servo].position.SetValue(pos[0] + (pos[1]<<8)) else: errors = errors + str(servo+1) + ", " errCount = errCount + 1.0 if self.curpose != "": self.parent.project.poses[self.curpose][servo] = self.servos[servo].position.GetValue() val = servo+2 dlg.Update(val) if errors != "could not read servos: ": self.parent.sb.SetStatusText(errors[0:-2],0) if errCount/self.parent.project.count > 0.1 and self.parent.port.ser.timeout < 10: self.parent.port.ser.timeout = self.parent.port.ser.timeout * 2.0 print "Raised timeout threshold to ", self.parent.port.ser.timeout | if self.port != None: if self.curpose != "": print "Capturing pose..." errors = "could not read servos: " errCount = 0.0 dlg = wx.ProgressDialog("capturing pose","this may take a few seconds, please wait...",self.parent.project.count + 1) dlg.Update(1) for servo in range(self.parent.project.count): pos = self.port.getReg(servo+1,P_PRESENT_POSITION_L, 2) if pos != -1 and len(pos) > 1: self.servos[servo].position.SetValue(pos[0] + (pos[1]<<8)) else: errors = errors + str(servo+1) + ", " errCount = errCount + 1.0 if self.curpose != "": self.parent.project.poses[self.curpose][servo] = self.servos[servo].position.GetValue() val = servo+2 dlg.Update(val) if errors != "could not read servos: ": self.parent.sb.SetStatusText(errors[0:-2],0) if errCount/self.parent.project.count > 0.1 and self.parent.port.ser.timeout < 10: self.parent.port.ser.timeout = self.parent.port.ser.timeout * 2.0 print "Raised timeout threshold to ", self.parent.port.ser.timeout else: self.parent.sb.SetStatusText("captured pose!",0) dlg.Destroy() self.parent.project.save = True | def capturePose(self, e=None): """ Downloads the current pose from the robot to the GUI. """ if self.port != None and self.curpose != "": print "Capturing pose..." errors = "could not read servos: " errCount = 0.0 dlg = wx.ProgressDialog("capturing pose","this may take a few seconds, please wait...",self.parent.project.count + 1) dlg.Update(1) for servo in range(self.parent.project.count): pos = self.port.getReg(servo+1,P_PRESENT_POSITION_L, 2) if pos != -1 and len(pos) > 1: self.servos[servo].position.SetValue(pos[0] + (pos[1]<<8)) else: errors = errors + str(servo+1) + ", " errCount = errCount + 1.0 if self.curpose != "": self.parent.project.poses[self.curpose][servo] = self.servos[servo].position.GetValue() val = servo+2 dlg.Update(val) if errors != "could not read servos: ": self.parent.sb.SetStatusText(errors[0:-2],0) # if we are failing a lot, raise the timeout if errCount/self.parent.project.count > 0.1 and self.parent.port.ser.timeout < 10: self.parent.port.ser.timeout = self.parent.port.ser.timeout * 2.0 print "Raised timeout threshold to ", self.parent.port.ser.timeout else: self.parent.sb.SetStatusText("captured pose!",0) dlg.Destroy() self.parent.project.save = True else: self.parent.sb.SetBackgroundColour('RED') self.parent.sb.SetStatusText("No Port Open",0) self.parent.timer.Start(20) | 54be78d70f9c90925d63d83ce44952a672cd568d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9859/54be78d70f9c90925d63d83ce44952a672cd568d/PoseEditor.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7477,
52,
2584,
12,
2890,
16,
425,
33,
7036,
4672,
3536,
13059,
87,
326,
783,
26948,
628,
326,
14386,
358,
326,
10978,
18,
3536,
309,
365,
18,
655,
480,
599,
471,
365,
18,
1397,
4150,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7477,
52,
2584,
12,
2890,
16,
425,
33,
7036,
4672,
3536,
13059,
87,
326,
783,
26948,
628,
326,
14386,
358,
326,
10978,
18,
3536,
309,
365,
18,
655,
480,
599,
471,
365,
18,
1397,
4150,
... |
elif ch == 'v': p += '\013' elif ch == 'a': p += '\007' | elif ch == 'v': p += u'\013' elif ch == 'a': p += u'\007' | def PyUnicode_DecodeUnicodeEscape(s, size, errors): if (size == 0): return u'' p = [] pos = 0 while (pos < size): | 16ed2024742360e02d96bf99f8cb5ce4527577cd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6934/16ed2024742360e02d96bf99f8cb5ce4527577cd/app_codecs.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4707,
16532,
67,
6615,
16532,
8448,
12,
87,
16,
963,
16,
1334,
4672,
225,
309,
261,
1467,
422,
374,
4672,
327,
582,
6309,
225,
293,
273,
5378,
949,
273,
374,
1323,
261,
917,
411,
963,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4707,
16532,
67,
6615,
16532,
8448,
12,
87,
16,
963,
16,
1334,
4672,
225,
309,
261,
1467,
422,
374,
4672,
327,
582,
6309,
225,
293,
273,
5378,
949,
273,
374,
1323,
261,
917,
411,
963,
... |
print svn_binary | print prog.svn_binary | def check_tools(): """Check that needed external tools can be found.""" if not os.path.exists(svn_binary): print 'Subversion binary could not be found from:' print svn_binary print 'Please check script configuration and/or make sure Subversion is installed.' return False vs_cmd = get_vs_ide_bin() if not os.path.exists(vs_cmd): print 'Cannot find Visual Studio IDE binary from:' print vs_cmd print 'Please check script configuration.' return False return True | 00a1fb5d68080ec3c9ad49bb9ef5c23d4d340126 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5344/00a1fb5d68080ec3c9ad49bb9ef5c23d4d340126/create_release.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
6642,
13332,
3536,
1564,
716,
3577,
3903,
8513,
848,
506,
1392,
12123,
225,
309,
486,
1140,
18,
803,
18,
1808,
12,
31505,
67,
8578,
4672,
1172,
296,
1676,
1589,
3112,
3377,
486,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
6642,
13332,
3536,
1564,
716,
3577,
3903,
8513,
848,
506,
1392,
12123,
225,
309,
486,
1140,
18,
803,
18,
1808,
12,
31505,
67,
8578,
4672,
1172,
296,
1676,
1589,
3112,
3377,
486,... |
assert str(e.reason) == 'insert_broken_csc' | assert str(e.reason) == 'insert_broken_add' | def insert_broken_csc(node): if node.op == theano.sparse.sd_csc: return [off_by_half(*node.inputs)] return False | 46305ee5eb1cd7beebd74117366afb20cd152a02 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12438/46305ee5eb1cd7beebd74117366afb20cd152a02/test_debugmode.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2243,
67,
70,
19906,
67,
71,
1017,
12,
2159,
4672,
309,
756,
18,
556,
422,
326,
31922,
18,
17472,
18,
6427,
67,
71,
1017,
30,
327,
306,
3674,
67,
1637,
67,
20222,
30857,
2159,
18,
10... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2243,
67,
70,
19906,
67,
71,
1017,
12,
2159,
4672,
309,
756,
18,
556,
422,
326,
31922,
18,
17472,
18,
6427,
67,
71,
1017,
30,
327,
306,
3674,
67,
1637,
67,
20222,
30857,
2159,
18,
10... |
@param name: @param arg: | @param name: name of the field @param arg: user defined argument | def _child_compute(self, cr, uid, ids, name, arg, context={}): """ @param self: The object pointer. @param cr: A database cursor @param uid: ID of the user currently logged in @param ids: the ID of mrp.production object @param name: @param arg: | 001d95a574dd31a9b0d995f0ac55365b570d3806 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/001d95a574dd31a9b0d995f0ac55365b570d3806/mrp.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
3624,
67,
9200,
12,
2890,
16,
4422,
16,
4555,
16,
3258,
16,
508,
16,
1501,
16,
819,
12938,
4672,
3536,
632,
891,
365,
30,
1021,
733,
4407,
18,
632,
891,
4422,
30,
432,
2063,
334... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
3624,
67,
9200,
12,
2890,
16,
4422,
16,
4555,
16,
3258,
16,
508,
16,
1501,
16,
819,
12938,
4672,
3536,
632,
891,
365,
30,
1021,
733,
4407,
18,
632,
891,
4422,
30,
432,
2063,
334... |
_fl_create_clock = cfuncproto(so_libforms, "fl_create_clock", | _fl_create_clock = cfuncproto(so_libforms, "fl_create_clock", | def fl_request_clipboard(ob, type, py_got_it_callback): """ fl_request_clipboard(ob, type, py_got_it_callback) -> num. """ c_got_it_callback = FL_SELECTION_CB(py_got_it_callback) _cfunc_refs[get_rand_dictkey()] = c_got_it_callback retval = _fl_request_clipboard(ob, type, c_got_it_callback) return retval | 9942dac8ce2b35a1e43615a26fd8e7054ef805d3 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/2429/9942dac8ce2b35a1e43615a26fd8e7054ef805d3/xformslib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1183,
67,
2293,
67,
14161,
3752,
12,
947,
16,
618,
16,
2395,
67,
13212,
67,
305,
67,
3394,
4672,
3536,
1183,
67,
2293,
67,
14161,
3752,
12,
947,
16,
618,
16,
2395,
67,
13212,
67,
305... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1183,
67,
2293,
67,
14161,
3752,
12,
947,
16,
618,
16,
2395,
67,
13212,
67,
305,
67,
3394,
4672,
3536,
1183,
67,
2293,
67,
14161,
3752,
12,
947,
16,
618,
16,
2395,
67,
13212,
67,
305... |
gtk.main_iteration() self.timer_id = gobject.timeout_add(2000, self.hide_cursor) if y > 0.9*self.available_image_height(): self.slideshow_controls_show() else: self.slideshow_controls_hide() return True def button_pressed(self, widget, event): if self.image_loaded == True: | gtk.main_iteration() self.timer_id = gobject.timeout_add(2000, self.hide_cursor) if y > 0.9*self.available_image_height(): self.slideshow_controls_show() else: self.slideshow_controls_hide() return True def button_pressed(self, widget, event): if self.image_loaded == True: | def mouse_moved(self, widget, event): # This handles the panning of the image | a4d6024d09d807bd2a142bf54daca1a7d1873c7f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2291/a4d6024d09d807bd2a142bf54daca1a7d1873c7f/mirage.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7644,
67,
81,
9952,
12,
2890,
16,
3604,
16,
871,
4672,
468,
1220,
7372,
326,
293,
10903,
434,
326,
1316,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7644,
67,
81,
9952,
12,
2890,
16,
3604,
16,
871,
4672,
468,
1220,
7372,
326,
293,
10903,
434,
326,
1316,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
""" used to send a typing notification. Upon receiving this message the official client will display a 'user is typing' message to all other users in the chat session for 10 seconds. The official client sends one of these every 5 seconds (I think) as long as you continue to type. | """ used to send a typing notification. Upon receiving this message the official client will display a 'user is typing' message to all other users in the chat session for 10 seconds. The official client sends one of these every 5 seconds (I think) as long as you continue to type. | def sendTypingNotification(self): """ used to send a typing notification. Upon receiving this message the official client will display a 'user is typing' message to all other users in the chat session for 10 seconds. The official client sends one of these every 5 seconds (I think) as long as you continue to type. """ m = MSNMessage() m.ack = m.MESSAGE_ACK_NONE m.setHeader('Content-Type', 'text/x-msmsgscontrol') m.setHeader('TypingUser', self.userHandle) m.message = "\r\n" self.sendMessage(m) | 9b2b7163e441396d9e5e7f46775cd7809b4655be /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12595/9b2b7163e441396d9e5e7f46775cd7809b4655be/msn.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1366,
18488,
310,
4386,
12,
2890,
4672,
3536,
1399,
358,
1366,
279,
13408,
3851,
18,
587,
500,
15847,
333,
883,
326,
3397,
22354,
1004,
903,
2562,
279,
296,
1355,
353,
13408,
11,
883,
35... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1366,
18488,
310,
4386,
12,
2890,
4672,
3536,
1399,
358,
1366,
279,
13408,
3851,
18,
587,
500,
15847,
333,
883,
326,
3397,
22354,
1004,
903,
2562,
279,
296,
1355,
353,
13408,
11,
883,
35... |
for mypkg in os.listdir(os.getcwd()+"/"+x+"/"+y): if mypkg[-7:] != ".ebuild": continue mypkg=mypkg[:-7] mykey=x+"/"+y fullpkg=x+"/"+mypkg if not portagedict.has_key(mykey): portagedict[mykey]=[] portagedict[mykey].append([fullpkg,catpkgsplit(fullpkg)]) | for mypkg in os.listdir(os.getcwd()+"/"+x+"/"+y): if mypkg[-7:] != ".ebuild": continue mypkg=mypkg[:-7] mykey=x+"/"+y fullpkg=x+"/"+mypkg if not portagedict.has_key(mykey): portagedict[mykey]=[] portagedict[mykey].append([fullpkg,catpkgsplit(fullpkg)]) | def port_porttree(): """ This function builds a dictionary of available ebuilds in the portage tree. Dictionary format is: mydict["cat/pkg"]=[ ["cat/fullpkgname",["cat","pkg","ver","rev"] ["cat/fullpkgname",["cat","pkg","ver2","rev2"] ] """ portagedict={} mydir=getsetting("PORTDIR") if not os.path.isdir(mydir): return origdir=os.getcwd() os.chdir(mydir) for x in categories: for y in os.listdir(os.getcwd()+"/"+x): if not os.path.isdir(os.getcwd()+"/"+x+"/"+y): continue if y=="CVS": continue for mypkg in os.listdir(os.getcwd()+"/"+x+"/"+y): if mypkg[-7:] != ".ebuild": continue mypkg=mypkg[:-7] mykey=x+"/"+y fullpkg=x+"/"+mypkg if not portagedict.has_key(mykey): portagedict[mykey]=[] portagedict[mykey].append([fullpkg,catpkgsplit(fullpkg)]) os.chdir(origdir) return portagedict | fc29702130c8047de55897e04d2d608bc69575cf /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2807/fc29702130c8047de55897e04d2d608bc69575cf/portage.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1756,
67,
655,
3413,
13332,
3536,
1220,
445,
10736,
279,
3880,
434,
2319,
425,
27324,
316,
326,
1756,
410,
2151,
18,
16447,
740,
353,
30,
3399,
1576,
9614,
2574,
19,
10657,
11929,
22850,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1756,
67,
655,
3413,
13332,
3536,
1220,
445,
10736,
279,
3880,
434,
2319,
425,
27324,
316,
326,
1756,
410,
2151,
18,
16447,
740,
353,
30,
3399,
1576,
9614,
2574,
19,
10657,
11929,
22850,
... |
def test_date_triple_value(self): | def test_date_tuple_value(self): | def test_date_triple_value(self): self.assertEqual(self.date.tuple(), (2004, 9, 4, 0, 17, 15.799999977461994)) | c94d72ca054bee8d59504a624c355033bffb3b39 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10979/c94d72ca054bee8d59504a624c355033bffb3b39/test_dates.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
712,
67,
8052,
67,
1132,
12,
2890,
4672,
365,
18,
11231,
5812,
12,
2890,
18,
712,
18,
8052,
9334,
261,
6976,
24,
16,
2468,
16,
1059,
16,
374,
16,
8043,
16,
4711,
18,
27,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
712,
67,
8052,
67,
1132,
12,
2890,
4672,
365,
18,
11231,
5812,
12,
2890,
18,
712,
18,
8052,
9334,
261,
6976,
24,
16,
2468,
16,
1059,
16,
374,
16,
8043,
16,
4711,
18,
27,
... |
commands = filter(lambda element: element.startswith('_cmd_'), | commands = filter(self._is_cmd_function, | def _get_cmd_list(self): """Returns a list of all known commands""" # find all command functions commands = filter(lambda element: element.startswith('_cmd_'), ConsoleIFace.__dict__.keys()) # strip the prefix from command functions return map(lambda element: element[5:], commands) | f24f12b48a571183e88090f9b172a58c7473e935 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2215/f24f12b48a571183e88090f9b172a58c7473e935/tel.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
4172,
67,
1098,
12,
2890,
4672,
3536,
1356,
279,
666,
434,
777,
4846,
4364,
8395,
468,
1104,
777,
1296,
4186,
4364,
273,
1034,
12,
2890,
6315,
291,
67,
4172,
67,
915,
16,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
4172,
67,
1098,
12,
2890,
4672,
3536,
1356,
279,
666,
434,
777,
4846,
4364,
8395,
468,
1104,
777,
1296,
4186,
4364,
273,
1034,
12,
2890,
6315,
291,
67,
4172,
67,
915,
16,... |
def TestUI(self): | def TestUIAll(self): | def TestUI(self): if not self._options.no_reinstrument: instrumentation_error = self.InstrumentDll() if instrumentation_error: return instrumentation_error return self.ScriptedTest("chrome", "chrome.exe", "ui_tests", ["ui_tests.exe", "--single-process", "--ui-test-timeout=180000", "--ui-test-action-timeout=80000", "--ui-test-action-max-timeout=180000", "--ui-test-sleep-timeout=40000"], multi=True) | b9ff2d73077956314c09e479fe77df4670c986df /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9392/b9ff2d73077956314c09e479fe77df4670c986df/chrome_tests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7766,
5370,
1595,
12,
2890,
4672,
309,
486,
365,
6315,
2116,
18,
2135,
67,
266,
22818,
30,
30668,
67,
1636,
273,
365,
18,
19228,
40,
2906,
1435,
309,
30668,
67,
1636,
30,
327,
30668,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7766,
5370,
1595,
12,
2890,
4672,
309,
486,
365,
6315,
2116,
18,
2135,
67,
266,
22818,
30,
30668,
67,
1636,
273,
365,
18,
19228,
40,
2906,
1435,
309,
30668,
67,
1636,
30,
327,
30668,
6... |
print "\t%s: %s," % (`record.doc[0]`, `str(record.clsid)`) print "}" print | print >> stream, "\t%s: %s," % (`record.doc[0]`, `str(record.clsid)`) print >> stream, "}" print >> stream | print "\t###%s: %s, # Typedef disabled because it doesn't have a non-null GUID" % (`record.doc[0]`, `str(record.clsid)`) | cfd84d318292d699ca53899ee423b08bfb797445 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/992/cfd84d318292d699ca53899ee423b08bfb797445/genpy.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1172,
1548,
88,
13151,
9,
87,
30,
738,
87,
16,
468,
1412,
536,
5673,
2724,
518,
3302,
1404,
1240,
279,
1661,
17,
2011,
20322,
6,
738,
21863,
3366,
18,
2434,
63,
20,
65,
9191,
1375,
701,
12... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1172,
1548,
88,
13151,
9,
87,
30,
738,
87,
16,
468,
1412,
536,
5673,
2724,
518,
3302,
1404,
1240,
279,
1661,
17,
2011,
20322,
6,
738,
21863,
3366,
18,
2434,
63,
20,
65,
9191,
1375,
701,
12... |
self('import "%s" %s'%(os.path.abspath(filename),options)) | self('import %s "%s"'%(options, os.path.abspath(filename))) | def import_patch(self, filename, options=''): """ Import an ordered set of patches from patch file, i.e., a plain text file created using the export command. | 1d69a67479f8640d407ca6259fe49a7a226a0153 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9890/1d69a67479f8640d407ca6259fe49a7a226a0153/hg.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1930,
67,
2272,
12,
2890,
16,
1544,
16,
702,
2218,
11,
4672,
3536,
6164,
392,
5901,
444,
434,
16482,
628,
4729,
585,
16,
277,
18,
73,
12990,
279,
7351,
977,
585,
2522,
1450,
326,
3359,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1930,
67,
2272,
12,
2890,
16,
1544,
16,
702,
2218,
11,
4672,
3536,
6164,
392,
5901,
444,
434,
16482,
628,
4729,
585,
16,
277,
18,
73,
12990,
279,
7351,
977,
585,
2522,
1450,
326,
3359,... |
print 'on_advance: %d %s' % (self.frame, mode) | def on_advance(self, mode): if not hasattr(self.atoms,'__iter__'): return | 4925c62c244990e67f4a414e6c2b5f4ab6334b52 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8914/4925c62c244990e67f4a414e6c2b5f4ab6334b52/atomeye.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
67,
361,
5882,
12,
2890,
16,
1965,
4672,
309,
486,
3859,
12,
2890,
18,
14937,
11189,
972,
2165,
7250,
4672,
327,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
67,
361,
5882,
12,
2890,
16,
1965,
4672,
309,
486,
3859,
12,
2890,
18,
14937,
11189,
972,
2165,
7250,
4672,
327,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... | |
"""Load in blender""" if self.confirmSave(): return if self.isNew(): return | """Load into blender""" | def load_in_blender(self): """Load in blender""" if self.confirmSave(): return if self.isNew(): return if self.parentPanel.checkBlender(): child = self.app.childActive answer = child.confirmSave('Only saved contents will be loaded in Blender.') if answer: import Blender #first: let's remove previous copies of this file from Blender's Texts for t in filter(lambda x: x.filename == child.fileName, Blender.Text.Get()) : Blender.Text.unlink(t) #second: let's load the just saved file. Blender.Text.Load(child.fileName) #BEWARE: this text will not be the selected text in Blender - #you still have to select it as the actual. self.setStatus(("File successfully loaded as a Blender's Text Editor item, named '%s'" % os.path.basename(child.fileName))) | fde13fc113d3937bf67c3b591d69f9a1b60bcece /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/2464/fde13fc113d3937bf67c3b591d69f9a1b60bcece/Child.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
67,
267,
67,
3083,
2345,
12,
2890,
4672,
3536,
2563,
1368,
2811,
2345,
8395,
309,
365,
18,
2938,
5537,
18,
1893,
4802,
2345,
13332,
1151,
282,
273,
365,
18,
2910,
18,
3624,
3896,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
67,
267,
67,
3083,
2345,
12,
2890,
4672,
3536,
2563,
1368,
2811,
2345,
8395,
309,
365,
18,
2938,
5537,
18,
1893,
4802,
2345,
13332,
1151,
282,
273,
365,
18,
2910,
18,
3624,
3896,
... |
self._error_check(line, out) | if error_check: self._error_check(line, out) | def _eval_line(self, line, reformat=True, allow_use_file=False, wait_for_prompt=True): if len(line) == 0: return '' line = line.rstrip() if line[-1] != '$' and line[-1] != ';': line += ';' | 27b52778f34947a11ed11c466f5d54d57304fd47 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9890/27b52778f34947a11ed11c466f5d54d57304fd47/maxima.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8622,
67,
1369,
12,
2890,
16,
980,
16,
283,
2139,
33,
5510,
16,
1699,
67,
1202,
67,
768,
33,
8381,
16,
2529,
67,
1884,
67,
13325,
33,
5510,
4672,
309,
562,
12,
1369,
13,
422,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8622,
67,
1369,
12,
2890,
16,
980,
16,
283,
2139,
33,
5510,
16,
1699,
67,
1202,
67,
768,
33,
8381,
16,
2529,
67,
1884,
67,
13325,
33,
5510,
4672,
309,
562,
12,
1369,
13,
422,
... |
if _v03: | if _noInstanceState: | def __delattr__( klas, attrname): 'deleting attributes of the klas comes here - try restore the StaticType' try: st = klas.StaticType[ attrname] except KeyError: _Base.__metaclass__.__delattr__( klas, attrname) else: setattr( klas, attrname, st) | 11b821f504d2225414d6487ef5b3b2c61c246af6 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/10808/11b821f504d2225414d6487ef5b3b2c61c246af6/sa2static.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
3771,
1747,
972,
12,
417,
9521,
16,
1604,
529,
4672,
296,
19003,
310,
1677,
434,
326,
417,
9521,
14535,
2674,
300,
775,
5217,
326,
10901,
559,
11,
775,
30,
384,
273,
417,
9521,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
3771,
1747,
972,
12,
417,
9521,
16,
1604,
529,
4672,
296,
19003,
310,
1677,
434,
326,
417,
9521,
14535,
2674,
300,
775,
5217,
326,
10901,
559,
11,
775,
30,
384,
273,
417,
9521,
1... |
print '--Call--' self.interaction(frame, None) | if self.stop_here(frame): print '--Call--' self.interaction(frame, None) | def user_call(self, frame, argument_list): """This method is called when there is the remote possibility that we ever need to stop in this function.""" print '--Call--' self.interaction(frame, None) | 01eb85cc0324a42160b676ee31a1a707bda579ee /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/01eb85cc0324a42160b676ee31a1a707bda579ee/pdb.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
729,
67,
1991,
12,
2890,
16,
2623,
16,
1237,
67,
1098,
4672,
3536,
2503,
707,
353,
2566,
1347,
1915,
353,
326,
2632,
25547,
716,
732,
14103,
1608,
358,
2132,
316,
333,
445,
12123,
1172,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
729,
67,
1991,
12,
2890,
16,
2623,
16,
1237,
67,
1098,
4672,
3536,
2503,
707,
353,
2566,
1347,
1915,
353,
326,
2632,
25547,
716,
732,
14103,
1608,
358,
2132,
316,
333,
445,
12123,
1172,
... |
if not (os.path.exists(dir_input)): print "No input files!" exit if (not os.path.exists(dir_output)): os.mkdir(dir_output) filenames_list = GetFilenames(dir_input) if not ((len(filenames_list[id_articles]) > 0) and (len(filenames_list[id_bodies]) > 0)): print "No articles!" exit if (len(filenames_list[id_announcements]) > 0): list_announcements = ReadAnnouncements(filenames_list[id_announcements]) else: list_announcements = [] if (len(filenames_list[id_events]) > 0): list_events = ReadEvents(filenames_list[id_events]) else: list_events = [] if (len(filenames_list[id_birthdays]) > 0): list_birthdays = ReadBirthdays(filenames_list[id_birthdays]) else: list_birthdays = [] if (len(filenames_list[id_hails]) > 0): list_hails = ReadHails(filenames_list[id_hails]) else: list_hails = [] list_articles = ReadArticles(filenames_list[id_articles], filenames_list[id_bodies]) xml_string = '<?xml version="1.0" encoding="utf-8"?>\n<chatter>\n <issue date="' + date_long + '" url="' + date_short + '">\n' xml_string = xml_string + WriteArticles(" ", 2, list_articles) xml_string = xml_string + WriteFeatures(" ", 2, list_announcements, list_events, list_birthdays, list_hails) xml_string = xml_string + """ </issue>\n</chatter>""" f_out = open(dir_output + file_output, "w") print >> f_out, xml_string f_out.close() | if (__name__ == '__main__'): if not (os.path.exists(dir_input)): print "No input files!" sys.exit(0) if (not os.path.exists(dir_output)): os.mkdir(dir_output) filenames_list = GetFilenames(dir_input) if not ((len(filenames_list[id_articles]) > 0) and (len(filenames_list[id_bodies]) > 0)): print "No articles!" sys.exit(0); if (not SetDates()): sys.exit(0); if (len(filenames_list[id_announcements]) > 0): list_announcements = ReadAnnouncements(filenames_list[id_announcements]) else: list_announcements = [] if (len(filenames_list[id_events]) > 0): list_events = ReadEvents(filenames_list[id_events]) else: list_events = [] if (len(filenames_list[id_birthdays]) > 0): list_birthdays = ReadBirthdays(filenames_list[id_birthdays]) else: list_birthdays = [] if (len(filenames_list[id_hails]) > 0): list_hails = ReadHails(filenames_list[id_hails]) else: list_hails = [] list_articles = ReadArticles(filenames_list[id_articles], filenames_list[id_bodies]) xml_string = '<?xml version="1.0" encoding="utf-8"?>\n<chatter>\n <issue date="' + date_long + '" url="' + date_short + '">\n' xml_string = xml_string + WriteArticles(" ", 2, list_articles) xml_string = xml_string + WriteFeatures(" ", 2, list_announcements, list_events, list_birthdays, list_hails) xml_string = xml_string + """ </issue>\n</chatter>""" f_out = open(dir_output + file_output, "w") print >> f_out, xml_string f_out.close() | def WriteFeatures(indent, lvl, announcement_list, event_list, birthday_list, hail_list): # avoid writing empty tags: if (len(announcement_list) + len(event_list) + len(birthday_list) + len(hail_list) == 0): return '' # write what we've got! out_str = indent*(lvl) + "<features>\n" # avoid writing empty tags: if (len(announcement_list) > 0): # fill in announcements out_str = out_str + indent*(lvl+1) + '<feature id="announcements">\n' out_str = out_str + indent*(lvl+2) + '<title>Announcements</title>\n' for announcement in announcement_list: out_str = out_str + indent*(lvl+2) + '<announcement id="' + announcement[0] + '">\n' out_str = out_str + indent*(lvl+3) + '<title>' + announcement[1] + '</title>\n' out_str = out_str + indent*(lvl+3) + '<body>\n' for body in announcement[2]: out_str = out_str + indent*(lvl+4) + '<p>' + body + '</p>\n' out_str = out_str + indent*(lvl+3) + '</body>\n' out_str = out_str + indent*(lvl+2) + '</announcement>\n' out_str = out_str + indent*(lvl+1) + '</feature>\n' # avoid writing empty tags: if (len(event_list) > 0): # fill in events # (Treat as rough draft, unless the events.txt has been pre-formated) out_str = out_str + indent*(lvl+1) + '<feature id="campus-events">\n' out_str = out_str + indent*(lvl+2) + '<title>Campus Events</title>\n' for date_group in event_list: out_str = out_str + indent*(lvl+2) + '<date-group date="' + date_group[0] + '">\n' for event in date_group[1]: out_str = out_str + indent*(lvl+3) + '<event>\n' out_str = out_str + indent*(lvl+4) + '<title>' + event[0] + '</title>\n' for index in [1,2,3,4,5]: item = 1 while item < len(event[index]): tag_o = '<' + event[index][0] + '>' tag_c = '</' + event[index][0] + '>\n' out_str = out_str + indent*(lvl+4) + tag_o + event[index][item] + tag_c item = item + 1 out_str = out_str + indent*(lvl+3) + '</event>\n' out_str = out_str + indent*(lvl+2) + '</date-group>\n' out_str = out_str + indent*(lvl+1) + '</feature>\n' # avoid writing empty tags: if (len(birthday_list) > 0): # fill in birthdays out_str = out_str + indent*(lvl+1) + '<feature id="birthdays" month="' + month + '">\n' out_str = out_str + indent*(lvl+2) + '<title>' + month + ' Birthdays</title>\n' out_str = out_str + indent*(lvl+2) + '<birthday-list>\n' for bday in birthday_list: out_str = out_str + indent*(lvl+3) + '<birthday date="' + bday[0] + '" names="' + bday[1] + '" />\n' out_str = out_str + indent*(lvl+2) + '</birthday-list>\n' out_str = out_str + indent*(lvl+1) + '</feature>\n' # avoid writing empty tags: if (len(hail_list) > 0): # fill in hail-and-farewells out_str = out_str + indent*(lvl+1) + '<feature id="hail-and-farewell">\n' out_str = out_str + indent*(lvl+2) + '<title>Hail And Farewell</title>\n' for hail in hail_list: out_str = out_str + indent*(lvl+2) + '<people for="' + hail[0] + '">\n' for person in hail[1]: out_str = out_str + indent*(lvl+3) + '<person name="' + person[0] + '" department="' + person[1] + '" position="' + person[2] + '" />\n' out_str = out_str + indent*(lvl+2) + '</people>\n' out_str = out_str + indent*(lvl+1) + '</feature>\n' # finish off the features closing tag out_str = out_str + indent*(lvl) + '</features>\n' return out_str | 3224681788f82607070ad977de5b5fb8c1c3a649 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/4025/3224681788f82607070ad977de5b5fb8c1c3a649/chatter_aio.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2598,
8696,
12,
9355,
16,
14864,
16,
25199,
67,
1098,
16,
871,
67,
1098,
16,
29950,
67,
1098,
16,
366,
671,
67,
1098,
4672,
468,
4543,
7410,
1008,
2342,
30,
309,
261,
1897,
12,
1072,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2598,
8696,
12,
9355,
16,
14864,
16,
25199,
67,
1098,
16,
871,
67,
1098,
16,
29950,
67,
1098,
16,
366,
671,
67,
1098,
4672,
468,
4543,
7410,
1008,
2342,
30,
309,
261,
1897,
12,
1072,
... |
elif PROCESSOR in ['x86', 'Power Macintosh']: algorithm = 'pari' | def number_of_partitions(n,k=None, algorithm='default'): r""" Returns the size of partitions_list(n,k). INPUT: n -- an integer k -- (default: None); if specified, instead returns the cardinality of the set of all (unordered) partitions of the positive integer n into sums with k summands. algorithm -- (default: 'default') 'default' -- if k is given use Gap. Otherwise, on x86 when n > 3000, use 'bober' On non x86 use 'pari'. 'bober' -- use Jonathon Bober's implementation 'gap' -- use GAP (VERY *slow*) 'pari' -- use PARI. Speed seems the same as GAP until $n$ is in the thousands, in which case PARI is faster. *But* PARI has a bug, e.g., on 64-bit Linux PARI-2.3.2 outputs numbpart(147007)%1000 as 536, but it should be 533!. So do not use this option. IMPLEMENTATION: Wraps GAP's NrPartitions or PARI's numbpart function. Use the function \code{partitions(n)} to return a generator over all partitions of $n$. It is possible to associate with every partition of the integer n a conjugacy class of permutations in the symmetric group on n points and vice versa. Therefore p(n) = NrPartitions(n) is the number of conjugacy classes of the symmetric group on n points. EXAMPLES: sage: v = list(partitions(5)); v [(1, 1, 1, 1, 1), (1, 1, 1, 2), (1, 2, 2), (1, 1, 3), (2, 3), (1, 4), (5,)] sage: len(v) 7 sage: number_of_partitions(5, algorithm='gap') 7 sage: number_of_partitions(5, algorithm='pari') 7 sage: number_of_partitions(5, algorithm='bober') 7 The input must be a nonnegative integer or a ValueError is raised. sage: number_of_partitions(-5) Traceback (most recent call last): ... ValueError: n (=-5) must be a nonnegative integer sage: number_of_partitions(10,2) 5 sage: number_of_partitions(10) 42 sage: number_of_partitions(3) 3 sage: number_of_partitions(10) 42 sage: number_of_partitions(3, algorithm='pari') 3 sage: number_of_partitions(10, algorithm='pari') 42 sage: number_of_partitions(40) 37338 sage: number_of_partitions(100) 190569292 sage: number_of_partitions(100000) 27493510569775696512677516320986352688173429315980054758203125984302147328114964173055050741660736621590157844774296248940493063070200461792764493033510116079342457190155718943509725312466108452006369558934464248716828789832182345009262853831404597021307130674510624419227311238999702284408609370935531629697851569569892196108480158600569421098519 A generating function for p(n) is given by the reciprocal of Euler's function: \[ \sum_{n=0}^\infty p(n)x^n = \prod_{k=1}^\infty \left(\frac {1}{1-x^k} \right). \] We use SAGE to verify that the first several coefficients do instead agree: sage: q = PowerSeriesRing(QQ, 'q', default_prec=9).gen() sage: prod([(1-q^k)^(-1) for k in range(1,9)]) ## partial product of 1 + q + 2*q^2 + 3*q^3 + 5*q^4 + 7*q^5 + 11*q^6 + 15*q^7 + 22*q^8 + O(q^9) sage: [number_of_partitions(k) for k in range(2,10)] [2, 3, 5, 7, 11, 15, 22, 30] REFERENCES: http://en.wikipedia.org/wiki/Partition_%28number_theory%29 TESTS: sage: n = 500 + randint(0,500) sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True sage: n = 1500 + randint(0,1500) sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True sage: n = 1000000 + randint(0,1000000) sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True sage: n = 1000000 + randint(0,1000000) sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True sage: n = 1000000 + randint(0,1000000) sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True sage: n = 1000000 + randint(0,1000000) sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True sage: n = 1000000 + randint(0,1000000) sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True sage: n = 1000000 + randint(0,1000000) sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True sage: n = 100000000 + randint(0,100000000) # takes a long time sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True sage: n = 100000000 + randint(0,100000000) # takes a long time sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True sage: n = 100000000 + randint(0,100000000) # takes a long time sage: number_of_partitions( n - (n % 385) + 369) % 385 == 0 True Another consistency test for n up to 500: sage: len([n for n in [1..500] if number_of_partitions(n) != number_of_partitions(n,algorithm='pari')]) 0 """ n = ZZ(n) if n < 0: raise ValueError, "n (=%s) must be a nonnegative integer"%n elif n == 0: return ZZ(1) global first_warning PROCESSOR = os.uname()[-1] bober_is_good = 'x86' in PROCESSOR if k is not None: algorithm = 'gap' elif algorithm == 'default': if bober_is_good: algorithm = 'bober' elif PROCESSOR in ['x86', 'Power Macintosh']: algorithm = 'pari' else: algorithm = 'gap' if algorithm == 'gap': if k is None: ans=gap.eval("NrPartitions(%s)"%(ZZ(n))) else: ans=gap.eval("NrPartitions(%s,%s)"%(ZZ(n),ZZ(k))) return ZZ(ans) if k is not None: raise ValueError, "only the GAP algorithm works if k is specified." if algorithm == 'bober': if not bober_is_good: if first_warning: print "*WARNING*: bober's implementation is broken on this platform or this size of n." first_warning=False return partitions_ext.number_of_partitions(n) elif algorithm == 'pari': if n > 3000 and 'x86_64' in PROCESSOR: if first_warning: print "*WARNING*: Pari's numbpart is very buggy on x86_64 (fixed in svn, so don't report to pari-dev)" first_warning = False return ZZ(pari(ZZ(n)).numbpart()) raise ValueError, "unknown algorithm '%s'"%algorithm | ea6709700b2d3afb6cf9a2244cb7740a082689cc /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9890/ea6709700b2d3afb6cf9a2244cb7740a082689cc/combinat.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1300,
67,
792,
67,
21275,
12,
82,
16,
79,
33,
7036,
16,
4886,
2218,
1886,
11,
4672,
436,
8395,
2860,
326,
963,
434,
10060,
67,
1098,
12,
82,
16,
79,
2934,
225,
12943,
30,
290,
1493,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1300,
67,
792,
67,
21275,
12,
82,
16,
79,
33,
7036,
16,
4886,
2218,
1886,
11,
4672,
436,
8395,
2860,
326,
963,
434,
10060,
67,
1098,
12,
82,
16,
79,
2934,
225,
12943,
30,
290,
1493,
... | |
obj.methods.append(self._parse_function(method, Function)) | obj.methods.append(self._parse_function_common(method, Function)) | def _parse_object_interface(self, node): ctor_args = [node.attrib['name'], node.attrib.get('parent'), node.attrib[_glibns('type-name')], node.attrib[_glibns('get-type')]] if node.tag == _corens('interface'): klass = GLibInterface elif node.tag == _corens('class'): klass = GLibObject is_abstract = node.attrib.get('abstract') is_abstract = is_abstract and is_abstract != '0' ctor_args.append(is_abstract) else: raise AssertionError(node) | 34db824e4656506e5dfa8df3d0446950b5a48e65 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/2770/34db824e4656506e5dfa8df3d0446950b5a48e65/girparser.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2670,
67,
1612,
67,
5831,
12,
2890,
16,
756,
4672,
15120,
67,
1968,
273,
306,
2159,
18,
14588,
3292,
529,
17337,
756,
18,
14588,
18,
588,
2668,
2938,
19899,
756,
18,
14588,
63,
67... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2670,
67,
1612,
67,
5831,
12,
2890,
16,
756,
4672,
15120,
67,
1968,
273,
306,
2159,
18,
14588,
3292,
529,
17337,
756,
18,
14588,
18,
588,
2668,
2938,
19899,
756,
18,
14588,
63,
67... |
[Ideal (y + 1, z^2 + 1) of Multivariate Polynomial Ring in x, y, z over Rational Field, Ideal (z^2 - y, y*z + 2, y^2 + 2*z) of Multivariate Polynomial Ring in x, y, z over Rational Field] | [Ideal (z^2 - y, y*z + 2, y^2 + 2*z) of Multivariate Polynomial Ring in x, y, z over Rational Field, Ideal (y + 1, z^2 + 1) of Multivariate Polynomial Ring in x, y, z over Rational Field] | def associated_primes(self, algorithm='sy'): """ EXAMPLES: sage: R.<x,y,z> = PolynomialRing(QQ, 3) sage: p = z^2 + 1; q = z^3 + 2 sage: I = (p*q^2, y-z^2)*R sage: I.associated_primes() [Ideal (y + 1, z^2 + 1) of Multivariate Polynomial Ring in x, y, z over Rational Field, Ideal (z^2 - y, y*z + 2, y^2 + 2*z) of Multivariate Polynomial Ring in x, y, z over Rational Field] """ return [P for _,P in self.complete_primary_decomposition(algorithm)] | 46f3f42ba21e62bc919b91320ff5b4482653b732 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9890/46f3f42ba21e62bc919b91320ff5b4482653b732/multi_polynomial_ideal.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3627,
67,
683,
4485,
12,
2890,
16,
4886,
2218,
9009,
11,
4672,
3536,
5675,
8900,
11386,
30,
272,
410,
30,
534,
22782,
92,
16,
93,
16,
94,
34,
273,
6730,
13602,
10369,
12,
53,
53,
16,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3627,
67,
683,
4485,
12,
2890,
16,
4886,
2218,
9009,
11,
4672,
3536,
5675,
8900,
11386,
30,
272,
410,
30,
534,
22782,
92,
16,
93,
16,
94,
34,
273,
6730,
13602,
10369,
12,
53,
53,
16,... |
elif not getattr(obj, doby, list()): | elif not getattr(obj, encode_page(doby), list()): | def node_filters(self, obj): # Node filters for filt, doby in [(self.filterorder, self.orderby), (self.filtercolor, self.colorby)]: | 0f239dff4cab548aefd29c73f1bfec5bbe6960b3 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/888/0f239dff4cab548aefd29c73f1bfec5bbe6960b3/ShowGraph.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
756,
67,
6348,
12,
2890,
16,
1081,
4672,
468,
2029,
3415,
364,
18553,
16,
741,
1637,
316,
306,
12,
2890,
18,
2188,
1019,
16,
365,
18,
25896,
3631,
261,
2890,
18,
2188,
3266,
16,
365,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
756,
67,
6348,
12,
2890,
16,
1081,
4672,
468,
2029,
3415,
364,
18553,
16,
741,
1637,
316,
306,
12,
2890,
18,
2188,
1019,
16,
365,
18,
25896,
3631,
261,
2890,
18,
2188,
3266,
16,
365,
... |
<td width="80%%"><textarea cols=100 rows=30 name="elfidesc" wrap="nowarp">%(el_fidesc)s</textarea></td> | <td width="80%%"><textarea cols="100" rows="30" name="elfidesc" wrap="nowarp">%(el_fidesc)s</textarea></td> | def tmpl_display_addelementform(self, elname="", elmarccode="", eltype="", elsize="", elrows="", elcols="", elmaxlength="", elval="", elfidesc="", elmodifytext="", elcookie="", elcd="", elmd="", perform_act="elementadd", user_msg="", el_use_tuple="" ): """Display Web form used to add a new element to the database @param elname: element name @param elmarccode: marc code of element @param eltype: element type (e.g. text, user-defined, etc) @param elsize: element size (e.g. for text input element) @param elrows: number of rows (e.g. for textarea element) @param elcols: number of columns (e.g. for textarea element) @param elmaxlength: maximum length of a text input field @param elval: value of element (e.g. for text input element) @param elfidesc: description for element (e.g. for user-defined element) @param elmodifytext: element's modification text @param elcookie: Is a cookie to be set for the element? @param elcd: creation date of element @param elmd: last modification date of element @param user_msg: Any message to be displayed on screen, such as a status report for the last task, etc. @param el_use_tuple: @return: HTML page body. """ ## First, get a rough preview of the element: output = "" etypes = {"D" : "User Defined Input", "F" : "File Input", "H" : "Hidden Input", "I" : "Text Input", \ "R" : "Response", "S" : "Select Box", "T" : "Text Area Element"} etypeids = etypes.keys() etypeids.sort() body_content = "" | e760c1fc3e16767c2b47a9e2f94b5484506a8af7 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12027/e760c1fc3e16767c2b47a9e2f94b5484506a8af7/websubmitadmin_templates.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10720,
67,
5417,
67,
1289,
2956,
687,
12,
2890,
16,
415,
529,
1546,
3113,
415,
3684,
71,
710,
1546,
3113,
415,
723,
1546,
3113,
415,
1467,
1546,
3113,
415,
3870,
1546,
3113,
415,
6842,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10720,
67,
5417,
67,
1289,
2956,
687,
12,
2890,
16,
415,
529,
1546,
3113,
415,
3684,
71,
710,
1546,
3113,
415,
723,
1546,
3113,
415,
1467,
1546,
3113,
415,
3870,
1546,
3113,
415,
6842,
... |
pull_url='%s/sage-%s'%(SAGE_INCOMING_SERVER, temp_in_branch_name), push_url='%s/sage-%s'%(SAGE_OUTGOING_SERVER, temp_out_branch_name), | pull_url='%s/sage-%s/'%(SAGE_INCOMING_SERVER, temp_in_branch_name), push_url='%s/sage-%s/'%(SAGE_OUTGOING_SERVER, temp_out_branch_name), | def bundle(self, filename, options='', url=None, base=None, to=None): r""" Create an hg changeset bundle with the given filename against the repository at the given url (which is by default the 'official' Sage repository, unless push_url() is changed in a setup file). If you have internet access, it's best to just do ``hg_sage.bundle(filename)``. If you don't find a revision r that you and the person unbundling both have (by looking at ``hg_sage.log()``), then do ``hg_sage.bundle(filename, base=r)``. Use self.inspect('file.bundle') to inspect the resulting bundle. This is a file that you should probably send to William Stein (wstein@gmail.com), post to a web page, or send to sage-devel. It will be written to the current directory. INPUT: - ``filename`` - output file in which to put bundle - ``options`` - pass to hg - ``url`` - url to bundle against (default: SAGE_SERVER, or push_url()) - ``base`` - a base changeset revision number to bundle against (doesn't require internet access) """ if not base is None: url = '' options = '--base=%s %s'%(int(base), options) if url is None: url = self.__push_url | b6cedfc3ec097dfea0a4c9b4ec56a612b77a1fd2 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/b6cedfc3ec097dfea0a4c9b4ec56a612b77a1fd2/hg.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3440,
12,
2890,
16,
1544,
16,
702,
2218,
2187,
880,
33,
7036,
16,
1026,
33,
7036,
16,
358,
33,
7036,
4672,
436,
8395,
1788,
392,
22576,
22463,
3440,
598,
326,
864,
1544,
5314,
326,
335... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3440,
12,
2890,
16,
1544,
16,
702,
2218,
2187,
880,
33,
7036,
16,
1026,
33,
7036,
16,
358,
33,
7036,
4672,
436,
8395,
1788,
392,
22576,
22463,
3440,
598,
326,
864,
1544,
5314,
326,
335... |
val = unpack("I", key)[0] | val = unpack("I", key)[0] | def read(self): # Never fails, attempts to recover as much as possible if self.is_read: return 1 self.is_read = 1 dbpath = self._getDBPath() if not os.path.isdir(dbpath): return 1 try: db = bsddb.hashopen(os.path.join(dbpath, "Packages"), "r") except bsddb.error: return 1 for key in db.keys(): rpmio = RpmFileIO(self.config, "dummy") pkg = package.RpmPackage(self.config, "dummy") data = db[key] try: val = unpack("I", key)[0] # FIXME: native endian? except struct.error: self.config.printError("Invalid key %s in rpmdb" % repr(key)) continue | e54b9fe621ece3e30d40352986bb0a4c529dbe38 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1143/e54b9fe621ece3e30d40352986bb0a4c529dbe38/io.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
855,
12,
2890,
4672,
468,
24496,
6684,
16,
7531,
358,
5910,
487,
9816,
487,
3323,
309,
365,
18,
291,
67,
896,
30,
327,
404,
365,
18,
291,
67,
896,
273,
404,
1319,
803,
273,
365,
6315... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
855,
12,
2890,
4672,
468,
24496,
6684,
16,
7531,
358,
5910,
487,
9816,
487,
3323,
309,
365,
18,
291,
67,
896,
30,
327,
404,
365,
18,
291,
67,
896,
273,
404,
1319,
803,
273,
365,
6315... |
SMTPRecipientsRefused The server rejected for ALL recipients | SMTPRecipientsRefused The server rejected ALL recipients | def sendmail(self, from_addr, to_addrs, msg, mail_options=[], rcpt_options=[]): """This command performs an entire mail transaction. | d25c1b73d2c8ab561476de849026aef7bd25808e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/d25c1b73d2c8ab561476de849026aef7bd25808e/smtplib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1366,
4408,
12,
2890,
16,
628,
67,
4793,
16,
358,
67,
23305,
16,
1234,
16,
4791,
67,
2116,
22850,
6487,
4519,
337,
67,
2116,
33,
8526,
4672,
3536,
2503,
1296,
11199,
392,
7278,
4791,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1366,
4408,
12,
2890,
16,
628,
67,
4793,
16,
358,
67,
23305,
16,
1234,
16,
4791,
67,
2116,
22850,
6487,
4519,
337,
67,
2116,
33,
8526,
4672,
3536,
2503,
1296,
11199,
392,
7278,
4791,
2... |
% offset | % offset_time | def get_moved_pages_redirects(self): '''generate redirects to recently-moved pages''' # this will run forever, until user interrupts it import datetime | be7a5bd35b2d089fafb83234ff3b989065d79878 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/4404/be7a5bd35b2d089fafb83234ff3b989065d79878/redirect.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
81,
9952,
67,
7267,
67,
22715,
12,
2890,
4672,
9163,
7163,
18064,
358,
19907,
17,
81,
9952,
4689,
26418,
468,
333,
903,
1086,
21238,
16,
3180,
729,
509,
30209,
1092,
518,
1930,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
81,
9952,
67,
7267,
67,
22715,
12,
2890,
4672,
9163,
7163,
18064,
358,
19907,
17,
81,
9952,
4689,
26418,
468,
333,
903,
1086,
21238,
16,
3180,
729,
509,
30209,
1092,
518,
1930,
... |
trace("-> stop") | def run(self, data = None, state=initialState): ''' This executes the current flow, given empty starting data and the default initial state. ''' trace("running flow") stack = self.stack if not(stack): stack.append((state, data, None, None, None)) while stack: (state, data, finish, itr, stop) = stack[-1] if finish: finish() stack.pop() continue elif itr: trace("-> iterator") try: data = itr.next() if data is stop: continue if not state: (state, data) = data if data is stop: continue except StopIteration: trace("-> stop") stack.pop() continue except YieldIteration: trace("-> yield") from twisted.internet import reactor reactor.callLater(self.waitInterval, self.run) return else: stack.pop() tpl = self._lookup(state) (func, state, isIterable, finish, failure, stop) = tpl if finish: stack.append((state, None, finish, None, stop)) data = func(data) trace("-> item") if isIterable: if data: data = iter(data) self.stack.append((state, None, None,data, stop)) else: if data is stop: continue if not state: (state, data) = data if data is stop: continue stack.append((state, data, None, None, stop)) trace("end flow") | 4cf3103eb894bde67fe5e1250725e8ef11de7f5a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12595/4cf3103eb894bde67fe5e1250725e8ef11de7f5a/flow.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
16,
501,
273,
599,
16,
919,
33,
6769,
1119,
4672,
9163,
1220,
11997,
326,
783,
4693,
16,
864,
1008,
5023,
501,
471,
326,
805,
2172,
919,
18,
9163,
2606,
2932,
8704,
469... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
16,
501,
273,
599,
16,
919,
33,
6769,
1119,
4672,
9163,
1220,
11997,
326,
783,
4693,
16,
864,
1008,
5023,
501,
471,
326,
805,
2172,
919,
18,
9163,
2606,
2932,
8704,
469... | |
self.prefs_tab.setCurrentIndex(8) | self.prefs_tab.setCurrentIndex(9) | def showDialog(self, pagename = 'General'): """ Display the Preferences dialog with page I{pagename}. @param pagename: name of the Preferences page. Default is "General". @type pagename: text """ | 7916a3f2a44e2d8e6b274c1dee72b20aaaef2a1c /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/11221/7916a3f2a44e2d8e6b274c1dee72b20aaaef2a1c/UserPrefs.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2405,
6353,
12,
2890,
16,
4262,
1069,
273,
296,
12580,
11,
4672,
3536,
9311,
326,
28310,
6176,
598,
1363,
467,
95,
9095,
1069,
5496,
225,
632,
891,
4262,
1069,
30,
508,
434,
326,
28310,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2405,
6353,
12,
2890,
16,
4262,
1069,
273,
296,
12580,
11,
4672,
3536,
9311,
326,
28310,
6176,
598,
1363,
467,
95,
9095,
1069,
5496,
225,
632,
891,
4262,
1069,
30,
508,
434,
326,
28310,
... |
init_sage = DOT_SAGE + 'init.sage' if os.path.exists(init_sage): A[init_sage] = 0 | def attached_files(self): try: A = self.__attached except AttributeError: A = {} | 9dab5d58718f9cb88256a96ca94d9a584e47fad6 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9417/9dab5d58718f9cb88256a96ca94d9a584e47fad6/worksheet.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7495,
67,
2354,
12,
2890,
4672,
775,
30,
432,
273,
365,
16186,
24157,
1335,
6394,
30,
432,
273,
2618,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7495,
67,
2354,
12,
2890,
4672,
775,
30,
432,
273,
365,
16186,
24157,
1335,
6394,
30,
432,
273,
2618,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... | |
menu = os.path.join(MENUDIR, "all", all_menus[x]) | menu = os.path.join(MENUDIR, all_menus[x]) | def clean_menu(): """Function removing unused menu entries, usually called at the end""" all_menus = [] for y in range(db.__len__()): if not all_menus.__contains__(db[y][2]): all_menus.append(db[y][2]) # This will "save" submenus entries for x in range(menus_used.__len__()): # Preserve menu if only submenu inside for single_entry in menus_used[x].split(" "): if all_menus.__contains__(single_entry): all_menus.remove(single_entry) if all_menus.__contains__(menus_used[x]): all_menus.remove(menus_used[x]) # This just generated the list of unused menu entries. for x in range(all_menus.__len__()): menu = os.path.join(MENUDIR, "all", all_menus[x]) if not options.simulate: if os.path.exists(menu): os.rmdir(menu) else: print warn + menu + " NOT FOUND" else: print warn + menu + " NOT FOUND" | 00021f5ca3968b28768c8c7c00779b392b4173dc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7161/00021f5ca3968b28768c8c7c00779b392b4173dc/genmenu.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2721,
67,
5414,
13332,
3536,
2083,
9427,
10197,
3824,
3222,
16,
11234,
2566,
622,
326,
679,
8395,
777,
67,
26065,
273,
5378,
364,
677,
316,
1048,
12,
1966,
16186,
1897,
972,
1435,
4672,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2721,
67,
5414,
13332,
3536,
2083,
9427,
10197,
3824,
3222,
16,
11234,
2566,
622,
326,
679,
8395,
777,
67,
26065,
273,
5378,
364,
677,
316,
1048,
12,
1966,
16186,
1897,
972,
1435,
4672,
... |
omega_m = (psi(m+2)*psi(m-1)**2 - psi(m-2)*psi(m+1)**2)/(4*y) y_coord = normalize(omega_m / psi_m**3) | if m == 2: lam = (3*x**2 + A)/(2*y) y_coord = normalize(-lam*x_coord - (y - lam*x)) else: omega_m = (psi(m+2)*psi(m-1)**2 - psi(m-2)*psi(m+1)**2)/(4*y) y_coord = normalize(omega_m / psi_m**3) | def normalize(f): return Q(f.numerator()).lift() / Q(f.denominator()).lift() | 84dad5dfc790d5dacf9589fdd0cfafb884726e83 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9417/84dad5dfc790d5dacf9589fdd0cfafb884726e83/ell_generic.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3883,
12,
74,
4672,
327,
2238,
12,
74,
18,
2107,
7385,
1435,
2934,
80,
2136,
1435,
342,
2238,
12,
74,
18,
13002,
26721,
1435,
2934,
80,
2136,
1435,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3883,
12,
74,
4672,
327,
2238,
12,
74,
18,
2107,
7385,
1435,
2934,
80,
2136,
1435,
342,
2238,
12,
74,
18,
13002,
26721,
1435,
2934,
80,
2136,
1435,
2,
-100,
-100,
-100,
-100,
-100,
-10... |
marker_match = perform_regex_search_upon_line_with_pattern_list(processed_line, marker_patterns) | marker_match = perform_regex_match_upon_line_with_pattern_list(processed_line, marker_patterns) | def convert_processed_reference_line_to_marc_xml(line): """Given a processed reference line, convert it to MARC XML. @param line: (string) - the processed reference line, in which the recognised citations have been tagged. @return: (tuple) - + xml_line (string) - the reference line with all of its identified citations marked up into the various subfields. + count_misc (integer) - number of sections of miscellaneous found in the line + count_title (integer) - number of title-citations found in the line + count_reportnum (integer) - number of report numbers found in the line + count_url (integer) - number of URLs found in the line """ count_misc = count_title = count_reportnum = count_url = 0 xml_line = "" previously_cited_item = None processed_line = line.lstrip() ## 1. Extract reference line marker (e.g. [1]) from start of line and tag it: ## get patterns to identify numeration markers at the start of lines: marker_patterns = get_reference_line_numeration_marker_patterns() marker_match = perform_regex_search_upon_line_with_pattern_list(processed_line, marker_patterns) if marker_match is not None: ## found a marker: marker_val = marker_match.group(u'mark') ## trim the marker from the start of the line: processed_line = processed_line[marker_match.end():].lstrip() else: marker_val = u" " ## Now display the marker in marked-up XML: xml_line += _refextract_markup_reference_line_marker_as_marcxml(marker_val) ## 2. Loop through remaining identified segments in line and tag them into MARC XML segments: cur_misc_txt = u"" ## a marker to hold gathered miscellaneous text before a citation tag_match = sre_tagged_citation.search(processed_line) while tag_match is not None: ## found a tag - process it: tag_match_start = tag_match.start() tag_match_end = tag_match.end() tag_type = tag_match.group(1) if tag_type == "TITLE": ## This tag is an identified journal TITLE. It should be followed by VOLUME, ## YEAR and PAGE tags. cur_misc_txt += processed_line[0:tag_match_start] ## extract the title from the line: idx_closing_tag = processed_line.find(CFG_REFEXTRACT_MARKER_CLOSING_TITLE, tag_match_end) ## Sanity check - did we find a closing TITLE tag? if idx_closing_tag == -1: ## no closing </cds.TITLE> tag found - strip the opening tag and move past it processed_line = processed_line[tag_match_end:] else: ## Closing tag was found: title_text = processed_line[tag_match_end:idx_closing_tag] ## Now trim this matched title and its tags from the start of the line: processed_line = processed_line[idx_closing_tag+len(CFG_REFEXTRACT_MARKER_CLOSING_TITLE):] ## Was this title followed by the tags of recognised VOLUME, YEAR and PAGE objects? numeration_match = sre_recognised_numeration_for_title.match(processed_line) if numeration_match is not None: ## recognised numeration immediately after the title - extract it: reference_volume = numeration_match.group(2) reference_year = numeration_match.group(3) reference_page = numeration_match.group(4) ## Skip past the matched numeration in the working line: processed_line = processed_line[numeration_match.end():] if previously_cited_item is None: ## There is no previously cited item - this should be added as the previously ## cited item: previously_cited_item = { 'type' : "TITLE", 'misc_txt' : cur_misc_txt, 'title' : title_text, 'volume' : reference_volume, 'year' : reference_year, 'page' : reference_page, } ## Now empty the miscellaneous text and title components: cur_misc_txt = "" title_text = "" reference_volume = "" reference_year = "" reference_page = "" elif (previously_cited_item is not None) and \ (previously_cited_item['type'] == "REPORTNUMBER") and \ (len(cur_misc_txt.lower().replace("arxiv", "").strip(".,:;- []")) == 0): ## This TITLE belongs with the REPORT NUMBER before it - add them both into ## the same datafield tag (REPORT NUMBER first, TITLE second): prev_report_num = previously_cited_item['report_num'] prev_misc_txt = previously_cited_item['misc_txt'].lstrip(".;, ").rstrip() xml_line += \ _refextract_markup_title_followed_by_report_number_as_marcxml(title_text, reference_volume, reference_year, reference_page, prev_report_num, prev_misc_txt) ## Increment the stats counters: if len(prev_misc_txt) > 0: count_misc += 1 count_title += 1 count_reportnum += 1 ## reset the various variables: previously_cited_item = None cur_misc_txt = u"" title_text = "" reference_volume = "" reference_year = "" reference_page = "" else: ## either the previously cited item is NOT a REPORT NUMBER, or this cited TITLE ## is preceeded by miscellaneous text. In either case, the two cited objects are ## not the same and do not belong together in the same datafield. if previously_cited_item['type'] == "REPORTNUMBER": ## previously cited item was a REPORT NUMBER. ## Add previously cited REPORT NUMBER to XML string: prev_report_num = previously_cited_item['report_num'] prev_misc_txt = previously_cited_item['misc_txt'].lstrip(".;, ").rstrip() xml_line += _refextract_markup_reportnumber_as_marcxml(prev_report_num, prev_misc_txt) ## Increment the stats counters: if len(prev_misc_txt) > 0: count_misc += 1 count_reportnum += 1 elif previously_cited_item['type'] == "TITLE": ## previously cited item was a TITLE. ## Add previously cited TITLE to XML string: prev_title = previously_cited_item['title'] prev_volume = previously_cited_item['volume'] prev_year = previously_cited_item['year'] prev_page = previously_cited_item['page'] prev_misc_txt = previously_cited_item['misc_txt'].lstrip(".;, ").rstrip() xml_line += _refextract_markup_title_as_marcxml(prev_title, prev_volume, prev_year, prev_page, prev_misc_txt) ## Increment the stats counters: if len(prev_misc_txt) > 0: count_misc += 1 count_title += 1 ## Now add the current cited item into the previously cited item marker previously_cited_item = { 'type' : "TITLE", 'misc_txt' : cur_misc_txt, 'title' : title_text, 'volume' : reference_volume, 'year' : reference_year, 'page' : reference_page, } ## empty miscellaneous text cur_misc_txt = u"" title_text = "" reference_volume = "" reference_year = "" reference_page = "" else: ## No numeration was recognised after the title. Add the title into misc and carry on: cur_misc_txt += " %s" % title_text elif tag_type == "REPORTNUMBER": ## This tag is an identified institutional report number: ## Account for the miscellaneous text before the citation: cur_misc_txt += processed_line[0:tag_match_start] ## extract the institutional report-number from the line: idx_closing_tag = processed_line.find(CFG_REFEXTRACT_MARKER_CLOSING_REPORT_NUM, tag_match_end) ## Sanity check - did we find a closing report-number tag? if idx_closing_tag == -1: ## no closing </cds.REPORTNUMBER> tag found - strip the opening tag and move past this ## recognised reportnumber as it is unreliable: processed_line = processed_line[tag_match_end:] else: ## closing tag was found report_num = processed_line[tag_match_end:idx_closing_tag] ## now trim this matched institutional report-number and its tags from the start of the line: processed_line = processed_line[idx_closing_tag+len(CFG_REFEXTRACT_MARKER_CLOSING_REPORT_NUM):] ## Now, if there was a previous TITLE citation and this REPORT NUMBER citation one has no ## miscellaneous text after punctuation has been stripped, the two refer to the same object, ## so group them under the same datafield: if previously_cited_item is None: ## There is no previously cited item - this should be added as the previously ## cited item: previously_cited_item = { 'type' : "REPORTNUMBER", 'misc_txt' : "%s" % cur_misc_txt, 'report_num' : "%s" % report_num, } ## empty miscellaneous text cur_misc_txt = u"" report_num = u"" elif (previously_cited_item is not None) and \ (previously_cited_item['type'] == "TITLE") and \ (len(cur_misc_txt.lower().replace("arxiv", "").strip(".,:;- []")) == 0): ## This REPORT NUMBER belongs with the title before it - add them both into ## the same datafield tag (TITLE first, REPORT NUMBER second): prev_title = previously_cited_item['title'] prev_volume = previously_cited_item['volume'] prev_year = previously_cited_item['year'] prev_page = previously_cited_item['page'] prev_misc_txt = previously_cited_item['misc_txt'].lstrip(".;, ").rstrip() xml_line += \ _refextract_markup_title_followed_by_report_number_as_marcxml(prev_title, prev_volume, prev_year, prev_page, report_num, prev_misc_txt) ## Increment the stats counters: if len(prev_misc_txt) > 0: count_misc += 1 count_title += 1 count_reportnum += 1 ## Reset variables: previously_cited_item = None cur_misc_txt = u"" else: ## either the previously cited item is NOT a TITLE, or this cited REPORT NUMBER ## is preceeded by miscellaneous text. In either case, the two cited objects are ## not the same and do not belong together in the same datafield. if previously_cited_item['type'] == "REPORTNUMBER": ## previously cited item was a REPORT NUMBER. ## Add previously cited REPORT NUMBER to XML string: prev_report_num = previously_cited_item['report_num'] prev_misc_txt = previously_cited_item['misc_txt'].lstrip(".;, ").rstrip() xml_line += _refextract_markup_reportnumber_as_marcxml(prev_report_num, prev_misc_txt) ## Increment the stats counters: if len(prev_misc_txt) > 0: count_misc += 1 count_reportnum += 1 elif previously_cited_item['type'] == "TITLE": ## previously cited item was a TITLE. ## Add previously cited TITLE to XML string: prev_title = previously_cited_item['title'] prev_volume = previously_cited_item['volume'] prev_year = previously_cited_item['year'] prev_page = previously_cited_item['page'] prev_misc_txt = previously_cited_item['misc_txt'].lstrip(".;, ").rstrip() xml_line += _refextract_markup_title_as_marcxml(prev_title, prev_volume, prev_year, prev_page, prev_misc_txt) ## Increment the stats counters: if len(prev_misc_txt) > 0: count_misc += 1 count_title += 1 ## Now add the current cited item into the previously cited item marker previously_cited_item = { 'type' : "REPORTNUMBER", 'misc_txt' : "%s" % cur_misc_txt, 'report_num' : "%s" % report_num, } ## empty miscellaneous text cur_misc_txt = u"" report_num = u"" elif tag_type == "URL": ## This tag is an identified URL: ## Account for the miscellaneous text before the URL: cur_misc_txt += processed_line[0:tag_match_start] ## extract the URL information from within the tags in the line: idx_closing_tag = processed_line.find(CFG_REFEXTRACT_MARKER_CLOSING_URL, tag_match_end) ## Sanity check - did we find a closing URL tag? if idx_closing_tag == -1: ## no closing </cds.URL> tag found - strip the opening tag and move past it processed_line = processed_line[tag_match_end:] else: ## Closing tag was found: ## First, get the URL string from between the tags: url_string = processed_line[tag_match_end:idx_closing_tag] ## Now, get the URL description string from within the opening cds tag. E.g.: ## from <cds.URL description="abc"> get the "abc" value: opening_url_tag = processed_line[tag_match_start:tag_match_end] if opening_url_tag.find(u"""<cds.URL description=\"""") != -1: ## the description is present - extract it: ## (Stop 2 characters before the end of the string - we assume they are the ## closing characters '">'. url_descr = opening_url_tag[22:-2] else: ## There is no description - description should now be the url string: url_descr = url_string ## now trim this URL and its tags from the start of the line: processed_line = processed_line[idx_closing_tag+len(CFG_REFEXTRACT_MARKER_CLOSING_URL):] ## Build the MARC XML representation of this identified URL: if previously_cited_item is not None: ## There was a previously cited item. We must convert it to XML before we can ## convert this URL to XML: if previously_cited_item['type'] == "REPORTNUMBER": ## previously cited item was a REPORT NUMBER. ## Add previously cited REPORT NUMBER to XML string: prev_report_num = previously_cited_item['report_num'] prev_misc_txt = previously_cited_item['misc_txt'].lstrip(".;, ").rstrip() xml_line += _refextract_markup_reportnumber_as_marcxml(prev_report_num, prev_misc_txt) ## Increment the stats counters: if len(prev_misc_txt) > 0: count_misc += 1 count_reportnum += 1 elif previously_cited_item['type'] == "TITLE": ## previously cited item was a TITLE. ## Add previously cited TITLE to XML string: prev_title = previously_cited_item['title'] prev_volume = previously_cited_item['volume'] prev_year = previously_cited_item['year'] prev_page = previously_cited_item['page'] prev_misc_txt = previously_cited_item['misc_txt'].lstrip(".;, ").rstrip() xml_line += _refextract_markup_title_as_marcxml(prev_title, prev_volume, prev_year, prev_page, prev_misc_txt) ## Increment the stats counters: if len(prev_misc_txt) > 0: count_misc += 1 count_title += 1 ## Empty the previously-cited item place-holder: previously_cited_item = None ## Now convert this URL to MARC XML cur_misc_txt = cur_misc_txt.lstrip(".;, ").rstrip() if url_string.find("http //") == 0: url_string = u"http://" + url_string[7:] elif url_string.find("ftp //") == 0: url_string = u"ftp://" + url_string[6:] if url_descr.find("http //") == 0: url_descr = u"http://" + url_descr[7:] elif url_descr.find("ftp //") == 0: url_descr = u"ftp://" + url_descr[6:] xml_line += \ _refextract_markup_url_as_marcxml(url_string, url_descr, cur_misc_txt) ## Increment the stats counters: if len(cur_misc_txt) > 0: count_misc += 1 count_url += 1 cur_misc_txt = u"" elif tag_type == "SER": ## This tag is a SERIES tag; Since it was not preceeded by a TITLE tag, ## it is useless - strip the tag and put it into miscellaneous: (cur_misc_txt, processed_line) = \ _convert_unusable_tag_to_misc(processed_line, cur_misc_txt, \ tag_match_start,tag_match_end, CFG_REFEXTRACT_MARKER_CLOSING_SERIES) elif tag_type == "VOL": ## This tag is a VOLUME tag; Since it was not preceeded by a TITLE tag, ## it is useless - strip the tag and put it into miscellaneous: (cur_misc_txt, processed_line) = \ _convert_unusable_tag_to_misc(processed_line, cur_misc_txt, \ tag_match_start,tag_match_end, CFG_REFEXTRACT_MARKER_CLOSING_VOLUME) elif tag_type == "YR": ## This tag is a YEAR tag; Since it's not preceeded by TITLE and VOLUME tags, it ## is useless - strip the tag and put the contents into miscellaneous: (cur_misc_txt, processed_line) = \ _convert_unusable_tag_to_misc(processed_line, cur_misc_txt, \ tag_match_start,tag_match_end, CFG_REFEXTRACT_MARKER_CLOSING_YEAR) elif tag_type == "PG": ## This tag is a PAGE tag; Since it's not preceeded by TITLE, VOLUME and YEAR tags, ## it is useless - strip the tag and put the contents into miscellaneous: (cur_misc_txt, processed_line) = \ _convert_unusable_tag_to_misc(processed_line, cur_misc_txt, \ tag_match_start,tag_match_end, CFG_REFEXTRACT_MARKER_CLOSING_PAGE) else: ## Unknown tag - discard as miscellaneous text: cur_misc_txt += processed_line[0:tag_match.end()] processed_line = processed_line[tag_match.end():] ## Look for the next tag in the processed line: tag_match = sre_tagged_citation.search(processed_line) ## If a previously cited item remains, convert it into MARC XML: if previously_cited_item is not None: if previously_cited_item['type'] == "REPORTNUMBER": ## previously cited item was a REPORT NUMBER. ## Add previously cited REPORT NUMBER to XML string: prev_report_num = previously_cited_item['report_num'] prev_misc_txt = previously_cited_item['misc_txt'].lstrip(".;, ").rstrip() xml_line += _refextract_markup_reportnumber_as_marcxml(prev_report_num, prev_misc_txt) ## Increment the stats counters: if len(prev_misc_txt) > 0: count_misc += 1 count_reportnum += 1 elif previously_cited_item['type'] == "TITLE": ## previously cited item was a TITLE. ## Add previously cited TITLE to XML string: prev_title = previously_cited_item['title'] prev_volume = previously_cited_item['volume'] prev_year = previously_cited_item['year'] prev_page = previously_cited_item['page'] prev_misc_txt = previously_cited_item['misc_txt'].lstrip(".;, ").rstrip() xml_line += _refextract_markup_title_as_marcxml(prev_title, prev_volume, prev_year, prev_page, prev_misc_txt) ## Increment the stats counters: if len(prev_misc_txt) > 0: count_misc += 1 count_title += 1 ## free up previously_cited_item: previously_cited_item = None ## place any remaining miscellaneous text into the appropriate MARC XML fields: cur_misc_txt += processed_line if len(cur_misc_txt.strip(" .;,")) > 0: ## The remaining misc text is not just a full-stop or semi-colon. Add it: xml_line += _refextract_markup_miscellaneous_text_as_marcxml(cur_misc_txt) ## Increment the stats counters: count_misc += 1 ## return the reference-line as MARC XML: return (xml_line, count_misc, count_title, count_reportnum, count_url) | 6258c992576ce615917d5d2bfa3f925bb23de3b5 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12027/6258c992576ce615917d5d2bfa3f925bb23de3b5/refextract.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1765,
67,
11005,
67,
6180,
67,
1369,
67,
869,
67,
3684,
71,
67,
2902,
12,
1369,
4672,
3536,
6083,
279,
5204,
2114,
980,
16,
1765,
518,
358,
490,
27206,
3167,
18,
632,
891,
980,
30,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1765,
67,
11005,
67,
6180,
67,
1369,
67,
869,
67,
3684,
71,
67,
2902,
12,
1369,
4672,
3536,
6083,
279,
5204,
2114,
980,
16,
1765,
518,
358,
490,
27206,
3167,
18,
632,
891,
980,
30,
2... |
cursor.execute("DELETE FROM session " "WHERE sid=%s AND authenticated=0", (self.sid,)) | cursor.execute(""" DELETE FROM session WHERE sid=%s AND authenticated=0 """, (self.sid,)) | def delete_session_cookie(db): cursor = db.cursor() if self._new: self.last_visit = now self._new = False # The session might already exist even if _new is True since # it could have been created by a concurrent request (#3563). try: cursor.execute("INSERT INTO session " " (sid,last_visit,authenticated)" " VALUES (%s,%s,%s)", (self.sid, self.last_visit, authenticated)) except Exception: self.env.log.warning('Session %s already exists', self.sid) db.rollback() return if self._old != self: attrs = [(self.sid, authenticated, k, v) for k, v in self.items()] cursor.execute("DELETE FROM session_attribute WHERE sid=%s", (self.sid,)) self._old = dict(self.items()) if attrs: cursor.executemany("INSERT INTO session_attribute " " (sid,authenticated,name,value) " " VALUES (%s,%s,%s,%s)", attrs) elif not authenticated: # No need to keep around empty unauthenticated sessions cursor.execute("DELETE FROM session " "WHERE sid=%s AND authenticated=0", (self.sid,)) return # Update the session last visit time if it is over an hour old, # so that session doesn't get purged if now - self.last_visit > UPDATE_INTERVAL: self.last_visit = now self.env.log.info("Refreshing session %s", self.sid) cursor.execute('UPDATE session SET last_visit=%s ' 'WHERE sid=%s AND authenticated=%s', (self.last_visit, self.sid, authenticated)) # Purge expired sessions. We do this only when the session was # changed as to minimize the purging. mintime = now - PURGE_AGE self.env.log.debug('Purging old, expired, sessions.') cursor.execute("DELETE FROM session_attribute " "WHERE authenticated=0 AND sid " "IN (SELECT sid FROM session WHERE " "authenticated=0 AND last_visit < %s)", (mintime,)) cursor.execute("DELETE FROM session WHERE " "authenticated=0 AND last_visit < %s", (mintime,)) | 4dbba84b306edf6862d2638abdd4283f65299711 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9317/4dbba84b306edf6862d2638abdd4283f65299711/session.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1430,
67,
3184,
67,
8417,
12,
1966,
4672,
3347,
273,
1319,
18,
9216,
1435,
309,
365,
6315,
2704,
30,
365,
18,
2722,
67,
11658,
273,
2037,
365,
6315,
2704,
273,
1083,
468,
1021,
1339,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1430,
67,
3184,
67,
8417,
12,
1966,
4672,
3347,
273,
1319,
18,
9216,
1435,
309,
365,
6315,
2704,
30,
365,
18,
2722,
67,
11658,
273,
2037,
365,
6315,
2704,
273,
1083,
468,
1021,
1339,
4... |
self.now = datetime.now() self.month = self.now.month self.months = [self.now] self.year = self.now.year y = self.year for x in range(11): m = self.month - x + 1 | today = date.today() self.months = [] m = today.month y = today.year for x in range(12): self.months.append(date(y, m, 1)) m = m - 1 | def __init__(self, context, request, year=None): self.context = context self.request = request portal_url = getToolByName(context, 'portal_url') self.portal = portal_url.getPortalObject() self.site_url = portal_url() self.catalog = getToolByName(context, 'portal_catalog') self.mtool = getToolByName(context, 'portal_membership') self.now = datetime.now() self.month = self.now.month self.months = [self.now] self.year = self.now.year # months is a list of datetime object of the past 12 months counting # from now. y = self.year for x in range(11): m = self.month - x + 1 if m <= 0: m += 12 if y == self.year: y -= 1 self.months.append(datetime(y, m, 1)) | 0941e64ae8d2ea0bcea576836a96afdcfa2fe53f /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/10443/0941e64ae8d2ea0bcea576836a96afdcfa2fe53f/employees.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
819,
16,
590,
16,
3286,
33,
7036,
4672,
365,
18,
2472,
273,
819,
365,
18,
2293,
273,
590,
11899,
67,
718,
273,
336,
6364,
5911,
12,
2472,
16,
296,
2438... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
819,
16,
590,
16,
3286,
33,
7036,
4672,
365,
18,
2472,
273,
819,
365,
18,
2293,
273,
590,
11899,
67,
718,
273,
336,
6364,
5911,
12,
2472,
16,
296,
2438... |
for i, t in enumerate(tokens): | for t in tokens: | def check_function_call_spaces(tokens, lines, warnings): "Check for function calls having the silly GNU spaces before (." keywlist = 'if', 'for', 'while', 'switch', 'return', 'case', 'goto' keywords = dict([(x, 1) for x in keywlist]) for i, t in enumerate(tokens): if t.bracelevel == 0: continue if t.typ != Token.punct or t.string != '(': continue prec = tokens[i-1] if prec.typ != Token.ident or prec.string in keywords: continue if prec.line == t.line and prec.end == t.col: continue m = t.matching if tokens[m+1].typ == Token.punct and tokens[m+1].string == '(': continue w = 'Space between function name and parenthesis (col %d): %s' warnings.append((t.line, w % (t.col, lines[t.line]))) | 6ef4f8f80b65ddd3424b5c94f1c8a2c566383415 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6176/6ef4f8f80b65ddd3424b5c94f1c8a2c566383415/check-coding-style.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
915,
67,
1991,
67,
9554,
12,
7860,
16,
2362,
16,
5599,
4672,
315,
1564,
364,
445,
4097,
7999,
326,
17714,
715,
611,
50,
57,
7292,
1865,
261,
1199,
498,
91,
1098,
273,
296,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
915,
67,
1991,
67,
9554,
12,
7860,
16,
2362,
16,
5599,
4672,
315,
1564,
364,
445,
4097,
7999,
326,
17714,
715,
611,
50,
57,
7292,
1865,
261,
1199,
498,
91,
1098,
273,
296,
4... |
for badobject in ("a string", 3+42j, randrange, NotAnIntNS(), NotAnIntOS()): self.assertRaises(struct.error, struct.pack, format, | for badobject in ("a string", 3+42j, randrange): self.assertRaises((TypeError, struct.error), struct.pack, self.format, | def __long__(self): return -163L | 30117f8486d0beab41eb919e163ec1a2e4d7df48 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12029/30117f8486d0beab41eb919e163ec1a2e4d7df48/test_struct.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
5748,
972,
12,
2890,
4672,
327,
300,
30886,
48,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
5748,
972,
12,
2890,
4672,
327,
300,
30886,
48,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
for taskID in sortLit(taskDict.keys()): | for taskID in sortList(taskDict.keys()): | def submitTasks(self,taskDict): submitted = 0 failed = 0 startTime = time.time() for taskID in sortLit(taskDict.keys()): taskDict[taskID] res = self.submitToExternal(taskDict[taskID]['TaskObject']) if res['OK']: taskDict[taskID]['ExternalID'] = res['Value'] taskDict[taskID]['Success'] = True submitted +=1 else: self.log.warn("Failed to submit task to WMS",res['Message']) taskDict[taskID]['Success'] = False failed += 1 self.log.info('submitTasks: Submitted %d tasks to RMS in %.1f seconds' % (submitted,time.time()-startTime)) self.log.info('submitTasks: Failed to submit %d tasks to RMS.' % (failed,time.time()-startTime)) return S_OK(taskDict) | 7e361d624cd9374c5dec904aad18c8eb176604b6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/7e361d624cd9374c5dec904aad18c8eb176604b6/TaskManager.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4879,
6685,
12,
2890,
16,
4146,
5014,
4672,
9638,
273,
374,
2535,
273,
374,
8657,
273,
813,
18,
957,
1435,
364,
29772,
316,
1524,
682,
12,
4146,
5014,
18,
2452,
1435,
4672,
1562,
5014,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4879,
6685,
12,
2890,
16,
4146,
5014,
4672,
9638,
273,
374,
2535,
273,
374,
8657,
273,
813,
18,
957,
1435,
364,
29772,
316,
1524,
682,
12,
4146,
5014,
18,
2452,
1435,
4672,
1562,
5014,
... |
details = "\n".join((error, details)) | info = _("If this problem persists consider reporting a bug using the link in the help menu.") details = "\n".join((error, info)) | def CreateDefaultErrorPane(self, error, details): message = _("A GStreamer error has occurred.") details = "\n".join((error, details)) msg_area = MessageArea.MessageArea() msg_area.add_button(gtk.STOCK_CLOSE, gtk.RESPONSE_CLOSE) msg_area.set_text_and_icon(gtk.STOCK_DIALOG_ERROR, message, details) msg_area.connect("response", self.OnMessageAreaReponse, msg_area) msg_area.connect("close", self.OnMessageAreaClose, msg_area) return msg_area | adb75921b66a6695c73cc7edcf6394a736a7ff1e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10033/adb75921b66a6695c73cc7edcf6394a736a7ff1e/RecordingView.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1788,
1868,
668,
8485,
12,
2890,
16,
555,
16,
3189,
4672,
883,
273,
389,
2932,
37,
611,
1228,
264,
555,
711,
7841,
1199,
13,
1123,
273,
389,
2932,
2047,
333,
6199,
13508,
1486,
5260,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1788,
1868,
668,
8485,
12,
2890,
16,
555,
16,
3189,
4672,
883,
273,
389,
2932,
37,
611,
1228,
264,
555,
711,
7841,
1199,
13,
1123,
273,
389,
2932,
2047,
333,
6199,
13508,
1486,
5260,
1... |
if mobj._date_name in mobj._columns: | if hasattr(mobj, '_date_name') and mobj._date_name in mobj._columns: | def _views_get_fnc(self, cr, uid, ids, name, arg, context={}): res={} for act in self.browse(cr, uid, ids): res[act.id]=[(view.view_id.id, view.view_mode) for view in act.view_ids] if (not act.view_ids): modes = act.view_mode.split(',') find = False if act.view_id: res[act.id].append((act.view_id.id, act.view_id.type)) for t in modes: if act.view_id and (t == act.view_id.type) and not find: find = True continue res[act.id].append((False, t)) | 209d2d67ff4549916f68983462d17c975b890aef /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12853/209d2d67ff4549916f68983462d17c975b890aef/ir_actions.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
7061,
67,
588,
67,
4293,
71,
12,
2890,
16,
4422,
16,
4555,
16,
3258,
16,
508,
16,
1501,
16,
819,
12938,
4672,
400,
12938,
364,
1328,
316,
365,
18,
25731,
12,
3353,
16,
4555,
16,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
7061,
67,
588,
67,
4293,
71,
12,
2890,
16,
4422,
16,
4555,
16,
3258,
16,
508,
16,
1501,
16,
819,
12938,
4672,
400,
12938,
364,
1328,
316,
365,
18,
25731,
12,
3353,
16,
4555,
16,... |
if debug_dragstuff: print "contentsDragLeaveEvent, event == %r" % event | def contentsDragLeaveEvent(self, event): if debug_dragstuff: print "contentsDragLeaveEvent, event == %r" % event # the following code is similar in dragLeave and Drop, not sure if identical self.last_dragMove_ok = False # not needed right now, but might matter # after the next dragEnter but before the next dragMove self.last_dragMove_cpos = None self.update_drop_point_highlighting() return | 8e76c375339a6ae063e685fcb0b4bee3c03d78cd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/8e76c375339a6ae063e685fcb0b4bee3c03d78cd/TreeWidget.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2939,
11728,
21300,
1133,
12,
2890,
16,
871,
4672,
309,
1198,
67,
15997,
334,
3809,
30,
1172,
315,
3980,
11728,
21300,
1133,
16,
871,
422,
738,
86,
6,
738,
871,
468,
326,
3751,
981,
35... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2939,
11728,
21300,
1133,
12,
2890,
16,
871,
4672,
309,
1198,
67,
15997,
334,
3809,
30,
1172,
315,
3980,
11728,
21300,
1133,
16,
871,
422,
738,
86,
6,
738,
871,
468,
326,
3751,
981,
35... | |
'kIICore', 'kGB0', 'kBigFive', 'kHKSCS', 'kHanyuPinyin', 'kJIS0', | 'kIICore', 'kGB0', 'kBigFive', 'kHKSCS', 'kHanyuPinyin', 'kJis0', | def generator(self): """Provides all data of one character per entry.""" columns = self.unihanGenerator.keys() for char, entryDict in self.unihanGenerator.generator(): newEntryDict = {UnihanBuilder.CHARACTER_COLUMN: char} for column in columns: if entryDict.has_key(column): newEntryDict[column] = entryDict[column] else: newEntryDict[column] = None yield newEntryDict | 6736272070f4e5c3b4376606c23c307009ad20a4 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11128/6736272070f4e5c3b4376606c23c307009ad20a4/builder.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4456,
12,
2890,
4672,
3536,
17727,
777,
501,
434,
1245,
3351,
1534,
1241,
12123,
2168,
273,
365,
18,
318,
7392,
304,
3908,
18,
2452,
1435,
364,
1149,
16,
1241,
5014,
316,
365,
18,
318,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4456,
12,
2890,
4672,
3536,
17727,
777,
501,
434,
1245,
3351,
1534,
1241,
12123,
2168,
273,
365,
18,
318,
7392,
304,
3908,
18,
2452,
1435,
364,
1149,
16,
1241,
5014,
316,
365,
18,
318,
... |
This test fails with GCC 3.4.x (see http://gcc.gnu.org/bugzilla/show_bug.cgi?id=20239 [3.4 Regression] ICE on empty preprocessed input)""" | This test fails with an internal compiler error in GCC 3.4.x for x < 5 (see http://gcc.gnu.org/bugzilla/show_bug.cgi?id=20239 [3.4 Regression] ICE on empty preprocessed input). But that's gcc's problem, not ours, so we make this test pass if gcc gets an ICE.""" | def setup(self): self.require_root() | ded623a8d8ba8f61a226fba0294a913a5a61aa29 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7392/ded623a8d8ba8f61a226fba0294a913a5a61aa29/testdistcc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
12,
2890,
4672,
365,
18,
6528,
67,
3085,
1435,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
12,
2890,
4672,
365,
18,
6528,
67,
3085,
1435,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
if cmd not in ("find", "info"): | if cmd not in my_commands: | def chanmsg(self, irc, msg, user): channel = self.__getChannel(msg.args[0]) text = self.space_re.subn(' ', msg.args[1].strip())[0] if text[0] != self.registryValue("prefixchar", channel): return text = text[1:] (cmd, rest) = (text.split(' ', 1) + [None])[:2] if cmd not in ("find", "info"): return if not rest: return (term, rest) = (rest.split(' ', 1) + [None])[:2] if cmd == "find": self.real_find(irc, msg, [], term, rest) else: self.real_info(irc, msg, [], term, rest) | 283652a263ab69ddf4dea37a7a8e0323c2a2eb2b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3102/283652a263ab69ddf4dea37a7a8e0323c2a2eb2b/plugin.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3861,
3576,
12,
2890,
16,
277,
1310,
16,
1234,
16,
729,
4672,
1904,
273,
365,
16186,
588,
2909,
12,
3576,
18,
1968,
63,
20,
5717,
977,
273,
365,
18,
2981,
67,
266,
18,
1717,
82,
2668... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3861,
3576,
12,
2890,
16,
277,
1310,
16,
1234,
16,
729,
4672,
1904,
273,
365,
16186,
588,
2909,
12,
3576,
18,
1968,
63,
20,
5717,
977,
273,
365,
18,
2981,
67,
266,
18,
1717,
82,
2668... |
for aaaaa in self.Address.keys(): | for aaaaa in self.WaysPassed: | def AddrTreeRecurse (top, wayback="» "): | cda6679173391cb972f4060eef8750b9189454a7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4644/cda6679173391cb972f4060eef8750b9189454a7/initialParse.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10216,
2471,
426,
17682,
261,
3669,
16,
4031,
823,
1546,
131,
124,
315,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10216,
2471,
426,
17682,
261,
3669,
16,
4031,
823,
1546,
131,
124,
315,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
self.previous_text = '\n' self.process_children(tag, tag_css) | self.current_block.must_append = True | def update_css(ncss): for key in ncss.keys(): if self.css.has_key(key): self.css[key].update(ncss[key]) else: self.css[key] = ncss[key] | e39dc4223f880d6b721e75d35aaa36ac5ad96971 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9125/e39dc4223f880d6b721e75d35aaa36ac5ad96971/convert_from.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
67,
5212,
12,
82,
5212,
4672,
364,
498,
316,
290,
5212,
18,
2452,
13332,
309,
365,
18,
5212,
18,
5332,
67,
856,
12,
856,
4672,
365,
18,
5212,
63,
856,
8009,
2725,
12,
82,
5212,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
67,
5212,
12,
82,
5212,
4672,
364,
498,
316,
290,
5212,
18,
2452,
13332,
309,
365,
18,
5212,
18,
5332,
67,
856,
12,
856,
4672,
365,
18,
5212,
63,
856,
8009,
2725,
12,
82,
5212,... |
value = self.getRaw(instance, raw=1) if IBaseUnit.isImplementedBy(value): return str(value.getContentType()) | accessor = self.getEditAccessor(instance) if accessor is not None: value = accessor(maybe_baseunit=1) | def getContentType(self, instance): """Return the mime type of object if known or can be guessed; otherwise, return None.""" value = '' value = self.getRaw(instance, raw=1) if IBaseUnit.isImplementedBy(value): return str(value.getContentType()) mimetype = getattr(aq_base(value), 'mimetype', None) if mimetype is None: mimetype, enc = guess_content_type('', str(value), None) else: # mimetype may be an imimetype object mimetype = str(mimetype) return mimetype | 6c49241026527d58ec33dad69b98fdf197a65ae4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12165/6c49241026527d58ec33dad69b98fdf197a65ae4/Field.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
17470,
12,
2890,
16,
791,
4672,
3536,
990,
326,
4892,
618,
434,
733,
309,
4846,
578,
848,
506,
3058,
281,
730,
31,
3541,
16,
327,
599,
12123,
460,
273,
875,
9789,
273,
365,
18,
588,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
17470,
12,
2890,
16,
791,
4672,
3536,
990,
326,
4892,
618,
434,
733,
309,
4846,
578,
848,
506,
3058,
281,
730,
31,
3541,
16,
327,
599,
12123,
460,
273,
875,
9789,
273,
365,
18,
588,
... |
apply(self.tk.call, (self._w, 'add', name) + self._options(cnf, kw)) self.subwidget_list[name] = TixSubWidget(self, name, check_intermediate=0) return self.subwidget_list[name] | apply(self.tk.call, (self._w, 'add', name) + self._options(cnf, kw)) self.subwidget_list[name] = TixSubWidget(self, name, check_intermediate=0) return self.subwidget_list[name] | def add(self, name, cnf={}, **kw): | 22710823fb554a796dc96c44885d7a9389426824 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/22710823fb554a796dc96c44885d7a9389426824/Tix.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
527,
12,
2890,
16,
508,
16,
23118,
28793,
2826,
9987,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
527,
12,
2890,
16,
508,
16,
23118,
28793,
2826,
9987,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
del self.daap_playlists[x.id] | del self.daap_playlists[x] | def handle_playlist_removed(self, obj, removed): for x in removed: del self.daap_playlists[x.id] | d0854cbc42b02a9b723215a904e1eca4bb1bc8e7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12354/d0854cbc42b02a9b723215a904e1eca4bb1bc8e7/sharing.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
67,
1601,
1098,
67,
14923,
12,
2890,
16,
1081,
16,
3723,
4672,
364,
619,
316,
3723,
30,
1464,
365,
18,
2414,
438,
67,
1601,
9772,
63,
92,
18,
350,
65,
2,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
67,
1601,
1098,
67,
14923,
12,
2890,
16,
1081,
16,
3723,
4672,
364,
619,
316,
3723,
30,
1464,
365,
18,
2414,
438,
67,
1601,
9772,
63,
92,
18,
350,
65,
2,
-100,
-100,
-100,
-100... |
elif issubclass(node.__class__, NamedNode): | elif issubclass(node.__class__, PandaNode): | def pickGeom3D(self, targetNodePath = render, origin = Point3(0), dir = Vec3(0,0,-1), fIntersectUnpickable = 0): self.collideWithGeom() numEntries = self.pick3D(targetNodePath, origin, dir) # Init self.cqIndex self.cqIndex = -1 # Pick out the closest object that isn't a widget for i in range(0,numEntries): entry = self.cq.getEntry(i) node = entry.getIntoNode() nodePath = targetNodePath.findPathTo(node) # Don't pick hidden nodes if nodePath.isHidden(): pass # Can pick unpickable, use the first visible node if fIntersectUnpickable: self.cqIndex = i break # Is it a named node?, If so, see if it has a name elif issubclass(node.__class__, NamedNode): name = node.getName() if name in self.unpickable: pass else: self.cqIndex = i break # Not hidden and not one of the widgets, use it else: self.cqIndex = i break # Did we hit an object? if(self.cqIndex >= 0): # Yes! # Find hit point in parent's space entry = self.cq.getEntry(self.cqIndex) hitPt = self.parentToHitPt(entry) hitPtDist = Vec3(hitPt - ZERO_POINT).length() return (node, hitPt, hitPtDist) else: return (None, ZERO_POINT, 0) | 6608efefe51f08c630783317d62424a94c1d3dbb /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8543/6608efefe51f08c630783317d62424a94c1d3dbb/DirectSelection.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6002,
16972,
23,
40,
12,
2890,
16,
25804,
743,
273,
1743,
16,
4026,
273,
4686,
23,
12,
20,
3631,
1577,
273,
12969,
23,
12,
20,
16,
20,
16,
17,
21,
3631,
284,
22541,
984,
11503,
429,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6002,
16972,
23,
40,
12,
2890,
16,
25804,
743,
273,
1743,
16,
4026,
273,
4686,
23,
12,
20,
3631,
1577,
273,
12969,
23,
12,
20,
16,
20,
16,
17,
21,
3631,
284,
22541,
984,
11503,
429,
... |
'\\mathrm{SL}_2(\\mathbf{Z})' | '\\mbox{\\rm SL}_2(\\mathbf{Z})' | def _latex_(self): r""" Return the LaTeX representation of self. | 6ba637c282ecbcc271600abf63d736f0343131a2 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9417/6ba637c282ecbcc271600abf63d736f0343131a2/congroup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
26264,
67,
12,
2890,
4672,
436,
8395,
2000,
326,
21072,
21575,
60,
4335,
434,
365,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
26264,
67,
12,
2890,
4672,
436,
8395,
2000,
326,
21072,
21575,
60,
4335,
434,
365,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
resource = memory.File(TEST_SYNTAX_ERROR) | def test_bad_syntax_csv_file_with_schema(self): resource = memory.File(TEST_SYNTAX_ERROR) handler = CSV() handler.columns = ['one', 'two', 'three'] handler.schema = {'one': Unicode, 'two': Unicode, 'three': Unicode} self.assertRaises(ValueError, handler.load_state, resource) | 7404f7e83f01f05f84fa82ab4b9b9fcd644b8aa8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12681/7404f7e83f01f05f84fa82ab4b9b9fcd644b8aa8/test_csv.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
8759,
67,
14308,
67,
6715,
67,
768,
67,
1918,
67,
4821,
12,
2890,
4672,
1838,
273,
10381,
1435,
1838,
18,
5112,
273,
10228,
476,
2187,
296,
15415,
2187,
296,
451,
992,
3546,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
8759,
67,
14308,
67,
6715,
67,
768,
67,
1918,
67,
4821,
12,
2890,
4672,
1838,
273,
10381,
1435,
1838,
18,
5112,
273,
10228,
476,
2187,
296,
15415,
2187,
296,
451,
992,
3546,
... | |
Nodes.versionId = Instances.versionId | Nodes.versionId = Instances.versionId AND Nodes.itemId = Instances.itemId | def troveNames(self, label): cu = self.db.cursor() cu.execute("""SELECT DISTINCT item FROM Labels JOIN LabelMap ON Labels.labelId = LabelMap.labelId JOIN Branches ON LabelMap.branchId = Branches.branchId JOIN Nodes ON Branches.branchId = Nodes.branchId JOIN Instances ON Nodes.versionId = Instances.versionId JOIN Items ON Instances.itemId = Items.itemId WHERE Labels.label=? AND isPresent=1 ORDER BY item """, label.asString()) | c91612bbb4a00e0e800693b16e1665671d071387 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8747/c91612bbb4a00e0e800693b16e1665671d071387/trovestore.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
23432,
537,
1557,
12,
2890,
16,
1433,
4672,
15985,
273,
365,
18,
1966,
18,
9216,
1435,
15985,
18,
8837,
2932,
3660,
4803,
23286,
761,
4571,
12395,
7645,
5287,
863,
6229,
12395,
18,
1925,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
23432,
537,
1557,
12,
2890,
16,
1433,
4672,
15985,
273,
365,
18,
1966,
18,
9216,
1435,
15985,
18,
8837,
2932,
3660,
4803,
23286,
761,
4571,
12395,
7645,
5287,
863,
6229,
12395,
18,
1925,
... |
s = '%s\tret = self._oleobj_.InvokeTypes(0x%x, LCID, %s, %s, %s%s)\n' % (linePrefix, id, fdesc[4], retDesc, `argsDesc`, _BuildArgList(fdesc, names)) | s = '%s\tret = self._oleobj_.InvokeTypes(%d, LCID, %s, %s, %s%s)\n' % (linePrefix, id, fdesc[4], retDesc, `argsDesc`, _BuildArgList(fdesc, names)) | s = linePrefix + 'def ' + name + '(self' + BuildCallList(fdesc, names, defNamedOptArg, defNamedNotOptArg, defUnnamedArg) + '):' | 976f9a30a4ef8b146cca4072b439ef3c29423cb0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/677/976f9a30a4ef8b146cca4072b439ef3c29423cb0/build.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
272,
273,
980,
2244,
397,
296,
536,
296,
397,
508,
397,
7747,
2890,
11,
397,
3998,
1477,
682,
12,
74,
5569,
16,
1257,
16,
1652,
7604,
6179,
4117,
16,
1652,
7604,
1248,
6179,
4117,
16,
1652,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
272,
273,
980,
2244,
397,
296,
536,
296,
397,
508,
397,
7747,
2890,
11,
397,
3998,
1477,
682,
12,
74,
5569,
16,
1257,
16,
1652,
7604,
6179,
4117,
16,
1652,
7604,
1248,
6179,
4117,
16,
1652,
... |
if symbol_ordinal*4<len(address_of_functions): | if symbol_ordinal*4 < len(address_of_functions): | def parse_export_directory(self, rva, size): """Parse the export directory. Given the RVA of the export directory, it will process all its entries. The exports will be made available through a list "exports" containing a tuple with the following elements: (ordinal, symbol_address, symbol_name) And also through a dictionary "exports_by_ordinal" whose keys will be the ordinals and the values tuples of the from: (symbol_address, symbol_name) The symbol addresses are relative, not absolute. """ try: export_dir = self.__unpack_data__( self.__IMAGE_EXPORT_DIRECTORY_format__, self.get_data( rva, Structure(self.__IMAGE_EXPORT_DIRECTORY_format__).sizeof() ), file_offset = self.get_offset_from_rva(rva) ) except PEFormatError: self.__warnings.append( 'Error parsing export directory at RVA: 0x%x' % ( rva ) ) return if not export_dir: return try: address_of_names = self.get_data( export_dir.AddressOfNames, export_dir.NumberOfNames*4) address_of_name_ordinals = self.get_data( export_dir.AddressOfNameOrdinals, export_dir.NumberOfNames*4) address_of_functions = self.get_data( export_dir.AddressOfFunctions, export_dir.NumberOfFunctions*4) except PEFormatError: self.__warnings.append( 'Error parsing export directory at RVA: 0x%x' % ( rva ) ) return exports = [] for i in xrange(export_dir.NumberOfNames): symbol_name = self.get_string_at_rva( self.get_dword_from_data(address_of_names, i)) symbol_ordinal = self.get_word_from_data( address_of_name_ordinals, i) if symbol_ordinal*4<len(address_of_functions): symbol_address = self.get_dword_from_data( address_of_functions, symbol_ordinal) else: # Corrupt? a bad pointer... we assume it's all # useless, no exports return None # If the funcion's RVA points within the export directory # it will point to a string with the forwarded symbol's string # instead of pointing the the function start address. if symbol_address>=rva and symbol_address<rva+size: forwarder_str = self.get_string_at_rva(symbol_address) else: forwarder_str = None exports.append( ExportData( ordinal = export_dir.Base+symbol_ordinal, address = symbol_address, name = symbol_name, forwarder = forwarder_str)) ordinals = [exp.ordinal for exp in exports] for idx in xrange(export_dir.NumberOfFunctions): if not idx+export_dir.Base in ordinals: symbol_address = self.get_dword_from_data( address_of_functions, idx) # # Checking for forwarder again. # if symbol_address>=rva and symbol_address<rva+size: forwarder_str = self.get_string_at_rva(symbol_address) else: forwarder_str = None exports.append( ExportData( ordinal = export_dir.Base+idx, address = symbol_address, name = None, forwarder = forwarder_str)) return ExportDirData( struct = export_dir, symbols = exports) | 34b416148a27cfed56459bf99d22f4b720a88d7f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4177/34b416148a27cfed56459bf99d22f4b720a88d7f/pefile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
6530,
67,
5149,
12,
2890,
16,
5633,
69,
16,
963,
4672,
3536,
3201,
326,
3359,
1867,
18,
225,
16803,
326,
534,
27722,
434,
326,
3359,
1867,
16,
518,
903,
1207,
777,
2097,
3222... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
6530,
67,
5149,
12,
2890,
16,
5633,
69,
16,
963,
4672,
3536,
3201,
326,
3359,
1867,
18,
225,
16803,
326,
534,
27722,
434,
326,
3359,
1867,
16,
518,
903,
1207,
777,
2097,
3222... |
cls.w_language_pl = cls.space.wrap("pl_PL") | cls.w_language_pl = cls.space.wrap("pl_PL.UTF-8") | def setup_class(cls): cls.space = gettestobjspace(usemodules=['_locale']) if sys.platform != 'win32': cls.w_language_en = cls.space.wrap("en_US") cls.w_language_utf8 = cls.space.wrap("en_US.UTF-8") cls.w_language_pl = cls.space.wrap("pl_PL") else: cls.w_language_en = cls.space.wrap("English_US") cls.w_language_utf8 = cls.space.wrap("English_US.65001") cls.w_language_pl = cls.space.wrap("Polish_Poland") | b2947f1d85855c2f4aef5ada5f0134e917951163 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6934/b2947f1d85855c2f4aef5ada5f0134e917951163/test_locale.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
67,
1106,
12,
6429,
4672,
2028,
18,
2981,
273,
336,
3813,
2603,
2981,
12,
1202,
6400,
33,
3292,
67,
6339,
19486,
309,
2589,
18,
9898,
480,
296,
8082,
1578,
4278,
2028,
18,
91,
67... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
67,
1106,
12,
6429,
4672,
2028,
18,
2981,
273,
336,
3813,
2603,
2981,
12,
1202,
6400,
33,
3292,
67,
6339,
19486,
309,
2589,
18,
9898,
480,
296,
8082,
1578,
4278,
2028,
18,
91,
67... |
relative = path[len(module):] | relative = path.rstrip('/')[len(module):] | def get_entry_from_path(path, module=self.module): # Given the repository url of this wc, say # "http://server/plone/CMFPlone/branches/Plone-2_0-branch" # extract the "entry" portion (a relative path) from what # svn log --xml says, ie # "/CMFPlone/branches/Plone-2_0-branch/tests/PloneTestCase.py" # that is to say "tests/PloneTestCase.py" if path.startswith(module): relative = path[len(module):] if relative.startswith('/'): return relative[1:] else: return relative | 97760ab25803583edfd4ef254182505f38b7f68c /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/11312/97760ab25803583edfd4ef254182505f38b7f68c/subversion.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
4099,
67,
2080,
67,
803,
12,
803,
16,
1605,
33,
2890,
18,
2978,
4672,
468,
16803,
326,
3352,
880,
434,
333,
14310,
16,
12532,
468,
282,
315,
2505,
2207,
3567,
19,
412,
476,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
4099,
67,
2080,
67,
803,
12,
803,
16,
1605,
33,
2890,
18,
2978,
4672,
468,
16803,
326,
3352,
880,
434,
333,
14310,
16,
12532,
468,
282,
315,
2505,
2207,
3567,
19,
412,
476,
... |
self.assertContains(res, 'Bars') | self.assertContains(res, 'Bars') self.assertContains(res, 'Users', 1) | def test_app_index(self): self.client.login(username='staff', password='123') res = self.client.get('/admin/test_app/foo/') self.assertEqual(res.status_code, 200) self.assertContains(res, 'Foos') self.assertNotContains(res, 'Bars') | 6cc360ec5504f72713a6359d8c450b89d95f6c9e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3856/6cc360ec5504f72713a6359d8c450b89d95f6c9e/tests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
2910,
67,
1615,
12,
2890,
4672,
365,
18,
2625,
18,
5819,
12,
5053,
2218,
31096,
2187,
2201,
2218,
12936,
6134,
400,
273,
365,
18,
2625,
18,
588,
2668,
19,
3666,
19,
3813,
67,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
2910,
67,
1615,
12,
2890,
4672,
365,
18,
2625,
18,
5819,
12,
5053,
2218,
31096,
2187,
2201,
2218,
12936,
6134,
400,
273,
365,
18,
2625,
18,
588,
2668,
19,
3666,
19,
3813,
67,... |
('print self', self._debug_printself), | def debug_menu_items(self): "[subclasses can in theory override this, but probably needn't and shouldn't]" import debug res = [ ('(debugging menu)', noop, 'disabled'), ('print self', self._debug_printself), # None, # separator ] if debug.exec_allowed(): #bruce 041217 made this item conditional on whether it will work res.extend( [ ('run py code', self._debug_runpycode), ] ) res.extend( [ ('enable ATOM_DEBUG', self._debug_enable_atom_debug ), ('disable ATOM_DEBUG', self._debug_disable_atom_debug ), ] ) if self._debug_win: res.extend( [ ('save window layout', self._debug_save_window_layout ), ('load window layout', self._debug_load_window_layout ), #bruce 050117 prototype "save window layout" here; when it works, move it elsewhere ] ) return res | 03a8cf9ca986e153701e1a9679e6bf5a4ee883d8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/03a8cf9ca986e153701e1a9679e6bf5a4ee883d8/debug.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1198,
67,
5414,
67,
3319,
12,
2890,
4672,
5158,
1717,
4701,
848,
316,
326,
630,
3849,
333,
16,
1496,
8656,
1608,
82,
1404,
471,
12044,
1404,
4279,
1930,
1198,
400,
273,
306,
7707,
12,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1198,
67,
5414,
67,
3319,
12,
2890,
4672,
5158,
1717,
4701,
848,
316,
326,
630,
3849,
333,
16,
1496,
8656,
1608,
82,
1404,
471,
12044,
1404,
4279,
1930,
1198,
400,
273,
306,
7707,
12,
... | |
EXAMPLES: We explicitly construct the alternating group on four elements. :: | EXAMPLES: We explicitly construct the alternating group on four elements:: | def _latex_(self): r""" Method for describing self in LaTeX. Encapsulates ``self.gens()`` in angle brackets to denote that self in generated by these elements. Called by the ``latex()`` function. EXAMPLES: We explicitly construct the alternating group on four elements. :: sage: A4 = PermutationGroup([[(1,2,3)],[(2,3,4)]]); A4 Permutation Group with generators [(2,3,4), (1,2,3)] sage: latex(A4) \langle (2,3,4), (1,2,3) \rangle sage: A4._latex_() '\\langle (2,3,4), (1,2,3) \\rangle' """ return '\\langle ' + \ ', '.join([x._latex_() for x in self.gens()]) + ' \\rangle' | 6ac30ecf5071d4fb909072e90d2311932d8c2165 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/6ac30ecf5071d4fb909072e90d2311932d8c2165/permgroup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
26264,
67,
12,
2890,
4672,
436,
8395,
2985,
364,
16868,
365,
316,
21072,
21575,
60,
18,
1374,
22984,
17099,
12176,
2890,
18,
23730,
1435,
10335,
316,
5291,
15597,
358,
5545,
1168,
716... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
26264,
67,
12,
2890,
4672,
436,
8395,
2985,
364,
16868,
365,
316,
21072,
21575,
60,
18,
1374,
22984,
17099,
12176,
2890,
18,
23730,
1435,
10335,
316,
5291,
15597,
358,
5545,
1168,
716... |
filter = '\n'.join([""" | filter = """ | def get_pyvalue(self): return generic_get_pyvalue(self) | 88f5c36fc2676649947284c1151aec84dabd0959 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3592/88f5c36fc2676649947284c1151aec84dabd0959/printutils.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
2074,
1132,
12,
2890,
4672,
327,
5210,
67,
588,
67,
2074,
1132,
12,
2890,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
2074,
1132,
12,
2890,
4672,
327,
5210,
67,
588,
67,
2074,
1132,
12,
2890,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
return None, 'cast (int %s to %s)'%(value, toptr) | return None, '%s cast (int %s to %s)'%(toptr, value, toptr) | def repr_constant(self, value): " returns node and repr as tuple " type_ = lltype.typeOf(value) if isinstance(type_, lltype.Primitive): repr = self.primitives.repr(type_, value) return None, "%s %s" % (self.repr_type(type_), repr) | 5e526fd22a01215f3b086f2c830d9d4d363d9e49 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6934/5e526fd22a01215f3b086f2c830d9d4d363d9e49/database.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8480,
67,
14384,
12,
2890,
16,
460,
4672,
315,
1135,
756,
471,
8480,
487,
3193,
315,
618,
67,
273,
6579,
723,
18,
723,
951,
12,
1132,
13,
309,
1549,
12,
723,
67,
16,
6579,
723,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8480,
67,
14384,
12,
2890,
16,
460,
4672,
315,
1135,
756,
471,
8480,
487,
3193,
315,
618,
67,
273,
6579,
723,
18,
723,
951,
12,
1132,
13,
309,
1549,
12,
723,
67,
16,
6579,
723,
18,
... |
if sunscription is None: | if subscription is None: | def add_user_subscription(self, user, subscription_type, subscription_id): user_subscription = model.UserSubscription(user=user, user_nickname=user.nickname, user_email=user.email, subscription_type=subscription_type, subscription_id=subscription_id) subscription = model.UserSubscription.all().filter('user', user).filter('subscription_type', subscription_type).filter('subscription_id', subscription_id).get() if sunscription is None: user_subscription.put() | 73e8022c2475a389c859c952f047ad5026dde744 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/2580/73e8022c2475a389c859c952f047ad5026dde744/BaseHandler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
527,
67,
1355,
67,
11185,
12,
2890,
16,
729,
16,
4915,
67,
723,
16,
4915,
67,
350,
4672,
729,
67,
11185,
273,
938,
18,
1299,
6663,
12,
1355,
33,
1355,
16,
729,
67,
17091,
529,
33,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
527,
67,
1355,
67,
11185,
12,
2890,
16,
729,
16,
4915,
67,
723,
16,
4915,
67,
350,
4672,
729,
67,
11185,
273,
938,
18,
1299,
6663,
12,
1355,
33,
1355,
16,
729,
67,
17091,
529,
33,
... |
commit=tpc_abort=tpc_begin | commit=tpc_begin | def tpc_begin(self, *ignored): pass | 1a0ea1b2fc6b9c821cf5e267d6c648ce97d43564 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9658/1a0ea1b2fc6b9c821cf5e267d6c648ce97d43564/TM.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
268,
2436,
67,
10086,
12,
2890,
16,
380,
24055,
4672,
1342,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
268,
2436,
67,
10086,
12,
2890,
16,
380,
24055,
4672,
1342,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
tags[self.tagnum]['title'] = fileref.tag().title | tags[self.tagnum]['title'] = tag.title | def tags_win_update(self, window, tags, entries, entries_names): self.updating_edit_entries = True # Populate tags(). Note that we only retrieve info from the # file if the info hasn't already been changed: fileref = tagpy.FileRef(tags[self.tagnum]['fullpath']) if not tags[self.tagnum]['title-changed']: tags[self.tagnum]['title'] = fileref.tag().title if not tags[self.tagnum]['artist-changed']: tags[self.tagnum]['artist'] = fileref.tag().artist if not tags[self.tagnum]['album-changed']: tags[self.tagnum]['album'] = fileref.tag().album if not tags[self.tagnum]['year-changed']: tags[self.tagnum]['year'] = fileref.tag().year if not tags[self.tagnum]['track-changed']: tags[self.tagnum]['track'] = fileref.tag().track if not tags[self.tagnum]['genre-changed']: tags[self.tagnum]['genre'] = fileref.tag().genre if not tags[self.tagnum]['comment-changed']: tags[self.tagnum]['comment'] = fileref.tag().comment # Update interface: entries[0].set_text(self.tags_get_tag(tags[self.tagnum], 'title')) entries[1].set_text(self.tags_get_tag(tags[self.tagnum], 'artist')) entries[2].set_text(self.tags_get_tag(tags[self.tagnum], 'album')) if self.tags_get_tag(tags[self.tagnum], 'year') != 0: entries[3].set_text(str(self.tags_get_tag(tags[self.tagnum], 'year'))) else: entries[3].set_text('') if self.tags_get_tag(tags[self.tagnum], 'track') != 0: entries[4].set_text(str(self.tags_get_tag(tags[self.tagnum], 'track'))) else: entries[4].set_text('') entries[5].set_text(self.tags_get_tag(tags[self.tagnum], 'genre')) entries[6].set_text(self.tags_get_tag(tags[self.tagnum], 'comment')) self.curr_mpdpath = gobject.filename_display_name(tags[self.tagnum]['mpdpath']) filename = self.curr_mpdpath if not self.use_mpdpaths: filename = os.path.basename(filename) entries[7].set_text(filename) entries[0].select_region(0, len(entries[0].get_text())) entries[0].grab_focus() window.set_title(_("Edit Tags") + " - " + str(self.tagnum+1) + " " + _("of") + " " + str(len(tags))) self.updating_edit_entries = False # Update text colors as appropriate: for i in range(len(entries)-1): if tags[self.tagnum][entries_names[i] + '-changed']: self.tags_win_entry_changed(entries[i]) else: self.tags_win_entry_revert_color(entries[i]) self.tags_win_set_sensitive(window.action_area) | f46c857512301d1836864c51b4a0d938e9c1cf02 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/2312/f46c857512301d1836864c51b4a0d938e9c1cf02/tagedit.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2342,
67,
8082,
67,
2725,
12,
2890,
16,
2742,
16,
2342,
16,
3222,
16,
3222,
67,
1973,
4672,
365,
18,
5533,
1776,
67,
4619,
67,
8219,
273,
1053,
468,
22254,
2342,
7675,
3609,
716,
732,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2342,
67,
8082,
67,
2725,
12,
2890,
16,
2742,
16,
2342,
16,
3222,
16,
3222,
67,
1973,
4672,
365,
18,
5533,
1776,
67,
4619,
67,
8219,
273,
1053,
468,
22254,
2342,
7675,
3609,
716,
732,
... |
for field in self.base_fields.iteritems(): if field.required: field.required = makeRequired | for field in self.base_fields.itervalues(): if field.required: field.required = self.makeRequired | def __init__(self, user = None, *args, **kwargs): if user is None or not (hasattr(user, 'other_user') and user.other_user): self.makeRequired = True else: self.makeRequired = False # set a few things... self.base_fields['e_mail'].widget.attrs['size'] = 25 | 8017ee2dce98074784d46c6e4d0bb3a35a009b75 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12041/8017ee2dce98074784d46c6e4d0bb3a35a009b75/user_profile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
729,
273,
599,
16,
380,
1968,
16,
2826,
4333,
4672,
309,
729,
353,
599,
578,
486,
261,
5332,
1747,
12,
1355,
16,
296,
3011,
67,
1355,
6134,
471,
729,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
729,
273,
599,
16,
380,
1968,
16,
2826,
4333,
4672,
309,
729,
353,
599,
578,
486,
261,
5332,
1747,
12,
1355,
16,
296,
3011,
67,
1355,
6134,
471,
729,
1... |
files = ['$(OPT)', config_c, frozen_c, frozenmain_c] + \ addfiles + libs + \ | files = ['$(OPT)', config_c, frozen_c] + \ supp_sources + addfiles + libs + \ | def main(): # overridable context prefix = PREFIX # settable with -p option extensions = [] path = sys.path # output files frozen_c = 'frozen.c' config_c = 'config.c' target = 'a.out' # normally derived from script name makefile = 'Makefile' # parse command line try: opts, args = getopt.getopt(sys.argv[1:], 'e:p:') except getopt.error, msg: usage('getopt error: ' + str(msg)) # proces option arguments for o, a in opts: if o == '-e': extensions.append(a) if o == '-p': prefix = a # locations derived from options binlib = os.path.join(prefix, 'lib/python/lib') incldir = os.path.join(prefix, 'include/Py') config_c_in = os.path.join(binlib, 'config.c.in') frozenmain_c = os.path.join(binlib, 'frozenmain.c') makefile_in = os.path.join(binlib, 'Makefile') defines = ['-DHAVE_CONFIG_H', '-DUSE_FROZEN', '-DNO_MAIN', '-DPYTHONPATH=\\"$(PYTHONPATH)\\"'] includes = ['-I' + incldir, '-I' + binlib] # sanity check of directories and files for dir in [prefix, binlib, incldir] + extensions: if not os.path.exists(dir): usage('needed directory %s not found' % dir) if not os.path.isdir(dir): usage('%s: not a directory' % dir) for file in config_c_in, makefile_in, frozenmain_c: if not os.path.exists(file): usage('needed file %s not found' % file) if not os.path.isfile(file): usage('%s: not a plain file' % file) for dir in extensions: setup = os.path.join(dir, 'Setup') if not os.path.exists(setup): usage('needed file %s not found' % setup) if not os.path.isfile(setup): usage('%s: not a plain file' % setup) # check that enough arguments are passed if not args: usage('at least one filename argument required') # check that file arguments exist for arg in args: if not os.path.exists(arg): usage('argument %s not found' % arg) if not os.path.isfile(arg): usage('%s: not a plain file' % arg) # process non-option arguments scriptfile = args[0] modules = args[1:] # derive target name from script name base = os.path.basename(scriptfile) base, ext = os.path.splitext(base) if base: if base != scriptfile: target = base else: target = base + '.bin' # Actual work starts here... dict = findmodules.findmodules(scriptfile, modules, path) backup = frozen_c + '~' try: os.rename(frozen_c, backup) except os.error: backup = None outfp = open(frozen_c, 'w') try: makefreeze.makefreeze(outfp, dict) finally: outfp.close() if backup: if cmp.cmp(backup, frozen_c): sys.stderr.write('%s not changed, not written\n' % frozen_c) os.rename(backup, frozen_c) builtins = [] unknown = [] mods = dict.keys() mods.sort() for mod in mods: if dict[mod] == '<builtin>': builtins.append(mod) elif dict[mod] == '<unknown>': unknown.append(mod) addfiles = [] if unknown: addfiles, addmods = \ checkextensions.checkextensions(unknown, extensions) for mod in addmods: unknown.remove(mod) builtins = builtins + addmods if unknown: sys.stderr.write('Warning: unknown modules remain: %s\n' % string.join(unknown)) builtins.sort() infp = open(config_c_in) backup = config_c + '~' try: os.rename(config_c, backup) except os.error: backup = None outfp = open(config_c, 'w') try: makeconfig.makeconfig(infp, outfp, builtins) finally: outfp.close() infp.close() if backup: if cmp.cmp(backup, config_c): sys.stderr.write('%s not changed, not written\n' % config_c) os.rename(backup, config_c) cflags = defines + includes + ['$(OPT)'] libs = [] for n in 'Modules', 'Python', 'Objects', 'Parser': n = 'lib%s.a' % n n = os.path.join(binlib, n) libs.append(n) makevars = parsesetup.getmakevars(makefile_in) somevars = {} for key in makevars.keys(): somevars[key] = makevars[key] somevars['CFLAGS'] = string.join(cflags) # override files = ['$(OPT)', config_c, frozen_c, frozenmain_c] + \ addfiles + libs + \ ['$(MODLIBS)', '$(LIBS)', '$(SYSLIBS)'] backup = makefile + '~' try: os.rename(makefile, backup) except os.error: backup = None outfp = open(makefile, 'w') try: makemakefile.makemakefile(outfp, somevars, files, target) finally: outfp.close() if backup: if not cmp.cmp(backup, makefile): print 'previous Makefile saved as', backup else: sys.stderr.write('%s not changed, not written\n' % makefile) os.rename(backup, makefile) # Done! print 'Now run make to build the target:', target | f6f28ea9023017ff20e8efaf99b071d5e906c352 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/f6f28ea9023017ff20e8efaf99b071d5e906c352/freeze.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
468,
5713,
27621,
819,
1633,
273,
17154,
1082,
202,
7,
444,
2121,
598,
300,
84,
1456,
4418,
273,
5378,
589,
273,
2589,
18,
803,
225,
468,
876,
1390,
12810,
67,
71,
273,
29... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
468,
5713,
27621,
819,
1633,
273,
17154,
1082,
202,
7,
444,
2121,
598,
300,
84,
1456,
4418,
273,
5378,
589,
273,
2589,
18,
803,
225,
468,
876,
1390,
12810,
67,
71,
273,
29... |
ikey = _ustr(tok[0]).strip() | if len(tok) == 0: continue ikey = tok[0] if isinstance(ikey,int): ikey = _ustr(tok[0]).strip() | def postParse( self, instring, loc, tokenlist ): for i,tok in enumerate(tokenlist): ikey = _ustr(tok[0]).strip() if len(tok)==1: tokenlist[ikey] = _ParseResultsWithOffset("",i) elif len(tok)==2 and not isinstance(tok[1],ParseResults): tokenlist[ikey] = _ParseResultsWithOffset(tok[1],i) else: dictvalue = tok.copy() #ParseResults(i) del dictvalue[0] if len(dictvalue)!= 1 or (isinstance(dictvalue,ParseResults) and dictvalue.keys()): tokenlist[ikey] = _ParseResultsWithOffset(dictvalue,i) else: tokenlist[ikey] = _ParseResultsWithOffset(dictvalue[0],i) | 8b7a5faecf2a697c2e60cea5d49043ec1eb7fe03 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12364/8b7a5faecf2a697c2e60cea5d49043ec1eb7fe03/pyparsing.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1603,
3201,
12,
365,
16,
316,
1080,
16,
1515,
16,
1147,
1098,
262,
30,
364,
277,
16,
17692,
316,
4241,
12,
2316,
1098,
4672,
309,
562,
12,
17692,
13,
422,
374,
30,
1324,
277,
856,
27... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1603,
3201,
12,
365,
16,
316,
1080,
16,
1515,
16,
1147,
1098,
262,
30,
364,
277,
16,
17692,
316,
4241,
12,
2316,
1098,
4672,
309,
562,
12,
17692,
13,
422,
374,
30,
1324,
277,
856,
27... |
def process_rs(self, service, backup=False): | def process_rs(self, owner, content, service, backup=False): | def process_rs(self, service, backup=False): """ Process data for a given virtual server and real server. | b7423ef5c8954a65153886a58999ac3d4a0e1ffd /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/13049/b7423ef5c8954a65153886a58999ac3d4a0e1ffd/cs.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
67,
5453,
12,
2890,
16,
3410,
16,
913,
16,
1156,
16,
5114,
33,
8381,
4672,
3536,
4389,
501,
364,
279,
864,
5024,
1438,
471,
2863,
1438,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
67,
5453,
12,
2890,
16,
3410,
16,
913,
16,
1156,
16,
5114,
33,
8381,
4672,
3536,
4389,
501,
364,
279,
864,
5024,
1438,
471,
2863,
1438,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,... |
return (isinstance(ex, boto.exception.BotoServerError) and ('Throttling' in ex.body or 'RequestExpired' in ex.body or 'Connection reset by peer' in ex.body)) | return ((isinstance(ex, boto.exception.BotoServerError) and ('Throttling' in ex.body or 'RequestExpired' in ex.body)) or (isinstance(ex, socket.error) and ex.args == (104, 'Connection reset by peer'))) | def retry_if(ex): """Retry if we get a server error indicating throttling.""" return (isinstance(ex, boto.exception.BotoServerError) and ('Throttling' in ex.body or 'RequestExpired' in ex.body or 'Connection reset by peer' in ex.body)) | 64f0ed09a1cc0ec105770a045e3cb70d03183c8f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14504/64f0ed09a1cc0ec105770a045e3cb70d03183c8f/emr.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3300,
67,
430,
12,
338,
4672,
3536,
7539,
309,
732,
336,
279,
1438,
555,
11193,
30700,
12123,
327,
261,
291,
1336,
12,
338,
16,
17228,
18,
4064,
18,
38,
6302,
8777,
13,
471,
7707,
1623... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3300,
67,
430,
12,
338,
4672,
3536,
7539,
309,
732,
336,
279,
1438,
555,
11193,
30700,
12123,
327,
261,
291,
1336,
12,
338,
16,
17228,
18,
4064,
18,
38,
6302,
8777,
13,
471,
7707,
1623... |
if (uid is None or ((not self._show_private) and uid.is_private()) or (not uid.is_variable() and not self._docmap.has_key(uid)) or (uid.is_variable() and not self._docmap.has_key(uid.parent()))): | if not self._documented(uid): | def _uid_to_href(self, uid, label=None, css_class=None, code=1): """ @return: The HTML code to link to the given UID. This code consists of an anchor with an href to the page for C{uid}. If C{label} is not C{None}, then it is used as the text for the link; otherwise, C{uid} is used as the text. If C{uid} is not documented, then the HTML code will just contain the name, and no href. @rtype: C{string} @type uid: L{uid.UID} @type label: C{string} @param code: Whether or not to include C{<code>...</code>} tags around the label. """ # We shouldn't need this, but include it just in case. if uid is None and label is None: return '<code>??</code>' # Find a default value for the label. if label is None: label = uid.name() | 54df16a310cfda72ff9768115cf38b98d8668613 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3512/54df16a310cfda72ff9768115cf38b98d8668613/html.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1911,
67,
869,
67,
7547,
12,
2890,
16,
4555,
16,
1433,
33,
7036,
16,
3747,
67,
1106,
33,
7036,
16,
981,
33,
21,
4672,
3536,
632,
2463,
30,
1021,
3982,
981,
358,
1692,
358,
326,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1911,
67,
869,
67,
7547,
12,
2890,
16,
4555,
16,
1433,
33,
7036,
16,
3747,
67,
1106,
33,
7036,
16,
981,
33,
21,
4672,
3536,
632,
2463,
30,
1021,
3982,
981,
358,
1692,
358,
326,
... |
spacing.append(1.5) | try: spacing.append(1.5) except(AttributeError): spacing = [1, 1, 1] | def SetParser(self, parser): self.level = parser.GetImageWindowLevel() self.window = parser.GetImageWindowWidth() self.position = parser.GetImagePosition() self.number = parser.GetImageNumber() self.spacing = spacing = parser.GetPixelSpacing() self.orientation_label = parser.GetImageOrientationLabel() self.file = parser.filename | 7d09e69fc82b0fdaff30d3a43e58f51de0cbcd41 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10228/7d09e69fc82b0fdaff30d3a43e58f51de0cbcd41/dicom.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1000,
2678,
12,
2890,
16,
2082,
4672,
365,
18,
2815,
273,
2082,
18,
967,
2040,
3829,
2355,
1435,
365,
18,
5668,
273,
2082,
18,
967,
2040,
3829,
2384,
1435,
365,
18,
3276,
273,
2082,
18... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1000,
2678,
12,
2890,
16,
2082,
4672,
365,
18,
2815,
273,
2082,
18,
967,
2040,
3829,
2355,
1435,
365,
18,
5668,
273,
2082,
18,
967,
2040,
3829,
2384,
1435,
365,
18,
3276,
273,
2082,
18... |
self.assert_(type(result) is dict, "result is a %s" % type(result)) filename = result['path'] self.assert_(isinstance(filename,str) or \ None == filename, str(result)) if isinstance(filename, str) and \ not filename[0:2] == 'HK': self.assert_ (os.path.lexists(filename), \ "In backend '%s' path does not exist: '%s' \ " % (key, filename)) | self.validate_result(result) | def test_get_commands(self): for key in sorted(backends): print "debug: test_get_commands: key='%s'" % (key, ) for (option_id, __name) in backends[key].get_options(): for cmd in backends[key].get_commands(option_id): for result in cmd.execute(really_delete = False): if result != True: break self.assert_(type(result) is dict, "result is a %s" % type(result)) filename = result['path'] self.assert_(isinstance(filename,str) or \ None == filename, str(result)) if isinstance(filename, str) and \ not filename[0:2] == 'HK': self.assert_ (os.path.lexists(filename), \ "In backend '%s' path does not exist: '%s' \ " % (key, filename)) # make sure trash and tmp don't return the same results if 'nt' == os.name: return def get_files(option_id): ret = [] for cmd in backends['system'].get_commands(option_id): result = cmd.execute(False).next() ret.append(result['path']) return ret trash_paths = get_files('trash') tmp_paths = get_files('tmp') for tmp_path in tmp_paths: self.assert_(tmp_path not in trash_paths) | 2ea7ae93957d422428c7d4c6765eb193280cd40e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7853/2ea7ae93957d422428c7d4c6765eb193280cd40e/Cleaner.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
588,
67,
7847,
12,
2890,
4672,
364,
498,
316,
3115,
12,
823,
5839,
4672,
1172,
315,
4148,
30,
1842,
67,
588,
67,
7847,
30,
498,
28713,
87,
4970,
738,
261,
856,
16,
262,
364... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
588,
67,
7847,
12,
2890,
4672,
364,
498,
316,
3115,
12,
823,
5839,
4672,
1172,
315,
4148,
30,
1842,
67,
588,
67,
7847,
30,
498,
28713,
87,
4970,
738,
261,
856,
16,
262,
364... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.