rem
stringlengths
2
226k
add
stringlengths
0
227k
context
stringlengths
8
228k
meta
stringlengths
156
215
input_ids
list
attention_mask
list
labels
list
pv = self.packageVersionApi.create(p.packageid, info['epoch'],
pv = self.packageVersionApi.create(p["packageid"], info['epoch'],
def _add_packages_from_dir(self, dir, repo): dirList = os.listdir(dir) packages = repo['packages'] package_count = 0 for fname in dirList: if (fname.endswith(".rpm")): try: info = getRPMInformation(dir + fname) #p = self.packageApi.package(info['name']) p = self.package(repo["id"], info['name']) if not p: p = self.create_package(repo["id"], info['name'], info['description']) pv = self.packageVersionApi.create(p.packageid, info['epoch'], info['version'], info['release'], info['arch']) for dep in info['requires']: pv.requires.append(dep) for dep in info['provides']: pv.provides.append(dep) self.packageVersionApi.update(pv) p.versions.append(pv) #self.packageApi.update(p) packages[p["packageid"]] = p log.debug("repo = %s" % (repo)) log.debug("packages = %s" % (packages)) self.update(repo) # Need to send package_count = package_count + 1 log.debug("Repo <%s> added package <%s> with %s versions" % (repo["id"], p["packageid"], len(p["versions"]))) except Exception, e: log.debug("%s" % (traceback.format_exc())) log.error("error reading package %s" % (dir + fname)) log.debug("read [%s] packages" % package_count) self._read_comps_xml(dir, repo)
b63b9f5dccf9939a1050996b70144329176bccfc /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10692/b63b9f5dccf9939a1050996b70144329176bccfc/repo.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 1289, 67, 10308, 67, 2080, 67, 1214, 12, 2890, 16, 1577, 16, 3538, 4672, 1577, 682, 273, 1140, 18, 1098, 1214, 12, 1214, 13, 5907, 273, 3538, 3292, 10308, 3546, 2181, 67, 1883, 27...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 1289, 67, 10308, 67, 2080, 67, 1214, 12, 2890, 16, 1577, 16, 3538, 4672, 1577, 682, 273, 1140, 18, 1098, 1214, 12, 1214, 13, 5907, 273, 3538, 3292, 10308, 3546, 2181, 67, 1883, 27...
"paragraph", "subparagraph", "description", "opcodedesc", "classdesc", "funcdesc", "methoddesc", "excdesc", "datadesc", "funcdescni", "methoddescni", "excdescni", "datadescni",
"paragraph", "subparagraph", "excdesc", "datadesc", "excdescni", "datadescni", ) RECURSE_INTO_PARA_CONTAINERS = ( "chapter", "section", "subsection", "subsubsection", "paragraph", "subparagraph", "abstract", "memberdesc", "memberdescni", "datadesc", "datadescni",
def move_elements_by_name(doc, source, dest, name, sep=None): nodes = [] for child in source.childNodes: if child.nodeType == xml.dom.core.ELEMENT and child.tagName == name: nodes.append(child) for node in nodes: source.removeChild(node) dest.appendChild(node) if sep: dest.appendChild(doc.createTextNode(sep))
6fd3b456df1736e218c2900c3479631306a4f4d7 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/6fd3b456df1736e218c2900c3479631306a4f4d7/docfixer.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3635, 67, 6274, 67, 1637, 67, 529, 12, 2434, 16, 1084, 16, 1570, 16, 508, 16, 5478, 33, 7036, 4672, 2199, 273, 5378, 364, 1151, 316, 1084, 18, 3624, 3205, 30, 309, 1151, 18, 2159, 55...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3635, 67, 6274, 67, 1637, 67, 529, 12, 2434, 16, 1084, 16, 1570, 16, 508, 16, 5478, 33, 7036, 4672, 2199, 273, 5378, 364, 1151, 316, 1084, 18, 3624, 3205, 30, 309, 1151, 18, 2159, 55...
this = apply(_quickfix.new_MDReqRejReason, args)
this = _quickfix.new_MDReqRejReason(*args)
def __init__(self, *args): this = apply(_quickfix.new_MDReqRejReason, args) try: self.this.append(this) except: self.this = this
7e632099fd421880c8c65fb0cf610d338d115ee9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8819/7e632099fd421880c8c65fb0cf610d338d115ee9/quickfix.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 380, 1968, 4672, 333, 273, 389, 19525, 904, 18, 2704, 67, 6188, 6113, 426, 78, 8385, 30857, 1968, 13, 775, 30, 365, 18, 2211, 18, 6923, 12, 2211, 13, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 380, 1968, 4672, 333, 273, 389, 19525, 904, 18, 2704, 67, 6188, 6113, 426, 78, 8385, 30857, 1968, 13, 775, 30, 365, 18, 2211, 18, 6923, 12, 2211, 13, 1...
j,k = 1,0
j, k = 1, 0
def validedChanges(*args): """Check for a least one selected kana portion (display of a message if not the case), catch parameters, then close the window.
64cfb9e0b60a3a976c72fa9d5d722987641133b9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3073/64cfb9e0b60a3a976c72fa9d5d722987641133b9/gtk_gui.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1244, 13898, 7173, 30857, 1968, 4672, 3536, 1564, 364, 279, 4520, 1245, 3170, 417, 13848, 14769, 261, 5417, 434, 279, 883, 309, 486, 326, 648, 3631, 1044, 1472, 16, 1508, 1746, 326, 2742, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1244, 13898, 7173, 30857, 1968, 4672, 3536, 1564, 364, 279, 4520, 1245, 3170, 417, 13848, 14769, 261, 5417, 434, 279, 883, 309, 486, 326, 648, 3631, 1044, 1472, 16, 1508, 1746, 326, 2742, ...
1/0
1./0
def f(p): 1/0
33715f480fc70c2d6211595f77c5fb72a0d256c5 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8125/33715f480fc70c2d6211595f77c5fb72a0d256c5/test_profilehooks.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 284, 12, 84, 4672, 404, 19, 20, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 284, 12, 84, 4672, 404, 19, 20, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
print data
sys.stdout.write(data) session_id = 1
def h(self,web,data): web.header('Content-Type', 'application/x-dmap-tagged') web.header('DAAP-Server', 'Simple') web.header('Expires', '-1') web.header('Cache-Control', 'no-cache') web.header('Accept-Ranges', 'bytes') web.header('Content-Language', 'en_us') if (hasattr(data, 'next')): try: web.header("Content-Length", str(os.stat(data.name).st_size)) except: pass return data else: try: web.header("Content-Length", str(len(data))) except: pass print data
7f1700be343da028c435d27dd80f8aeb85c12223 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8691/7f1700be343da028c435d27dd80f8aeb85c12223/server.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 366, 12, 2890, 16, 4875, 16, 892, 4672, 3311, 18, 3374, 2668, 1350, 17, 559, 2187, 296, 3685, 19, 92, 17, 72, 1458, 17, 23023, 6134, 3311, 18, 3374, 2668, 9793, 2203, 17, 2081, 2187, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 366, 12, 2890, 16, 4875, 16, 892, 4672, 3311, 18, 3374, 2668, 1350, 17, 559, 2187, 296, 3685, 19, 92, 17, 72, 1458, 17, 23023, 6134, 3311, 18, 3374, 2668, 9793, 2203, 17, 2081, 2187, ...
points::
points, though which one creates more points is heavily dependent upon the function being plotted. ::
def adaptive_refinement(f, p1, p2, adaptive_tolerance=0.01, adaptive_recursion=5, level=0): r""" The adaptive refinement algorithm for plotting a function f. See the docstring for plot for a description of the algorithm. INPUT: - ``f`` - a function of one variable - ``p1, p2`` - two points to refine between - ``adaptive_recursion`` - (default: 5) how many levels of recursion to go before giving up when doing adaptive refinement. Setting this to 0 disables adaptive refinement. - ``adaptive_tolerance`` - (default: 0.01) how large a difference should be before the adaptive refinement code considers it significant. See the documentation for plot() for more information. OUTPUT: - ``list`` - a list of points to insert between p1 and p2 to get a better linear approximation between them TESTS:: sage: from sage.plot.plot import adaptive_refinement sage: adaptive_refinement(sin, (0,0), (pi,0), adaptive_tolerance=0.01, adaptive_recursion=0) [] sage: adaptive_refinement(sin, (0,0), (pi,0), adaptive_tolerance=0.01) [(0.125*pi, 0.38268343236508978), (0.1875*pi, 0.55557023301960218), (0.25*pi, 0.707106781186547...), (0.3125*pi, 0.831469612302545...), (0.375*pi, 0.92387953251128674), (0.4375*pi, 0.98078528040323043), (0.5*pi, 1.0), (0.5625*pi, 0.98078528040323043), (0.625*pi, 0.92387953251128674), (0.6875*pi, 0.831469612302545...), (0.75*pi, 0.70710678118654757), (0.8125*pi, 0.55557023301960218), (0.875*pi, 0.3826834323650898...)] This shows that lowering adaptive_tolerance and raising adaptive_recursion both increase the number of subdivision points:: sage: x = var('x') sage: f(x) = sin(1/x) sage: n1 = len(adaptive_refinement(f, (0,0), (pi,0), adaptive_tolerance=0.01)); n1 15 sage: n2 = len(adaptive_refinement(f, (0,0), (pi,0), adaptive_recursion=10, adaptive_tolerance=0.01)); n2 79 sage: n3 = len(adaptive_refinement(f, (0,0), (pi,0), adaptive_tolerance=0.001)); n3 26 """ if level >= adaptive_recursion: return [] x = (p1[0] + p2[0])/2.0 try: y = float(f(x)) except (ZeroDivisionError, TypeError, ValueError, OverflowError), msg: sage.misc.misc.verbose("%s\nUnable to compute f(%s)"%(msg, x), 1) # give up for this branch return [] # this distance calculation is not perfect. if abs((p1[1] + p2[1])/2.0 - y) > adaptive_tolerance: return adaptive_refinement(f, p1, (x, y), adaptive_tolerance=adaptive_tolerance, adaptive_recursion=adaptive_recursion, level=level+1) \ + [(x, y)] + \ adaptive_refinement(f, (x, y), p2, adaptive_tolerance=adaptive_tolerance, adaptive_recursion=adaptive_recursion, level=level+1) else: return []
aa6916134b0b3764422bd53e781315007a519805 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/aa6916134b0b3764422bd53e781315007a519805/plot.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5855, 688, 67, 266, 926, 820, 12, 74, 16, 293, 21, 16, 293, 22, 16, 5855, 688, 67, 25456, 33, 20, 18, 1611, 16, 5855, 688, 67, 31347, 33, 25, 16, 1801, 33, 20, 4672, 436, 8395, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5855, 688, 67, 266, 926, 820, 12, 74, 16, 293, 21, 16, 293, 22, 16, 5855, 688, 67, 25456, 33, 20, 18, 1611, 16, 5855, 688, 67, 31347, 33, 25, 16, 1801, 33, 20, 4672, 436, 8395, 1...
self.config['g_series'] = cfile['file_seriesname']+'.%(ext)s'
self.config['g_series'] = cfile['file_seriesname']+self.graphic_suffix[rel_type]+'.%(ext)s'
def _getSecondarySourceGraphics(self, cfile, graphic_type, watched=False): '''Download from secondary source such as movieposter.com return None return full qualified path and filename of downloaded graphic ''' if not len(self.config['myth_secondary_sources']): return None
147cd3ef0bfc1beba805d496a3aeed4923b921a1 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/13713/147cd3ef0bfc1beba805d496a3aeed4923b921a1/jamu.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 588, 14893, 1830, 17558, 12, 2890, 16, 276, 768, 16, 19548, 67, 723, 16, 23135, 33, 8381, 4672, 9163, 7109, 628, 9946, 1084, 4123, 487, 7344, 522, 881, 29811, 18, 832, 327, 599, 3...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 588, 14893, 1830, 17558, 12, 2890, 16, 276, 768, 16, 19548, 67, 723, 16, 23135, 33, 8381, 4672, 9163, 7109, 628, 9946, 1084, 4123, 487, 7344, 522, 881, 29811, 18, 832, 327, 599, 3...
result.append(newSpec)
def readSpecFile(filename): """ _readSpecFile_ Tool for extracting multiple RequestSpecs from a file Returns a list of spec instances found """ result = [] # ignore failed parsed requests ignore = 0 node = loadIMProvFile(filename) node = loadIMProvFile(filename) specQ = IMProvQuery("RequestSpec") specNodes = specQ(node) for snode in specNodes: newSpec = RequestSpec() try: newSpec.load(snode) except Exception,ex: logging.debug('ERROR loading a requestspec, will ignore this requestspec :'+str(ex)) ignore+=1 result.append(newSpec) return (result,ignore)
3869ae41eb2148ee2f5f65c38a278dfa90c684e0 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8886/3869ae41eb2148ee2f5f65c38a278dfa90c684e0/RequestSpec.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 855, 1990, 812, 12, 3459, 4672, 3536, 389, 896, 1990, 812, 67, 225, 13288, 364, 27117, 3229, 1567, 15999, 628, 279, 585, 225, 2860, 279, 666, 434, 857, 3884, 1392, 3536, 563, 273, 5378, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 855, 1990, 812, 12, 3459, 4672, 3536, 389, 896, 1990, 812, 67, 225, 13288, 364, 27117, 3229, 1567, 15999, 628, 279, 585, 225, 2860, 279, 666, 434, 857, 3884, 1392, 3536, 563, 273, 5378, ...
PyErr_Format(PyExc_TypeError, "%.400s constructor takes either 0 or " "%d positional argument%s", Py_TYPE(self)->tp_name,
PyErr_Format(PyExc_TypeError, "%.400s constructor takes %s" "%" PY_FORMAT_SIZE_T "d positional argument%s", Py_TYPE(self)->tp_name, numfields == 0 ? "" : "either 0 or ",
def visitModule(self, mod): self.emit("""
549a2cbf7e049de06b3d8b45dc419577921ed25b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8125/549a2cbf7e049de06b3d8b45dc419577921ed25b/asdl_c.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3757, 3120, 12, 2890, 16, 681, 4672, 365, 18, 18356, 2932, 3660, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3757, 3120, 12, 2890, 16, 681, 4672, 365, 18, 18356, 2932, 3660, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
except ValueError: print "Cannot extract TLS group number."
except ValueError: print "Cannot extract TLS group number:\n ", record, file_name return []
def extract_tls_parameters(remark_3_records):
c7414c7bd0f10baaa7d702b1454c3d4335e68afe /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/696/c7414c7bd0f10baaa7d702b1454c3d4335e68afe/remark_3_interpretation.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2608, 67, 17116, 67, 3977, 12, 2764, 1313, 67, 23, 67, 7094, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2608, 67, 17116, 67, 3977, 12, 2764, 1313, 67, 23, 67, 7094, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -1...
(r"(\s|.)", Name.Builtin),
(r"(\s|.)", Name.Builtin)
def word_callback(lexer, match): word = match.group()
cfe64ea65eacd18eae0cdcbaec6bc03b39d3ea28 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/2885/cfe64ea65eacd18eae0cdcbaec6bc03b39d3ea28/other.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2076, 67, 3394, 12, 31731, 16, 845, 4672, 2076, 273, 845, 18, 1655, 1435, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2076, 67, 3394, 12, 31731, 16, 845, 4672, 2076, 273, 845, 18, 1655, 1435, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
self.annotation.fromDom(i)
self.annotation.fromDom(contents[indx]) elif (component == 'attribute'): if contents[indx].hasattr('name'): content.append(AttributeDeclaration()) elif contents[indx].hasattr('ref'): content.append(AttributeReference()) else: raise SchemaError, 'Unknown attribute type' content[-1].fromDom(contents[indx]) elif (component == 'attributeGroup'): content.append(AttributeGroupReference()) content[-1].fromDom(contents[indx]) elif (component == 'anyAttribute') and (len(contents) == x+1): content.append(AttributeWildCard()) content[-1].fromDom(contents[indx])
def fromDom(self, node): self.setAttributes(node) contents = self.getContents(node)
66c6baf2288f065b45db940b63cfc5dba1aa9fb0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/13054/66c6baf2288f065b45db940b63cfc5dba1aa9fb0/XMLSchema.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 628, 8832, 12, 2890, 16, 756, 4672, 365, 18, 542, 2498, 12, 2159, 13, 2939, 273, 365, 18, 588, 6323, 12, 2159, 13, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 628, 8832, 12, 2890, 16, 756, 4672, 365, 18, 542, 2498, 12, 2159, 13, 2939, 273, 365, 18, 588, 6323, 12, 2159, 13, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
tree = oe.mailbox.dbxTree(dbxStream, address, entries)
tree = oe_mailbox.dbxTree(dbxStream, address, entries)
def _convertOutlookExpressToMbox(self, content): """Check if the uploaded mailbox file is an Outlook Express DBX one.
b939df774e8b09ad51d2d92670490e6b1622f831 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9857/b939df774e8b09ad51d2d92670490e6b1622f831/UserInterface.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 6283, 1182, 7330, 424, 1028, 774, 49, 2147, 12, 2890, 16, 913, 4672, 3536, 1564, 309, 326, 9140, 14187, 585, 353, 392, 2976, 7330, 18482, 2383, 60, 1245, 18, 2, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 6283, 1182, 7330, 424, 1028, 774, 49, 2147, 12, 2890, 16, 913, 4672, 3536, 1564, 309, 326, 9140, 14187, 585, 353, 392, 2976, 7330, 18482, 2383, 60, 1245, 18, 2, -100, -100, -100, ...
mydir = (os.path.split(sys.argv[0]))[0]
try: mydir = os.path.join(os.environ['APPDATA']) except KeyError: mydir,x = os.path.split(sys.argv[0])
def __init__(self, frame): self.frame = frame self.log("Loading plugin handler") self.myUsername = self.frame.np.config.sections["server"]["login"] self.plugins = [] if WIN32: mydir = (os.path.split(sys.argv[0]))[0] self.plugindir = os.path.join(mydir, "plugins") else: self.plugindir = os.path.join(os.path.expanduser("~"),'.nicotine','plugins') if os.path.isdir(self.plugindir): self.load(self.plugindir) else: self.log("It appears '%s' is not a directory, not loading plugins." % self.plugindir)
24801fbfef00af6047bd97682bef614051766b3a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8738/24801fbfef00af6047bd97682bef614051766b3a/pluginsystem.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 2623, 4672, 365, 18, 3789, 273, 2623, 365, 18, 1330, 2932, 10515, 1909, 1838, 7923, 365, 18, 4811, 8575, 273, 365, 18, 3789, 18, 6782, 18, 1425, 18, 1165...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 2623, 4672, 365, 18, 3789, 273, 2623, 365, 18, 1330, 2932, 10515, 1909, 1838, 7923, 365, 18, 4811, 8575, 273, 365, 18, 3789, 18, 6782, 18, 1425, 18, 1165...
return s[:-4] + '...' return '...' + s[4:]
return s[:length-4] + ' ...' return '... ' + s[len(s)-length+4:]
def trunc(s, length, end=1): """Truncate a string s to length length, by cutting off the last (length-4) characters and replacing them with ' ...' With end=0, truncate from the front instead. """ if not s: return '' if len(s) > length: if end: return s[:-4] + '...' return '...' + s[4:] return s
d8165d152b789bb2f73e04c9c2311882ca86746f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2493/d8165d152b789bb2f73e04c9c2311882ca86746f/pbpscript.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6956, 12, 87, 16, 769, 16, 679, 33, 21, 4672, 3536, 25871, 279, 533, 272, 358, 769, 769, 16, 635, 6391, 1787, 3397, 326, 1142, 261, 2469, 17, 24, 13, 3949, 471, 13993, 2182, 598, 296...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6956, 12, 87, 16, 769, 16, 679, 33, 21, 4672, 3536, 25871, 279, 533, 272, 358, 769, 769, 16, 635, 6391, 1787, 3397, 326, 1142, 261, 2469, 17, 24, 13, 3949, 471, 13993, 2182, 598, 296...
dryfunc(options.dry, getCommandOut, "find \"%s\" -name .svn -exec rm -rf \"{}\"" % repospathofversionedfile)
dryfunc(options.dry, getPipedCommandOut, "find \"%s\" -name .svn -exec rm -rf \"{}\" \\;" % repospathofversionedfile)
def unversionthis(filetoversion): try: if accesscontrollist.hasacl(filetoversion) and not options.ignoreacl: err = "filetoversion has a 'deny' in ACL permissions (ls -lde %s: %s) \n \ This program is currently not clever enough to check if you have permission to move/delete this file. \n \ To avoid this problem remove deny permissions from the access control entries \n \ or rerun this command with --ignoreacl" % (filetoversion, accesscontrollist.getacl(filetoversion)) raise SyncherException(err) # TODO: verify that this file is not already added logging.info("should: check for dups") filetoversionpath, repospathofversionedfile, repospathtoputnewfilein = settings.getFileToVersionPathes(filetoversion) util.remove(filetoversionpath) if subversion.isinsvn(repospathofversionedfile): subversion.deleteinreposonly(repospathofversionedfile) if subversion.isinsvn(repospathofversionedfile): if options.autocommit: logging.info("committing delete of %s" % repospathofversionedfile) subversion.commit(repospathofversionedfile, "removed via unversionthis") else: raise SyncherException("repository will be messed up if we don't commit removal svn for this path") if os.path.isdir(repospathofversionedfile): logging.info("deleting .svn files in %s" % repospathofversionedfile) dryfunc(options.dry, getCommandOut, "find \"%s\" -name .svn -exec rm -rf \"{}\"" % repospathofversionedfile) util.removesymlink(filetoversionpath) acl = None if options.ignoreacl: acl = removeacl(repospathofversionedfile) util.move(repospathofversionedfile, filetoversionpath)#repospathtoputnewfilein) if acl is not None: accesscontrollist.setacl(filetoversion, acl) #created = util.makedirs(repospathtoputnewfilein) except Exception as e: logging.warn("ROLLING BACK because of %s" % e) undo.rollback() raise
4c733124dc4bf6d8fbe589d37fdcd75266656273 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10713/4c733124dc4bf6d8fbe589d37fdcd75266656273/unversionthis.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 640, 1589, 2211, 12, 7540, 278, 1527, 722, 4672, 775, 30, 309, 2006, 7098, 1098, 18, 5332, 10150, 12, 7540, 278, 1527, 722, 13, 471, 486, 702, 18, 6185, 10150, 30, 393, 273, 315, 7540,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 640, 1589, 2211, 12, 7540, 278, 1527, 722, 4672, 775, 30, 309, 2006, 7098, 1098, 18, 5332, 10150, 12, 7540, 278, 1527, 722, 13, 471, 486, 702, 18, 6185, 10150, 30, 393, 273, 315, 7540,...
audio.cancel() audio.speakMessage(str(self.dom.readyState))
if winUser.isDescendantWindow(self.NVDAObject.windowHandle,api.getFocusObject().windowHandle): audio.cancel() audio.speakMessage(str(self.dom.readyState))
def isDocumentComplete(self): documentComplete=True if self.dom.readyState!="complete": audio.cancel() audio.speakMessage(str(self.dom.readyState)) documentComplete=False for frameNum in range(self.dom.frames.length): try: if self.dom.frames.item(frameNum).document.readyState!="complete": audio.cancel() audio.speakMessage(str(self.dom.frames.item(frameNum).document.readyState)) documentComplete=False except: pass return documentComplete
2ccbd29a498b50e15a00cf571a2d239c140eb639 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9340/2ccbd29a498b50e15a00cf571a2d239c140eb639/MSHTML.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 353, 2519, 6322, 12, 2890, 4672, 1668, 6322, 33, 5510, 309, 365, 18, 9859, 18, 1672, 1119, 5, 1546, 6226, 6877, 309, 5657, 1299, 18, 291, 29014, 3829, 12, 2890, 18, 11679, 9793, 921, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 353, 2519, 6322, 12, 2890, 4672, 1668, 6322, 33, 5510, 309, 365, 18, 9859, 18, 1672, 1119, 5, 1546, 6226, 6877, 309, 5657, 1299, 18, 291, 29014, 3829, 12, 2890, 18, 11679, 9793, 921, 1...
docs.sort(lambda x,y: _cmp_name(x, y))
docs.sort(lambda x,y,s=self: s._cmp_name(x, y))
def _sort(self, docs, sort_order=None): """ Sort and filter a list of C{ObjDoc}s. In particular, if C{sort_order} is not C{None}, then sort according to its contents; otherwise, sort using L{_cmp_name}. If L{_show_private} is true, then filter out all private objects; otherwise, perform no filtering.
62ff4e473663c51d0a945a451e36120f11d9450a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11420/62ff4e473663c51d0a945a451e36120f11d9450a/html.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 3804, 12, 2890, 16, 3270, 16, 1524, 67, 1019, 33, 7036, 4672, 3536, 5928, 471, 1034, 279, 666, 434, 385, 95, 2675, 1759, 97, 87, 18, 225, 657, 6826, 16, 309, 385, 95, 3804, 67, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 3804, 12, 2890, 16, 3270, 16, 1524, 67, 1019, 33, 7036, 4672, 3536, 5928, 471, 1034, 279, 666, 434, 385, 95, 2675, 1759, 97, 87, 18, 225, 657, 6826, 16, 309, 385, 95, 3804, 67, ...
if not [x for x in groups if x['name'] in ('cvsextras', 'cvsadmin')] and person['id'] > MAXSYSTEMUID: return dict(status=False, message='%s is not in a group that' ' is allowed to own a package' % owner)
try: self._acl_can_be_held_by_user('owner', (person, groups)) except AclNotAllowedError, e return dict(status=False, message=str(e))
def add_package(self, package, owner, summary): '''Add a new package to the database. ''' # Check that the tg.identity is allowed to set an owner if not identity.in_any_group('cvsadmin'): return dict(status=False, message='User must be in cvsadmin')
89e61b70031507809123a3070a9c56e059cbbf35 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9953/89e61b70031507809123a3070a9c56e059cbbf35/dispatcher.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 67, 5610, 12, 2890, 16, 2181, 16, 3410, 16, 4916, 4672, 9163, 986, 279, 394, 2181, 358, 326, 2063, 18, 9163, 468, 2073, 716, 326, 18289, 18, 10781, 353, 2935, 358, 444, 392, 3410,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 67, 5610, 12, 2890, 16, 2181, 16, 3410, 16, 4916, 4672, 9163, 986, 279, 394, 2181, 358, 326, 2063, 18, 9163, 468, 2073, 716, 326, 18289, 18, 10781, 353, 2935, 358, 444, 392, 3410,...
sage: HK.coleman_integral(w*x^3,S,P) 2*5^2 + 4*5^3 + 5^6 + 2*5^7 + O(5^8) sage: negP = HK(0,-3) sage: HK.coleman_integral(w, P,negP, algorithm='teichmuller') 2*5^2 + 3*5^3 + 2*5^6 + 4*5^8 + O(5^9) sage: HK.coleman_integral(w, P, negP) 2*5^2 + 3*5^3 + 2*5^6 + 4*5^8 + O(5^9) AUTHORS:
sage: HK.coleman_integral(w*x^3,S,P) 2*5^2 + 4*5^3 + 5^6 + 2*5^7 + O(5^8) sage: negP = HK(0,-3) sage: HK.coleman_integral(w, P,negP, algorithm='teichmuller') 2*5^2 + 3*5^3 + 2*5^6 + 4*5^8 + O(5^9) sage: HK.coleman_integral(w, P, negP) 2*5^2 + 3*5^3 + 2*5^6 + 4*5^8 + O(5^9) AUTHORS:
def coleman_integral(self, w, P, Q, algorithm = 'None'): """ INPUT: - w differential (if one of P,Q is Weierstrass, w must be odd) - P point on self - Q point on self - algorithm (optional) = None (uses Frobenius) or teichmuller (uses Teichmuller points)
5067c9f3282b0d03ba4df2e4d4149da37b9c1331 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9890/5067c9f3282b0d03ba4df2e4d4149da37b9c1331/hyperelliptic_padic_field.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 276, 790, 4728, 67, 14970, 23811, 12, 2890, 16, 341, 16, 453, 16, 2238, 16, 4886, 273, 296, 7036, 11, 4672, 3536, 12943, 30, 300, 341, 15221, 2001, 261, 430, 1245, 434, 453, 16, 53, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 276, 790, 4728, 67, 14970, 23811, 12, 2890, 16, 341, 16, 453, 16, 2238, 16, 4886, 273, 296, 7036, 11, 4672, 3536, 12943, 30, 300, 341, 15221, 2001, 261, 430, 1245, 434, 453, 16, 53, ...
c.Run()
def call_blockdev_snapshot(self, node, cf_bdev): """Request a snapshot of the given block device.
ecfe94918845683627f97b646872fde40247176b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7542/ecfe94918845683627f97b646872fde40247176b/rpc.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 745, 67, 2629, 5206, 67, 11171, 12, 2890, 16, 756, 16, 6080, 67, 70, 5206, 4672, 3536, 691, 279, 4439, 434, 326, 864, 1203, 2346, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 745, 67, 2629, 5206, 67, 11171, 12, 2890, 16, 756, 16, 6080, 67, 70, 5206, 4672, 3536, 691, 279, 4439, 434, 326, 864, 1203, 2346, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100,...
actions = [self.ACTION_ANIMATE_AND_OPERATE, self.ACTION_ANIMATE_ONLY]
actions = (self.ACTION_ANIMATE_AND_OPERATE, self.ACTION_ANIMATE_ONLY) if self.is_locked: actions = (self.ACTION_ANIMATE_ONLY, ) if what_to_do in actions: self.toggle(new_state) if self.animate_id is None: self.animate_id = gobject.timeout_add(10, self.animate) actions = (self.ACTION_ANIMATE_AND_OPERATE, self.ACTION_OPERATE_ONLY)
def action(self, event=None): what_to_do = self.ACTION_ANIMATE_AND_OPERATE new_state = None if event is not None and event.content is not None: what_to_do = event.content if isinstance(event.content, dict): what_to_do = event.content['action'] new_state = event.content['state'] actions = [self.ACTION_ANIMATE_AND_OPERATE, self.ACTION_ANIMATE_ONLY] if what_to_do in actions: if not self.is_locked: self.lock_for(self.cooldown_ms) self.launch_event(self.event_type, self.is_on) self.toggle(new_state) self._animate_timeout_id = gobject.timeout_add(10, self.animate) else: if self.cooldown_ms >= 1000: self.launch_event('infobar_message', 'Please wait %d seconds' %(self.cooldown_ms / 1000)) if what_to_do == self.ACTION_ANIMATE_ONLY: self.toggle(new_state) self.animate()
ccb593c8b7e8f82b3dc9caeb447999b88b75f2e7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5280/ccb593c8b7e8f82b3dc9caeb447999b88b75f2e7/icon.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1301, 12, 2890, 16, 871, 33, 7036, 4672, 4121, 67, 869, 67, 2896, 273, 365, 18, 12249, 67, 1258, 3445, 1777, 67, 4307, 67, 13950, 1777, 394, 67, 2019, 273, 599, 309, 871, 353, 486, 5...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1301, 12, 2890, 16, 871, 33, 7036, 4672, 4121, 67, 869, 67, 2896, 273, 365, 18, 12249, 67, 1258, 3445, 1777, 67, 4307, 67, 13950, 1777, 394, 67, 2019, 273, 599, 309, 871, 353, 486, 5...
if search_hidden:
if search_hidden is not None:
def upload(self, filename=None, imageData=None, title=None, desc=None, tags=None, is_public=None, is_family=None, is_friend=None, safety=None, search_hidden=None): # Sanity check the arguments if filename is None and imageData is None: raise ValueError("Need to pass either filename or imageData") if filename and imageData: raise ValueError("Cannot pass both filename and imageData")
f1eb4fc29e06f5b6b54d4fc3fbd2f857a5d191dc /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/11023/f1eb4fc29e06f5b6b54d4fc3fbd2f857a5d191dc/flickrest.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3617, 12, 2890, 16, 1544, 33, 7036, 16, 26229, 33, 7036, 16, 2077, 33, 7036, 16, 3044, 33, 7036, 16, 2342, 33, 7036, 16, 353, 67, 482, 33, 7036, 16, 353, 67, 9309, 33, 7036, 16, 35...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3617, 12, 2890, 16, 1544, 33, 7036, 16, 26229, 33, 7036, 16, 2077, 33, 7036, 16, 3044, 33, 7036, 16, 2342, 33, 7036, 16, 353, 67, 482, 33, 7036, 16, 353, 67, 9309, 33, 7036, 16, 35...
def get_next_song_index(self):
def get_next_random_song_filename(self):
def get_next_song_index(self): """Choose a song at random. Be sure it is not already in the playlist and it does not match the 'exclude' regexp. """
05a2689316661b4838403bc54160a8a4b5da8f7c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6855/05a2689316661b4838403bc54160a8a4b5da8f7c/mpdrandomd.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 4285, 67, 9188, 67, 816, 75, 67, 3459, 12, 2890, 4672, 3536, 24529, 279, 17180, 622, 2744, 18, 4823, 3071, 518, 353, 486, 1818, 316, 326, 16428, 471, 518, 1552, 486, 845, 326,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 4285, 67, 9188, 67, 816, 75, 67, 3459, 12, 2890, 4672, 3536, 24529, 279, 17180, 622, 2744, 18, 4823, 3071, 518, 353, 486, 1818, 316, 326, 16428, 471, 518, 1552, 486, 845, 326,...
else: other_invitation_ids = [] for att2 in self.browse(cr, uid, other_invitation_ids): att_infos.append(((att2.user_id and att2.user_id.name) or \ (att2.partner_id and att2.partner_id.name) or \ att2.email) + ' - Status: ' + att2.state.title()) body_vals = {'name': res_obj.name, 'start_date': res_obj.date, 'end_date': res_obj.date_deadline or False, 'description': res_obj.description or '-', 'location': res_obj.location or '-', 'attendees': '<br>'.join(att_infos), 'user': res_obj.user_id and res_obj.user_id.name or 'OpenERP User', 'sign': sign, 'company': company } body = html_invitation % body_vals if mail_to and email_from: attach = self.get_ics_file(cr, uid, res_obj, context=context) tools.email_send( email_from, mail_to, sub, body, attach=attach and [('invitation.ics', attach)] or None, subtype='html', reply_to=email_from )
for att2 in self.browse(cr, uid, other_invitation_ids): att_infos.append(((att2.user_id and att2.user_id.name) or \ (att2.partner_id and att2.partner_id.name) or \ att2.email) + ' - Status: ' + att2.state.title()) body_vals = {'name': res_obj.name, 'start_date': res_obj.date, 'end_date': res_obj.date_deadline or False, 'description': res_obj.description or '-', 'location': res_obj.location or '-', 'attendees': '<br>'.join(att_infos), 'user': res_obj.user_id and res_obj.user_id.name or 'OpenERP User', 'sign': sign, 'company': company } body = html_invitation % body_vals if mail_to and email_from: attach = self.get_ics_file(cr, uid, res_obj, context=context) tools.email_send( email_from, mail_to, sub, body, attach=attach and [('invitation.ics', attach)] or None, subtype='html', reply_to=email_from )
def _send_mail(self, cr, uid, ids, mail_to, email_from=tools.config.get('email_from', False), context=None): """ Send mail for event invitation to event attendees. @param cr: the current row, from the database cursor, @param uid: the current user’s ID for security checks, @param ids: List of attendee’s IDs. @param email_from: Email address for user sending the mail @param context: A standard dictionary for contextual values @return: True """ if not context: context = {}
568351b86acf37bfade00da7c8f1ba5065f03fc2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/568351b86acf37bfade00da7c8f1ba5065f03fc2/base_calendar.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 4661, 67, 4408, 12, 2890, 16, 4422, 16, 4555, 16, 3258, 16, 4791, 67, 869, 16, 2699, 67, 2080, 33, 6642, 18, 1425, 18, 588, 2668, 3652, 67, 2080, 2187, 1083, 3631, 819, 33, 7036...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 4661, 67, 4408, 12, 2890, 16, 4422, 16, 4555, 16, 3258, 16, 4791, 67, 869, 16, 2699, 67, 2080, 33, 6642, 18, 1425, 18, 588, 2668, 3652, 67, 2080, 2187, 1083, 3631, 819, 33, 7036...
dayling savings and -08:00 otherwise. Zone is coincides with what
dayling savings and -08:00 otherwise. Zone coincides with what
def gettzd(timeinseconds = None): """Return w3's timezone identification string. Expresed as [+/-]hh:mm. For instance, PDT is -07:00 during dayling savings and -08:00 otherwise. Zone is coincides with what localtime(), etc., use. If no argument given, use the current time. """ if timeinseconds is None: timeinseconds = time.time() dst_in_effect = time.daylight and time.localtime(timeinseconds)[8] if dst_in_effect: offset = -time.altzone/60 else: offset = -time.timezone/60 if offset > 0: prefix = "+" elif offset < 0: prefix = "-" else: return "Z" # time is already in UTC hours, minutes = map(abs, divmod(offset, 60)) assert 0 <= hours <= 23 assert 0 <= minutes <= 59 return "%s%02d:%02d" % (prefix, hours, minutes)
6248edd386076d761a1e48c22f98e74aa807c93e /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/783/6248edd386076d761a1e48c22f98e74aa807c93e/Time.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 12994, 72, 12, 957, 267, 7572, 273, 599, 4672, 3536, 990, 341, 23, 1807, 6371, 24739, 533, 18, 225, 1312, 1484, 730, 487, 306, 31298, 17, 65, 21622, 30, 7020, 18, 225, 2457, 791, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 12994, 72, 12, 957, 267, 7572, 273, 599, 4672, 3536, 990, 341, 23, 1807, 6371, 24739, 533, 18, 225, 1312, 1484, 730, 487, 306, 31298, 17, 65, 21622, 30, 7020, 18, 225, 2457, 791, ...
self.format_url(url), s, msg)
self.format_url(url), s, msg)
def report_errors(self): if not self.bad: self.message("\nNo errors") return self.message("\nError Report:") sources = self.errors.keys() sources.sort() for source in sources: triples = self.errors[source] self.message("") if len(triples) > 1: self.message("%d Errors in %s", len(triples), source) else: self.message("Error in %s", source) # Begin SLB 2/24/99: Modified this loop to # call self.format_url() instead of referring # to the URL directly, since the URLs in these # triples is now a (URL, fragment) pair. The value # of the "source" variable comes from the list of # origins, and is a URL, not a pair. for url, rawlink, msg in triples: if rawlink != self.format_url(url): s = " (%s)" % rawlink else: s = "" self.message(" HREF %s%s\n msg %s", self.format_url(url), s, msg) # End SLB 2/24/99.
5445e352b244db461527b895255328eaa1a7dc6f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/5445e352b244db461527b895255328eaa1a7dc6f/wcnew.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2605, 67, 4324, 12, 2890, 4672, 309, 486, 365, 18, 8759, 30, 365, 18, 2150, 31458, 82, 2279, 1334, 7923, 327, 365, 18, 2150, 31458, 82, 668, 8706, 2773, 13, 5550, 273, 365, 18, 4324, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2605, 67, 4324, 12, 2890, 4672, 309, 486, 365, 18, 8759, 30, 365, 18, 2150, 31458, 82, 2279, 1334, 7923, 327, 365, 18, 2150, 31458, 82, 668, 8706, 2773, 13, 5550, 273, 365, 18, 4324, ...
elif val[1] == "network-route":
elif sub_val[0] == "network-route":
def getXendNetConfig(): # Find out what environment we're in: bridge, nat, or route xconfig = os.getenv("XEND_CONFIG") if not xconfig: xconfig = "/etc/xen/xend-config.sxp" try: configfile = open(xconfig, 'r') except: return "bridge" S = configfile.read() pin = Parser() pin.input(S) pin.input_eof() val = pin.get_val() while val[0] != 'network-script': val = pin.get_val() if val[1] == "network-bridge": netenv = "bridge" elif val[1] == "network-route": netenv = "route" elif val[1] == "network-nat": netenv = "nat" else: raise NetworkError("Failed to get network env from xend config") configfile.close() return netenv
30d1d39adad997219d9266653a1f7265cb591f26 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6195/30d1d39adad997219d9266653a1f7265cb591f26/NetConfig.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6538, 409, 7308, 809, 13332, 468, 4163, 596, 4121, 3330, 732, 4565, 316, 30, 10105, 16, 10535, 16, 578, 1946, 619, 1425, 273, 1140, 18, 588, 3074, 2932, 60, 4415, 67, 7203, 7923, 309, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6538, 409, 7308, 809, 13332, 468, 4163, 596, 4121, 3330, 732, 4565, 316, 30, 10105, 16, 10535, 16, 578, 1946, 619, 1425, 273, 1140, 18, 588, 3074, 2932, 60, 4415, 67, 7203, 7923, 309, ...
['sage/gsl/interpolation.pyx'], libraries = ['gsl', BLAS, BLAS2], define_macros=[('GSL_DISABLE_DEPRECATED','1')])
['sage/gsl/interpolation.pyx'], libraries = ['gsl', BLAS, BLAS2], define_macros=[('GSL_DISABLE_DEPRECATED','1')])
def is_newer(file1, file2): """ Return True if either file2 does not exist or is older than file1. If file1 does not exist, always return False. """ if not os.path.exists(file1): return False if not os.path.exists(file2): return True if os.path.getmtime(file2) < os.path.getmtime(file1): return True return False
12aff705aa1e2b2dc6e401b763fe2e9e51bdf02d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/12aff705aa1e2b2dc6e401b763fe2e9e51bdf02d/setup.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 353, 67, 2704, 264, 12, 768, 21, 16, 585, 22, 4672, 3536, 2000, 1053, 309, 3344, 585, 22, 1552, 486, 1005, 578, 353, 12156, 2353, 585, 21, 18, 225, 971, 585, 21, 1552, 486, 1005, 16,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 353, 67, 2704, 264, 12, 768, 21, 16, 585, 22, 4672, 3536, 2000, 1053, 309, 3344, 585, 22, 1552, 486, 1005, 578, 353, 12156, 2353, 585, 21, 18, 225, 971, 585, 21, 1552, 486, 1005, 16,...
pkglist ])
] + pkglist)
def remove_packages(pkglist): proc = subprocess.call([ 'emerge', '--unmerge', pkglist ]) return proc == os.EX_OK
c51ba60c8826cc86748400edf41a8edb17be64f5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10378/c51ba60c8826cc86748400edf41a8edb17be64f5/tinderbox.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1206, 67, 10308, 12, 10657, 1098, 4672, 5418, 273, 6652, 18, 1991, 3816, 296, 351, 1583, 2187, 3534, 318, 2702, 2187, 308, 397, 3475, 1098, 13, 327, 5418, 422, 1140, 18, 2294, 67, 3141, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1206, 67, 10308, 12, 10657, 1098, 4672, 5418, 273, 6652, 18, 1991, 3816, 296, 351, 1583, 2187, 3534, 318, 2702, 2187, 308, 397, 3475, 1098, 13, 327, 5418, 422, 1140, 18, 2294, 67, 3141, ...
wikipedia.output(newPhotoDescription) if not skip: bot = upload.UploadRobot(photoUrl, description=newPhotoDescription, useFilename=newFilename, keepFilename=True, verifyDescription=False) bot.upload_image(debug=False) return 1 return 0
wikipedia.output(newPhotoDescription) if not skip: bot = upload.UploadRobot(photoUrl, description=newPhotoDescription, useFilename=newFilename, keepFilename=True, verifyDescription=False) bot.upload_image(debug=False) return 1 return 0
def processPhoto(flickr=None, photo_id=u'', flickrreview=False, reviewer=u'', addCategory=u'', removeCategories=False, autonomous=False): ''' Process a single Flickr photo ''' if photo_id: print photo_id (photoInfo, photoSizes) = flickrripper.getPhoto(flickr, photo_id) if flickrripper.isAllowedLicense(photoInfo): #Get the url of the largest photo photoUrl = flickrripper.getPhotoUrl(photoSizes) #Should download the photo only once photo = flickrripper.downloadPhoto(photoUrl) #Don't upload duplicate images, should add override option duplicates = flickrripper.findDuplicateImages(photo) if duplicates: wikipedia.output(u'Found duplicate image at %s' % duplicates.pop()) else: filename = flickrripper.getFilename(photoInfo, project=u'WLM') flinfoDescription = flickrripper.getFlinfoDescription(photo_id) rijksmonumentid = getRijksmonumentid(photoInfo) photoDescription = buildDescription(flinfoDescription, flickrreview, reviewer, addCategory, removeCategories, rijksmonumentid) #wikipedia.output(photoDescription) if not autonomous: (newPhotoDescription, newFilename, skip)=Tkdialog(photoDescription, photo, filename).run() else: newPhotoDescription=photoDescription newFilename=filename skip=False wikipedia.output(newPhotoDescription) if not skip: bot = upload.UploadRobot(photoUrl, description=newPhotoDescription, useFilename=newFilename, keepFilename=True, verifyDescription=False) bot.upload_image(debug=False) return 1 return 0
bc27f27a5b0fbb3bb90a31410f477020f5d574c1 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7426/bc27f27a5b0fbb3bb90a31410f477020f5d574c1/flickrripper_wlm.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1207, 19934, 12, 2242, 16254, 33, 7036, 16, 10701, 67, 350, 33, 89, 11, 2187, 1183, 16254, 16041, 33, 8381, 16, 10725, 264, 33, 89, 11, 2187, 527, 4457, 33, 89, 11, 2187, 1206, 10487, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1207, 19934, 12, 2242, 16254, 33, 7036, 16, 10701, 67, 350, 33, 89, 11, 2187, 1183, 16254, 16041, 33, 8381, 16, 10725, 264, 33, 89, 11, 2187, 527, 4457, 33, 89, 11, 2187, 1206, 10487, ...
"""http://www.mol.fi/paikat/Job.do?jobID=*"""
"""http://www.mol.fi/paikat/Job.do?*"""
def _handle_mol(url): """http://www.mol.fi/paikat/Job.do?jobID=*""" bs = getUrl(url).getBS() if not bs: return title = bs.first("div", {'class':'otsikko'}).string return title
17d9c8810c54ee0df18e3226edf801bb7fd4bd0a /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/14781/17d9c8810c54ee0df18e3226edf801bb7fd4bd0a/module_urltitle.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 4110, 67, 21260, 12, 718, 4672, 3536, 2505, 2207, 5591, 18, 21260, 18, 22056, 19, 8774, 1766, 270, 19, 2278, 18, 2896, 35, 14, 8395, 7081, 273, 8928, 12, 718, 2934, 588, 12692, 14...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 4110, 67, 21260, 12, 718, 4672, 3536, 2505, 2207, 5591, 18, 21260, 18, 22056, 19, 8774, 1766, 270, 19, 2278, 18, 2896, 35, 14, 8395, 7081, 273, 8928, 12, 718, 2934, 588, 12692, 14...
str += I+'<TH CLASS="navbar">&nbsp;&nbsp;&nbsp;Class' +\ '&nbsp;&nbsp;&nbsp;</TH>\n'
str += I+'<th class="navbar">&nbsp;&nbsp;&nbsp;Class' str += '&nbsp;&nbsp;&nbsp;</th>\n'
def _navbar(self, where, uid=None): """ @param where: What page the navbar is being displayed on.. """ str = self._start_of('Navbar') str += '<TABLE CLASS="navbar" BORDER="0" WIDTH="100%"' str += ' CELLPADDING="0" BGCOLOR="#a0c0ff" CELLSPACING="0">\n' str += ' <TR>\n' str += ' <TD WIDTH="100%">\n' str += ' <TABLE BORDER="0" CELLPADDING="0" CELLSPACING="0">\n' str += ' <TR VALIGN="top">\n'
1b52b16120989cc834166571a9a1e9b8676d968f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11420/1b52b16120989cc834166571a9a1e9b8676d968f/html.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 11589, 3215, 12, 2890, 16, 1625, 16, 4555, 33, 7036, 4672, 3536, 632, 891, 1625, 30, 18734, 1363, 326, 8775, 3215, 353, 3832, 10453, 603, 838, 3536, 609, 273, 365, 6315, 1937, 67, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 11589, 3215, 12, 2890, 16, 1625, 16, 4555, 33, 7036, 4672, 3536, 632, 891, 1625, 30, 18734, 1363, 326, 8775, 3215, 353, 3832, 10453, 603, 838, 3536, 609, 273, 365, 6315, 1937, 67, ...
AbelianGroup( 4, [0, 2, 3, 5])
Abelian Group isomorphic to Z x Z/2Z x Z/3Z x Z/5Z
def __init__(self, ambient, gens, names="f"): """ EXAMPLES: sage: F = AbelianGroup(5,[30,64,729],names = list("abcde")) sage: a,b,c,d,e = F.gens() sage: F.subgroup([a^3,b]) AbelianGroup( 2, [0, 0]) sage: F.subgroup([c]) AbelianGroup( 3, [2, 3, 5]) sage: F.subgroup([a,c]) AbelianGroup( 4, [0, 2, 3, 5]) sage: F.subgroup([a,b*c]) AbelianGroup( 2, [0, 0]) sage: F.subgroup([b*c,d]) AbelianGroup( 2, [0, 64]) sage: F.subgroup([a*b,c^6,d],names = list("xyz")) AbelianGroup( 3, [0, 5, 64]) sage: G = F.subgroup([a*b,c^6,d],names = list("xyz")) sage: G AbelianGroup( 3, [0, 5, 64]) sage: print G Subgroup of Abelian group on 5 generators (a, b, c, d, e) with invariants [0, 0, 30, 64, 729] generated by [a*b, c^6, d] with elementary divisors [0, 5, 64] sage: x,y,z = G.gens() sage: x.order() Infinity sage: y.order() 5 sage: z.order() 64 sage: A = AbelianGroup(5,[3, 5, 5, 7, 8], names = "abcde") sage: a,b,c,d,e = A.gens() sage: A.subgroup([a,b]) AbelianGroup( 2, [3, 5]) sage: A.subgroup([a,b,c,d^2,e]) AbelianGroup( 5, [3, 5, 5, 7, 8]) sage: A.subgroup([a,b,c,d^2,e^2]) AbelianGroup( 5, [3, 4, 5, 5, 7]) sage: B = A.subgroup([a^3,b,c,d,e^2]); B AbelianGroup( 4, [4, 5, 5, 7]) sage: B.invariants() [4, 5, 5, 7] sage: A = AbelianGroup(4,[1009, 2003, 3001, 4001], names = "abcd") sage: a,b,c,d = A.gens() sage: B = A.subgroup([a^3,b,c,d]) sage: B.invariants() [1009, 2003, 3001, 4001] sage: A.order() 24266473210027 sage: B.order() 24266473210027 sage: A = AbelianGroup(4,[1008, 2003, 3001, 4001], names = "abcd") sage: a,b,c,d = A.gens() sage: B = A.subgroup([a^3,b,c,d]) sage: B AbelianGroup( 6, [3, 7, 16, 2003, 3001, 4001]) sage: print B Subgroup of Abelian group on 4 generators (a, b, c, d) with invariants [1008, 2003, 3001, 4001] generated by [a^3, b, c, d] with elementary divisors [3, 7, 16, 2003, 3001, 4001] Infinite groups can also be handled: sage: G = AbelianGroup([3,4,0], names = "abc") sage: a,b,c = G.gens() sage: F = G.subgroup([a,b^2,c]); F AbelianGroup( 3, [0, 3, 4]) sage: F.invariants() [0, 3, 4] sage: F.gens() [a, b^2, c] sage: F.order() Infinity
474e9cd21e5543c10385340d5e9747ad7d0f1773 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9890/474e9cd21e5543c10385340d5e9747ad7d0f1773/abelian_group.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 13232, 1979, 16, 314, 773, 16, 1257, 1546, 74, 6, 4672, 3536, 225, 5675, 8900, 11386, 30, 272, 410, 30, 478, 273, 9771, 292, 2779, 1114, 12, 25, 16, 63...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 13232, 1979, 16, 314, 773, 16, 1257, 1546, 74, 6, 4672, 3536, 225, 5675, 8900, 11386, 30, 272, 410, 30, 478, 273, 9771, 292, 2779, 1114, 12, 25, 16, 63...
utils.system("start tcsd", ignore_status=True) raise error.TestError("expecting %s = %s, receiving %s = %s" %
raise error.TestError("expecting %s = %s, observing %s = %s" %
def expect(d, key, value): if (d[key] != value): utils.system("start tcsd", ignore_status=True) raise error.TestError("expecting %s = %s, receiving %s = %s" % (key, value, key, d[key]))
7bfabc008cf3f580a5d1adfd6ef7bb7142a706d0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7177/7bfabc008cf3f580a5d1adfd6ef7bb7142a706d0/hardware_TPMCheck.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4489, 12, 72, 16, 498, 16, 460, 4672, 309, 261, 72, 63, 856, 65, 480, 460, 4672, 1002, 555, 18, 4709, 668, 2932, 12339, 310, 738, 87, 273, 738, 87, 16, 8681, 310, 738, 87, 273, 738...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4489, 12, 72, 16, 498, 16, 460, 4672, 309, 261, 72, 63, 856, 65, 480, 460, 4672, 1002, 555, 18, 4709, 668, 2932, 12339, 310, 738, 87, 273, 738, 87, 16, 8681, 310, 738, 87, 273, 738...
self.sb.error.SetDimensions(rect.x+2, rect.y+2,
self.sb.error.SetDimensions(rect.x+2, rect.y+2,
def interaction(self, frame, info=None): self.frame = frame code = frame.f_code file = code.co_filename base = os.path.basename(file) lineno = frame.f_lineno
872f754eafee36b7b9b20954604570707e5cd36c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4325/872f754eafee36b7b9b20954604570707e5cd36c/Debugger.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13581, 12, 2890, 16, 2623, 16, 1123, 33, 7036, 4672, 365, 18, 3789, 273, 2623, 981, 273, 2623, 18, 74, 67, 710, 585, 273, 981, 18, 2894, 67, 3459, 1026, 273, 1140, 18, 803, 18, 13909...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13581, 12, 2890, 16, 2623, 16, 1123, 33, 7036, 4672, 365, 18, 3789, 273, 2623, 981, 273, 2623, 18, 74, 67, 710, 585, 273, 981, 18, 2894, 67, 3459, 1026, 273, 1140, 18, 803, 18, 13909...
Abelian group on 2 generators (x, y) with invariants [2, 3]
AbelianGroup( 2, [2, 3])
def kernel(self): """ Only works for finite groups.
1ca18bca290f184bea6cb10d93bc2595c7e0e5fd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9417/1ca18bca290f184bea6cb10d93bc2595c7e0e5fd/abelian_group_morphism.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5536, 12, 2890, 4672, 3536, 5098, 6330, 364, 25922, 3252, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5536, 12, 2890, 4672, 3536, 5098, 6330, 364, 25922, 3252, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10...
print 'putting file %s as %s.%s' % (file_path, bucket_name, key.name)
print 'putting file %s as %s.%s' % (file_path, bucket_name, key.name)
def put_file(self, bucket_name, file_path): self.log(method='put_file', bucket_name=bucket_name, file_path=file_path) successful = False num_tries = 0 while not successful and num_tries < self.RetryCount: try: num_tries += 1 bucket = self.get_bucket(bucket_name) if self.preserve_file_name: key_name = os.path.split(file_path)[1] else: key_name = None key = bucket.new_key(key_name) key.set_contents_from_filename(file_path) print 'putting file %s as %s.%s' % (file_path, bucket_name, key.name) successful = True except S3ResponseError, e: print 'caught S3Error' print e time.sleep(self.RetryDelay) return key
48b9f392bcbc4a7ea6c31710206d637d024b727a /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1098/48b9f392bcbc4a7ea6c31710206d637d024b727a/service.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1378, 67, 768, 12, 2890, 16, 2783, 67, 529, 16, 585, 67, 803, 4672, 365, 18, 1330, 12, 2039, 2218, 458, 67, 768, 2187, 2783, 67, 529, 33, 7242, 67, 529, 16, 585, 67, 803, 33, 768, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1378, 67, 768, 12, 2890, 16, 2783, 67, 529, 16, 585, 67, 803, 4672, 365, 18, 1330, 12, 2039, 2218, 458, 67, 768, 2187, 2783, 67, 529, 33, 7242, 67, 529, 16, 585, 67, 803, 33, 768, ...
res.append((imgName, grades[0][0], grades[0][1], grades[0][2], grades[0][3], grades[0][4], grades[0][5], grades[0][6]))
res.append((imgName, grades[0][0], grades[0][2], grades[0][3], grades[0][4], grades[0][1], grades[0][5], grades[0][6]))
def get_grades(res_output_dir = None, idList = None): """ Returns all images grades in the database. Only the latest grade, when available, is returned for each image. Both user-defined and previous release grades are checked for grading information and comment. @return a list of tuples [(image name, grade, comment), ...] """ if res_output_dir and type(res_output_dir) != types.StringType and type(res_output_dir) != types.UnicodeType: raise TypeError, "res_output_dir must be a string" if idList and type(idList) != types.ListType: raise TypeError, "idList must be a list of id [id1, id2,...]" from django.db import connection cur = connection.cursor() q = """ SELECT f.id, f.prevrelgrade, f.prevrelcomment, i.name, i.id, i.checksum FROM youpi_plugin_fitsin AS f, youpi_processing_task AS t, youpi_rel_it AS r, youpi_image AS i WHERE f.task_id=t.id AND r.task_id=f.task_id AND r.image_id = i.id AND t.success=1""" if idList: q += """ AND i.id IN (%s)""" % ','.join([str(r) for r in idList]) if res_output_dir: q += """ AND t.results_output_dir="%s";""" % res_output_dir if q.find(';') == -1: q += ';' cur.execute(q) res = cur.fetchall() tmp = {} for r in res: # Latest grade grades = FirstQEval.objects.filter(fitsin__id = r[0]).order_by('-date') imgName = r[3] # Looks for user-defined grades if grades: if not tmp.has_key(imgName): tmp[imgName] = [] # Keep custom comment in priority, if any if grades[0].custom_comment: tmp[imgName].append((grades[0].grade, grades[0].date, grades[0].custom_comment, grades[0].fitsin.id, r[4], r[5], grades[0].user)) else: # Keep predefined comment tmp[imgName].append((grades[0].grade, grades[0].date, grades[0].comment.comment, grades[0].fitsin.id, r[4], r[5], grades[0].user)) # Now looks for a previous release grade if r[1]: if not tmp.has_key(imgName): tmp[imgName] = [] # Add a fake datetime object to make sure this grade is the older one tmp[imgName].append([r[1], datetime.datetime(datetime.MINYEAR, 1, 1), r[2], r[0], r[4], r[5]]) # Now, only keep the latest grade for each image header = ('Image name', 'Grade', 'Datetime', 'Comment', 'Image ID', 'Fitsin ID', 'Checksum', 'Graded by',) res = [] for imgName, grades in tmp.iteritems(): if len(grades) == 1: res.append((imgName, grades[0][0], grades[0][1], grades[0][2], grades[0][3], grades[0][4], grades[0][5], grades[0][6])) continue latest = grades[0] for k in range(1, len(grades)): if grades[k][1] > latest[1]: latest = grades[k] res.append((imgName, latest[0], latest[1], latest[2], latest[3], latest[4], latest[5], grades[0][6])) # Sort by image name res.sort(cmp = lambda x,y: cmp(x[0], y[0])) res.insert(0, header) return res
06bbf02326e39180c5d0c43ecf29eafadd9d283a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11651/06bbf02326e39180c5d0c43ecf29eafadd9d283a/grading.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 13088, 12, 455, 67, 2844, 67, 1214, 273, 599, 16, 29174, 273, 599, 4672, 3536, 2860, 777, 4602, 24396, 316, 326, 2063, 18, 5098, 326, 4891, 7324, 16, 1347, 2319, 16, 353, 2106...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 13088, 12, 455, 67, 2844, 67, 1214, 273, 599, 16, 29174, 273, 599, 4672, 3536, 2860, 777, 4602, 24396, 316, 326, 2063, 18, 5098, 326, 4891, 7324, 16, 1347, 2319, 16, 353, 2106...
print >> output, "data pointer: %s%s" % (hex(obj.ctypes._as_parameter_), extra)
print >> output, "data pointer: %s%s" % (hex(obj.ctypes._as_parameter_.value), extra)
def info(obj, output=sys.stdout, numpy=0): if numpy: bp = lambda x: x else: bp = lambda x: int(x) cls = getattr(obj, '__class__', type(obj)) if numpy: nm = getattr(cls, '__name__', cls) else: nm = cls print >> output, "class: ", nm print >> output, "shape: ", obj.shape strides = obj.strides print >> output, "strides: ", strides if not numpy: print >> output, "byteoffset: 0" if len(strides) > 0: bs = obj.strides[0] else: bs = obj.itemsize print >> output, "bytestride: ", bs print >> output, "itemsize: ", obj.itemsize print >> output, "aligned: ", bp(obj.flags.aligned) print >> output, "contiguous: ", bp(obj.flags.contiguous) if numpy: print >> output, "fortran: ", obj.flags.fortran if not numpy: print >> output, "buffer: ", repr(obj.data) if not numpy: extra = " (DEBUG ONLY)" tic = "'" else: extra = "" tic = "" print >> output, "data pointer: %s%s" % (hex(obj.ctypes._as_parameter_), extra) print >> output, "byteorder: ", endian = obj.dtype.byteorder if endian in ['|','=']: print >> output, "%s%s%s" % (tic, sys.byteorder, tic) byteswap = False elif endian == '>': print >> output, "%sbig%s" % (tic, tic) byteswap = sys.byteorder != "big" else: print >> output, "%slittle%s" % (tic, tic) byteswap = sys.byteorder != "little" print >> output, "byteswap: ", bp(byteswap) if not numpy: print >> output, "type: ", typefrom(obj).name else: print >> output, "type: %s" % obj.dtype
4c49c1891382a2320ab8c3dcfce0e80487906cff /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/14925/4c49c1891382a2320ab8c3dcfce0e80487906cff/functions.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1123, 12, 2603, 16, 876, 33, 9499, 18, 10283, 16, 3972, 33, 20, 4672, 309, 3972, 30, 9107, 273, 3195, 619, 30, 619, 469, 30, 9107, 273, 3195, 619, 30, 509, 12, 92, 13, 2028, 273, 3...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1123, 12, 2603, 16, 876, 33, 9499, 18, 10283, 16, 3972, 33, 20, 4672, 309, 3972, 30, 9107, 273, 3195, 619, 30, 619, 469, 30, 9107, 273, 3195, 619, 30, 509, 12, 92, 13, 2028, 273, 3...
def test_variableschema(self): dummy = self._dummy dummy.update(id='dummy1') self.assertEqual(dummy.getId(),'dummy1')
def test_variableschema(self): self.folder.dummy = Dummy(oid='dummy') dummy = self.folder.dummy dummy.setTitle('dummy1') self.assertEqual(dummy.Title(),'dummy1')
def test_variableschema(self):
ad7091df760c0a7489c2dac6e9fbaa5eb5f08655 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12165/ad7091df760c0a7489c2dac6e9fbaa5eb5f08655/test_varschema.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 7528, 1243, 12, 2890, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 7528, 1243, 12, 2890, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10...
enable_plugin(env, plugin_name)
_enable_plugin(env, plugin_name)
def _load_py_files(env, search_path, auto_enable=None): for path in search_path: plugin_files = glob(os.path.join(path, '*.py')) for plugin_file in plugin_files: try: plugin_name = os.path.basename(plugin_file[:-3]) env.log.debug('Loading file plugin %s from %s' % \ (plugin_name, plugin_file)) module = imp.load_source(plugin_name, plugin_file) if path == auto_enable: enable_plugin(env, plugin_name) except Exception, e: env.log.error('Failed to load plugin from %s', plugin_file, exc_info=True)
20976725cc7a73b4a447b0ca585c1ea5c4f8fe86 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2831/20976725cc7a73b4a447b0ca585c1ea5c4f8fe86/loader.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 945, 67, 2074, 67, 2354, 12, 3074, 16, 1623, 67, 803, 16, 3656, 67, 7589, 33, 7036, 4672, 364, 589, 316, 1623, 67, 803, 30, 1909, 67, 2354, 273, 4715, 12, 538, 18, 803, 18, 57...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 945, 67, 2074, 67, 2354, 12, 3074, 16, 1623, 67, 803, 16, 3656, 67, 7589, 33, 7036, 4672, 364, 589, 316, 1623, 67, 803, 30, 1909, 67, 2354, 273, 4715, 12, 538, 18, 803, 18, 57...
if not prop._compile_started:
if prop.parent is self and not prop._compile_started:
def _post_configure_properties(self): """Call the ``init()`` method on all ``MapperProperties`` attached to this mapper.
7af9d1c04b431a98c1e19dd234a00fb202283c0a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1074/7af9d1c04b431a98c1e19dd234a00fb202283c0a/mapper.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2767, 67, 14895, 67, 4738, 12, 2890, 4672, 3536, 1477, 326, 12176, 2738, 1435, 10335, 707, 603, 777, 12176, 4597, 2297, 10335, 7495, 358, 333, 5815, 18, 2, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2767, 67, 14895, 67, 4738, 12, 2890, 4672, 3536, 1477, 326, 12176, 2738, 1435, 10335, 707, 603, 777, 12176, 4597, 2297, 10335, 7495, 358, 333, 5815, 18, 2, -100, -100, -100, -100, -...
-p, --port=PORT: port, a number, defualt = 0 or a device name
-p, --port=PORT: port, a number, default = 0 or a device name
def usage(): print >>sys.stderr, """USAGE: %s [options] Simple Terminal Programm for the serial port. options: -p, --port=PORT: port, a number, defualt = 0 or a device name -b, --baud=BAUD: baudrate, default 9600 -r, --rtscts: enable RTS/CTS flow control (default off) -x, --xonxoff: enable software flow control (default off) -e, --echo: enable local echo (default off) -c, --cr: disable CR -> CR+LF translation """ % sys.argv[0]
b786cd543c0d1d1d0891ed88728d89d9950c58cb /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10955/b786cd543c0d1d1d0891ed88728d89d9950c58cb/miniterm.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4084, 13332, 1172, 1671, 9499, 18, 11241, 16, 3536, 29550, 30, 738, 87, 306, 2116, 65, 4477, 18778, 13586, 81, 364, 326, 2734, 1756, 18, 225, 702, 30, 300, 84, 16, 1493, 655, 33, 6354,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4084, 13332, 1172, 1671, 9499, 18, 11241, 16, 3536, 29550, 30, 738, 87, 306, 2116, 65, 4477, 18778, 13586, 81, 364, 326, 2734, 1756, 18, 225, 702, 30, 300, 84, 16, 1493, 655, 33, 6354,...
device_rp = rbdir.append("aux") if device_rp.lstat(): local_edd = 1 else: local_edd = 0
try: device_rp = rbdir.append("aux") if device_rp.lstat(): local_edd = 1 else: local_edd = 0 except (OSError): local_edd = 1
def set_must_escape_dos_devices(self, rbdir): """If local edd or src edd, then must escape """ if getattr(self, "src_fsa", None) is not None: src_edd = self.src_fsa.escape_dos_devices else: src_edd = 0 device_rp = rbdir.append("aux") if device_rp.lstat(): local_edd = 1 else: local_edd = 0 SetConnections.UpdateGlobal('must_escape_dos_devices', \ src_edd or local_edd) log.Log("Restore: must_escape_dos_devices = %d" % \ (src_edd or local_edd), 4)
b8a8f0407b1b2a63b2a9edd2543147cf27293aa2 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/783/b8a8f0407b1b2a63b2a9edd2543147cf27293aa2/fs_abilities.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 444, 67, 11926, 67, 6939, 67, 19219, 67, 12506, 12, 2890, 16, 7138, 1214, 4672, 3536, 2047, 1191, 1675, 72, 578, 1705, 1675, 72, 16, 1508, 1297, 4114, 3536, 309, 3869, 12, 2890, 16, 31...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 444, 67, 11926, 67, 6939, 67, 19219, 67, 12506, 12, 2890, 16, 7138, 1214, 4672, 3536, 2047, 1191, 1675, 72, 578, 1705, 1675, 72, 16, 1508, 1297, 4114, 3536, 309, 3869, 12, 2890, 16, 31...
IEnumStr.__init__(self, ["RGB","RGBA"])
IEnumStr.__init__(self, ["RGB", "RGBA"])
def __init__ (self) : IEnumStr.__init__(self, ["RGB","RGBA"])
e6b2e44036d3a8b58fb73fd3b51f4165a59478b1 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/4914/e6b2e44036d3a8b58fb73fd3b51f4165a59478b1/interface.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 261, 2890, 13, 294, 467, 3572, 1585, 16186, 2738, 972, 12, 2890, 16, 8247, 11343, 15937, 23263, 6, 5717, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 261, 2890, 13, 294, 467, 3572, 1585, 16186, 2738, 972, 12, 2890, 16, 8247, 11343, 15937, 23263, 6, 5717, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
self.SetTitle(self.title_base)
self.SetTitle(APPLICATION_NAME)
def __init__(self): wx.Frame.__init__(self, None, size=config.get_window_size(), style=wx.DEFAULT_FRAME_STYLE) self._set_initial_values_to_member_variables() self._create_gui() self.Maximize(config.get_window_maximized()) self.SetTitle(self.title_base) self.mnu_view_sidebar.Check(config.get_show_sidebar()) self._enable_disable_menus()
f9ba1251fba4c6e1fd79f0257c2701642e518762 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5252/f9ba1251fba4c6e1fd79f0257c2701642e518762/gui.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 4672, 7075, 18, 3219, 16186, 2738, 972, 12, 2890, 16, 599, 16, 963, 33, 1425, 18, 588, 67, 5668, 67, 1467, 9334, 2154, 33, 27226, 18, 5280, 67, 15072, 67, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 4672, 7075, 18, 3219, 16186, 2738, 972, 12, 2890, 16, 599, 16, 963, 33, 1425, 18, 588, 67, 5668, 67, 1467, 9334, 2154, 33, 27226, 18, 5280, 67, 15072, 67, ...
sys.stderr.write("Removing rows with too many NAs...")
sys.stderr.write("Removing rows with NA rate >= %s ..."%(row_cutoff))
def remove_rows_with_too_many_NAs(cls, data_matrix, row_cutoff, debug=0): """ 2008-05-08 become classmethod """ sys.stderr.write("Removing rows with too many NAs...") no_of_rows, no_of_cols = data_matrix.shape rows_with_too_many_NAs_set = Set() strain_index2no_of_NAs = {} for i in range(no_of_rows): no_of_NAs = 0.0 for j in range(no_of_cols): if data_matrix[i][j] == 0: no_of_NAs += 1 NA_ratio = no_of_NAs/no_of_cols strain_index2no_of_NAs[i] = NA_ratio if NA_ratio >= row_cutoff: rows_with_too_many_NAs_set.add(i) if debug: print print 'rows_with_too_many_NAs_set' print rows_with_too_many_NAs_set sys.stderr.write("%s strains removed, done.\n"%len(rows_with_too_many_NAs_set)) return rows_with_too_many_NAs_set, strain_index2no_of_NAs
9c37de4822440ddecb9d7de412265d707bdb9dbc /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9645/9c37de4822440ddecb9d7de412265d707bdb9dbc/FilterStrainSNPMatrix.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1206, 67, 3870, 67, 1918, 67, 16431, 67, 9353, 67, 50, 1463, 12, 6429, 16, 501, 67, 5667, 16, 1027, 67, 16131, 16, 1198, 33, 20, 4672, 3536, 4044, 28, 17, 6260, 17, 6840, 12561, 667,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1206, 67, 3870, 67, 1918, 67, 16431, 67, 9353, 67, 50, 1463, 12, 6429, 16, 501, 67, 5667, 16, 1027, 67, 16131, 16, 1198, 33, 20, 4672, 3536, 4044, 28, 17, 6260, 17, 6840, 12561, 667,...
self.__lock.acquire()
self._available.acquire()
def shutdown(self): self.__lock.acquire() try: for con in self.__cnxs: con.cnx.close() finally: self.__lock.release()
4439746dd49f45fdd6d37cb8534b75a39afcdad0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9317/4439746dd49f45fdd6d37cb8534b75a39afcdad0/db.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5731, 12, 2890, 4672, 365, 6315, 5699, 18, 1077, 1039, 1435, 775, 30, 364, 356, 316, 365, 16186, 10305, 13713, 30, 356, 18, 10305, 92, 18, 4412, 1435, 3095, 30, 365, 16186, 739, 18, 93...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5731, 12, 2890, 4672, 365, 6315, 5699, 18, 1077, 1039, 1435, 775, 30, 364, 356, 316, 365, 16186, 10305, 13713, 30, 356, 18, 10305, 92, 18, 4412, 1435, 3095, 30, 365, 16186, 739, 18, 93...
if prec is infinity:
if prec is infinity and self.degree() > 0:
def __invert__(self): """ Inverse of the power series, which we assume to have nonzero constant term so that the inverse is again a power series.
40f3d9145118e324dcadda3d7dfd8365723ad073 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9890/40f3d9145118e324dcadda3d7dfd8365723ad073/power_series_ring_element.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 267, 1097, 972, 12, 2890, 4672, 3536, 657, 2476, 434, 326, 7212, 4166, 16, 1492, 732, 6750, 358, 1240, 16966, 5381, 2481, 1427, 716, 326, 8322, 353, 3382, 279, 7212, 4166, 18, 2, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 267, 1097, 972, 12, 2890, 4672, 3536, 657, 2476, 434, 326, 7212, 4166, 16, 1492, 732, 6750, 358, 1240, 16966, 5381, 2481, 1427, 716, 326, 8322, 353, 3382, 279, 7212, 4166, 18, 2, ...
pass
self.body.append(self.context.pop())
def depart_image(self, node): pass
1c335118dca38df949bc8cb376ce377f00ca3a57 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5620/1c335118dca38df949bc8cb376ce377f00ca3a57/html4css1.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 26000, 67, 2730, 12, 2890, 16, 756, 4672, 1342, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 26000, 67, 2730, 12, 2890, 16, 756, 4672, 1342, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100,...
resource_id = re.match(urn_pattern, ia_urn).group(0)
resource_id = re.match(urn_pattern, ia_urn).group(1)
def update_loan_status(self): """Update the loan status based off the status in ACS4""" urn_pattern = r'acs:\w+:(.*)' for ia_urn in self.get_lending_resources(): resource_id = re.match(urn_pattern, ia_urn).group(0) borrow.update_loan_status(resource_id)
0d86ced97a5cdeb05624044c7dad2cd74cf09ccf /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3913/0d86ced97a5cdeb05624044c7dad2cd74cf09ccf/models.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 67, 383, 304, 67, 2327, 12, 2890, 4672, 3536, 1891, 326, 28183, 1267, 2511, 3397, 326, 1267, 316, 432, 4596, 24, 8395, 18412, 67, 4951, 273, 436, 11, 28965, 5581, 91, 15, 30, 744...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 67, 383, 304, 67, 2327, 12, 2890, 4672, 3536, 1891, 326, 28183, 1267, 2511, 3397, 326, 1267, 316, 432, 4596, 24, 8395, 18412, 67, 4951, 273, 436, 11, 28965, 5581, 91, 15, 30, 744...
try:
if pcl_data.nokits:
def __init__(self, appdir):
a557af5bd5f16aa8e745c49124b6913355238d30 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/4494/a557af5bd5f16aa8e745c49124b6913355238d30/devide.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 595, 1214, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 595, 1214, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100,...
f78 = open("work/PApars.pre-flags", "w")
def PApars(rngbrf, nlap, anomaly_stream): g.log = open('log/PApars.GHCN.CL.1000.20.log','w') g.rngbrf = rngbrf g.nlap = nlap g.nstat = g.sumch = g.nyrs = g.nstap = g.nyrsp = g.sumchp = 0 g.nsl1 = g.nsl2 = g.ndsl = g.nswitch = g.nsw0 = g.nok = g.nokx = g.nsta = 0 g.nshort = 0 ts = [0.0] * 900 f = [0.0] * 900 x = [0.0] * 900 w = [0.0] * 900 yr = [0.0] * 900 fpar = [0.0] * 20 rmsp = [0.0] * 20 g.rngbrh = g.rngbrf / 2 g.rbyrcf = earth.radius / g.rngbrf g.rbyrch = earth.radius / g.rngbrh g.cscrif = math.cos(g.rngbrf / earth.radius) g.cscrih = math.cos(g.rngbrh / earth.radius) g.isrData = [] g.isuData = [] # Open output files for some extra logging # # Combination info f66 = open("log/PApars.statn.log.GHCN.CL.1000.20", "w") # Station usage stats f77 = open("log/PApars.statn.use.GHCN.CL.1000.20", "w") # Isolated urban stations f79 = open("log/PApars.noadj.stations.list", "w") # for debugging f78 = open("work/PApars.pre-flags", "w") last_year = int(open('work/GHCN.last_year', 'r').read().strip()) iyoff = 1879 iyrm = last_year - iyoff nyrsin = iyrm g.mbad = 9999 g.xbad = float(9999) isu = 0 isr = 0 i1snow = 1 g.isrData = [] g.isuData = [] idata = [] for (dict, anomalies) in anomaly_stream: idata.extend(anomalies) mfcur = dict['first'] - iyoff length = len(anomalies) # throw away some precision in lat/lon for compatibility with Fortran lat = 0.1 * math.floor(dict['lat']* 10 + 0.5) lon = 0.1 * math.floor(dict['lon']* 10 + 0.5) # note isr = rural stations, isu = non-rural stations if ((dict['US-brightness'] == ' ' and dict['pop'] == 'R') or dict['US-brightness'] == '1'): d = Struct() g.isrData.append(d) d.mfsr = mfcur d.i1sr = i1snow d.ilsr = i1snow + length - 1 d.cslatr = math.cos(lat * g.pi180) d.snlatr = math.sin(lat * g.pi180) d.cslonr = math.cos(lon * g.pi180) d.snlonr = math.sin(lon * g.pi180) d.idr = int(dict['id'][3:]) else: d = Struct() g.isuData.append(d) d.mfsu = mfcur d.i1su = i1snow d.ilsu = i1snow + length - 1 d.cslatu = math.cos(lat * g.pi180) d.snlatu = math.sin(lat * g.pi180) d.cslonu = math.cos(lon * g.pi180) d.snlonu = math.sin(lon * g.pi180) d.idu = int(dict['id'][3:]) d.cc = dict['id'][:3] i1snow = i1snow + length nstau = len(g.isuData) # total number of bright / urban or dim / sm.town stations nstar = len(g.isrData) # total number of dark / rural stations ldtot = i1snow - 1 # total length of idata used g.log.write(" number of rural/urban stations %11d %11d\n" % (nstar, nstau)) # Convert data to real numbers (ann avgs were multiplied by 10) g.rdata = [] for v in idata: if v == g.mbad: g.rdata.append(g.xbad) else: g.rdata.append(0.1 * v) # Sort the rural stations according to the length of the time record # (ignoring gaps). lengths = [] for i, station in enumerate(g.isrData): a, b = station.i1sr - 1, station.ilsr station.recLen = len([v for v in idata[a:b] if v != g.mbad]) station.index = i for i, station in enumerate(g.isrData): g.log.write(" rural station: %11d id: %11d #ok %11d\n" % ( i + 1, station.idr, station.recLen)) g.isrData.sort(key=lambda s:s.recLen) g.isrData.reverse() for i, station in enumerate(g.isrData): g.log.write(" rural station: %11d id: %11d #ok %11d\n" % ( i + 1, station.idr, station.recLen)) # Combine time series for rural stations around each urban station g.nuseid = [0] * nstar for us in g.isuData: usingFullRadius = False dropStation = False needNewNeighbours = True while True: if needNewNeighbours: data = getNeighbours(us, iyoff, full=usingFullRadius) is0, combined, rngbr, iyu1, iyu2, rbyrc = data if is0 == 0: if usingFullRadius: dropStation = True g.log.write(' no rural neighbors for %9d\n' % us.idu) f79.write(" no rural neighbors for %9d\n" % (us.idu)) break usingFullRadius = True needNewNeighbours = True continue wt, iwt, urb, avg = func2(us, iyrm, is0, iyoff, rngbr, combined) iy1 = 1 needNewNeighbours = False if iy1 == 1: f66.write("year dTs-urban dTs-rural StnID=%9d\n" % ( us.idu)) tmean, n3l, nxy, n3, n3f, nxy3, tm3 = func3( iy1, iyrm, avg, urb, iwt, ts, f, iyoff, yr, x, w, f66) if n3 < g.ncrit: if usingFullRadius: f79.write("%3s%09d good years: %4d total years: %4d" " too little rural-neighbors-overlap" " - drop station 9999\n" % ( us.cc, us.idu, n3, n3l - n3f + 1)) dropStation = True break usingFullRadius = True needNewNeighbours = True continue if float(n3) >= g.xcrit * (n3l - n3f + 1. - .1): break # not enough good years for the given range (<66%) # the - 0.1 is to prevent equality with potential uncertainty # due to hardware rounding or compiler behaviour. # nick barnes, ravenbrook limited, 2008 - 09 - 10 # try to save cases in which the gaps are in the early part: iy1 = int(n3l - (n3 - 1) / g.xcrit) if iy1 < n3f + 1: iy1 = n3f + 1 # avoid infinite loop f79.write("%3s%09d drop early years %4d-%4d\n" % ( us.cc, us.idu, 1 + iyoff, iy1 - 1 + iyoff)) if dropStation: continue #=== c subtract urban station and call a curve fitting program tmean = tm3 / nxy3 nxy = nxy3 getfit(nxy, x, f, fpar, rmsp) # find extended range iyxtnd = int(round(n3 / g.xcrit)) - (n3l - n3f + 1) g.log.write(" possible range increase %11d %11d %11d\n" % ( iyxtnd, n3, n3l - n3f + 1)) n1x = n3f + iyoff n2x = n3l + iyoff if iyxtnd < 0: sys.exit('impossible') if iyxtnd > 0: lxend = iyu2 - (n3l + iyoff) if iyxtnd <= lxend: n2x = n2x + lxend else: n1x = n1x - (iyxtnd - lxend) if n1x < iyu1: n1x = iyu1 n2x = iyu2 flag = flags(fpar[0], fpar[1], int(fpar[2] + g.x0), fpar[4], n3f + iyoff, n3l + iyoff) yield (us.cc, us.idu, fpar[0], fpar[1], int(fpar[2] + g.x0), fpar[3], fpar[4], fpar[5], rmsp[0], rmsp[1], n3f + iyoff, n3l + iyoff, n1x, n2x, flag) nuse = 0 tempRs = sorted(g.isrData, key=lambda s: s.index) for (used, tempRs) in itertools.izip(g.nuseid, tempRs): if used > 0: f77.write(" used station %11d %11d times\n" % ( tempRs.idr, used)) nuse += 1 f77.write("%12d rural stations were used\n" % (nuse)) log_stats()
04764b9605b1025116b2db5b3602a74bb6031fc2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6890/04764b9605b1025116b2db5b3602a74bb6031fc2/step2.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 15662, 8954, 12, 86, 3368, 2848, 74, 16, 290, 3762, 16, 392, 26988, 67, 3256, 4672, 314, 18, 1330, 273, 1696, 2668, 1330, 19, 4066, 8954, 18, 16715, 12821, 18, 5017, 18, 18088, 18, 346...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 15662, 8954, 12, 86, 3368, 2848, 74, 16, 290, 3762, 16, 392, 26988, 67, 3256, 4672, 314, 18, 1330, 273, 1696, 2668, 1330, 19, 4066, 8954, 18, 16715, 12821, 18, 5017, 18, 18088, 18, 346...
cols = intersect(self._inherit_fields.keys(), fields)
cols = intersect(self._inherit_fields.keys(), fields_to_read)
def convert_field(f): if f in ('create_date', 'write_date'): return "date_trunc('second', %s) as %s" % (f, f) if isinstance(self._columns[f], fields.binary) and context.get('get_binary_size', True): return "length(%s) as %s" % (f,f) return '"%s"' % (f,)
8ae5a06e35207b8c35182160e11718968900a444 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7397/8ae5a06e35207b8c35182160e11718968900a444/orm.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1765, 67, 1518, 12, 74, 4672, 309, 284, 316, 7707, 2640, 67, 712, 2187, 296, 2626, 67, 712, 11, 4672, 327, 315, 712, 67, 23558, 2668, 8538, 2187, 738, 87, 13, 487, 738, 87, 6, 738, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1765, 67, 1518, 12, 74, 4672, 309, 284, 316, 7707, 2640, 67, 712, 2187, 296, 2626, 67, 712, 11, 4672, 327, 315, 712, 67, 23558, 2668, 8538, 2187, 738, 87, 13, 487, 738, 87, 6, 738, ...
try: blob=self.recv() if blob is None: pass elif blob == '': self.handle_close() else: self.buffer = self.buffer + blob except: self.close()
if self._ssl_accepting: s = self.socket.accept_ssl() if s: self._ssl_accepting = 0 else: try: blob = self.recv(4096) if blob is None: pass elif blob == '': self.handle_close() else: self.buffer = self.buffer + blob except SSL.SSLError, what: if str(what) == 'unexpected eof': self.handle_close() return else: raise
def handle_read(self): try: blob=self.recv() if blob is None: pass elif blob == '': self.handle_close() else: self.buffer = self.buffer + blob except: self.close()
1cf0b9cc3867f97e1e442063b268bff37d2bdb10 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8055/1cf0b9cc3867f97e1e442063b268bff37d2bdb10/echod-async.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1640, 67, 896, 12, 2890, 4672, 775, 30, 4795, 33, 2890, 18, 18334, 1435, 309, 4795, 353, 599, 30, 1342, 1327, 4795, 422, 875, 30, 365, 18, 4110, 67, 4412, 1435, 469, 30, 365, 18, 410...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1640, 67, 896, 12, 2890, 4672, 775, 30, 4795, 33, 2890, 18, 18334, 1435, 309, 4795, 353, 599, 30, 1342, 1327, 4795, 422, 875, 30, 365, 18, 4110, 67, 4412, 1435, 469, 30, 365, 18, 410...
self.transaction = True
def __init__(self, store): global handle if handle != None: self.transaction = True self.sh = handle else: self.sh=get_handle(store) self.transaction = False
35490f2fa56ad10a42e0df031fd435fe3fc932f6 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7147/35490f2fa56ad10a42e0df031fd435fe3fc932f6/seobject.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 1707, 4672, 2552, 1640, 225, 309, 1640, 480, 599, 30, 365, 18, 674, 273, 1640, 469, 30, 365, 18, 674, 33, 588, 67, 4110, 12, 2233, 13, 365, 18, 7958, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 1707, 4672, 2552, 1640, 225, 309, 1640, 480, 599, 30, 365, 18, 674, 273, 1640, 469, 30, 365, 18, 674, 33, 588, 67, 4110, 12, 2233, 13, 365, 18, 7958, ...
run_configure(configure, do_script)
def main(): build_all = "-a" in sys.argv if sys.argv[1] == "Release": arch = "x86" debug = False configure = "VC-WIN32" do_script = "ms\\do_masm" makefile = "ms\\nt.mak" elif sys.argv[1] == "Debug": arch = "x86" debug = True configure = "VC-WIN32" do_script = "ms\\do_masm" makefile="ms\\d32.mak" elif sys.argv[1] == "ReleaseItanium": arch = "ia64" debug = False configure = "VC-WIN64I" do_script = "ms\\do_win64i" makefile = "ms\\nt.mak" os.environ["VSEXTCOMP_USECL"] = "MS_ITANIUM" elif sys.argv[1] == "ReleaseAMD64": arch="amd64" debug=False configure = "VC-WIN64A" do_script = "ms\\do_win64a" makefile = "ms\\nt.mak" os.environ["VSEXTCOMP_USECL"] = "MS_OPTERON" make_flags = "" if build_all: make_flags = "-a" # perl should be on the path, but we also look in "\perl" and "c:\\perl" # as "well known" locations perls = find_all_on_path("perl.exe", ["\\perl\\bin", "C:\\perl\\bin"]) perl = find_working_perl(perls) if perl is None: sys.exit(1) print "Found a working perl at '%s'" % (perl,) sys.stdout.flush() # Look for SSL 2 levels up from pcbuild - ie, same place zlib etc all live. ssl_dir = find_best_ssl_dir(("..\\..",)) if ssl_dir is None: sys.exit(1) old_cd = os.getcwd() try: os.chdir(ssl_dir) # If the ssl makefiles do not exist, we invoke Perl to generate them. if not os.path.isfile(makefile): print "Creating the makefiles..." sys.stdout.flush() # Put our working Perl at the front of our path os.environ["PATH"] = os.path.dirname(perl) + \ os.pathsep + \ os.environ["PATH"] if arch=="x86" and debug: # the do_masm script in openssl doesn't generate a debug # build makefile so we generate it here: os.system("perl util\mk1mf.pl debug "+configure+" >"+makefile) run_configure(configure, do_script) # Now run make. print "Executing nmake over the ssl makefiles..." sys.stdout.flush() rc = os.system("nmake /nologo PERL=\"%s\" -f \"%s\"" %(perl, makefile)) if rc: print "Executing "+makefile+" failed" print rc sys.exit(rc) finally: os.chdir(old_cd) # And finally, we can build the _ssl module itself for Python. defs = "SSL_DIR=\"%s\"" % (ssl_dir,) if debug: defs = defs + " " + "DEBUG=1" makeCommand = 'nmake /nologo -f _ssl.mak ' + defs + " " + make_flags print "Executing:", makeCommand sys.stdout.flush() rc = os.system(makeCommand) sys.exit(rc)
8f2484bf2f1bacabd5606bc1dbda5a5197a45bd1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/8f2484bf2f1bacabd5606bc1dbda5a5197a45bd1/build_ssl.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 1361, 67, 454, 273, 3701, 69, 6, 316, 2589, 18, 19485, 309, 2589, 18, 19485, 63, 21, 65, 422, 315, 7391, 6877, 6637, 273, 315, 92, 5292, 6, 1198, 273, 1083, 5068, 273, 3...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 1361, 67, 454, 273, 3701, 69, 6, 316, 2589, 18, 19485, 309, 2589, 18, 19485, 63, 21, 65, 422, 315, 7391, 6877, 6637, 273, 315, 92, 5292, 6, 1198, 273, 1083, 5068, 273, 3...
usage() sys.exit(-1)
usage() sys.exit(-1)
def usage(): print "Usage: missing.py [--nocolor] [--hidetrans] langXX"
6a4c4fb149ce545711f2ea5cbc01572ef63b9c03 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/206/6a4c4fb149ce545711f2ea5cbc01572ef63b9c03/missing.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4084, 13332, 1172, 315, 5357, 30, 3315, 18, 2074, 306, 413, 31470, 355, 280, 65, 306, 413, 76, 350, 16354, 634, 65, 3303, 5619, 6, 225, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4084, 13332, 1172, 315, 5357, 30, 3315, 18, 2074, 306, 413, 31470, 355, 280, 65, 306, 413, 76, 350, 16354, 634, 65, 3303, 5619, 6, 225, 2, -100, -100, -100, -100, -100, -100, -100, -10...
if os.path.isfile(self.dir + '/.password'):
if os.path.isfile(self.dir + '/.password') and config.NEW_SKIN:
def cwd(self, arg=None, menuw=None): """ make a menu item for each file in the directory """ # are we on a ROM_DRIVE and have to mount it first? for media in config.REMOVABLE_MEDIA: if string.find(self.dir, media.mountdir) == 0: util.mount(self.dir) self.media = media
fbe4ad64f5190b15428fb58a8f41e5d9e06338a9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11399/fbe4ad64f5190b15428fb58a8f41e5d9e06338a9/mediamenu.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7239, 12, 2890, 16, 1501, 33, 7036, 16, 3824, 91, 33, 7036, 4672, 3536, 1221, 279, 3824, 761, 364, 1517, 585, 316, 326, 1867, 3536, 225, 468, 854, 732, 603, 279, 534, 1872, 67, 40, 2...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7239, 12, 2890, 16, 1501, 33, 7036, 16, 3824, 91, 33, 7036, 4672, 3536, 1221, 279, 3824, 761, 364, 1517, 585, 316, 326, 1867, 3536, 225, 468, 854, 732, 603, 279, 534, 1872, 67, 40, 2...
time.sleep(5)
def main(): """ Intended for internal use. If you want to run the installer from another script, use install.install(prog_path). Parses command line arguments and calls install() accordingly. """ global SILENT_MODE opts = None args = None try: opts, args = getopt.getopt(sys.argv[1:], "hs") except getopt.GetoptError, err: print str(err) usage() return if in_opts(opts, "-h"): usage() return if in_opts(opts, "-s"): SILENT_MODE = True # The install directory defaults to the current directory. install_dir = "." if len(args) > 1: install_dir = args[1] try: install(install_dir) except UnsupportedPythonError, err: print "Your version of Python (" + err.version + ") is not supported." except AlreadyInstalledError: print "seattle was already installed."
afe52e13cb76fb3d04696f9c6931db6e0621e809 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7263/afe52e13cb76fb3d04696f9c6931db6e0621e809/install.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 3536, 657, 8140, 364, 2713, 999, 18, 971, 1846, 2545, 358, 1086, 326, 16050, 628, 4042, 2728, 16, 999, 3799, 18, 5425, 12, 14654, 67, 803, 2934, 2280, 2420, 1296, 980, 1775,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 3536, 657, 8140, 364, 2713, 999, 18, 971, 1846, 2545, 358, 1086, 326, 16050, 628, 4042, 2728, 16, 999, 3799, 18, 5425, 12, 14654, 67, 803, 2934, 2280, 2420, 1296, 980, 1775,...
print "-- starting process for class: %s" % classId
def reap_processes(wait=False): # reap the current processes (wait==False: if they are finished) #print "-- entering reap_processes with len: %d" % len(processes) reaped = False while True: for pos, pid in enumerate(processes.keys()): if not wait and pid.poll() == None: # None = process hasn't terminated #print pid.poll() continue #print "checking pos: %d" % pos #self._console.progress(pos, length) output, errout = pid.communicate() rcode = pid.returncode cpos = processes[pid][0] if rcode == 0: #tf = processes[pid][1].read() #print output[:30] #print tf[:30] contA[cpos][CONTENT] = output.decode('utf-8') #contA[cpos] = tf else: raise RuntimeError("Problems compiling %s: %s" % (classes[cpos], errout)) print "-- terminating process for class: %s" % classes[cpos] del processes[pid] reaped = True
83dce8af86e08a211989c0be7a9958c50f26647a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5718/83dce8af86e08a211989c0be7a9958c50f26647a/TreeCompiler.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 283, 438, 67, 18675, 12, 7048, 33, 8381, 4672, 468, 283, 438, 326, 783, 8488, 261, 7048, 631, 8381, 30, 309, 2898, 854, 6708, 13, 468, 1188, 5238, 19014, 283, 438, 67, 18675, 598, 562,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 283, 438, 67, 18675, 12, 7048, 33, 8381, 4672, 468, 283, 438, 326, 783, 8488, 261, 7048, 631, 8381, 30, 309, 2898, 854, 6708, 13, 468, 1188, 5238, 19014, 283, 438, 67, 18675, 598, 562,...
first_doc = encode_link(dosc_n_slots) docs_n_slots = docs_slots_n + 3 + frequency
first_doc = encode_link(docs_n_slots) docs_n_slots = docs_n_slots + 3 + frequency
def index_term(self, tree_file, docs_file, word, documents): """ Indexes the given documents for the given words.
4e709a2fc5d2cbb8b456e3dd46a12604ca37f7ba /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12681/4e709a2fc5d2cbb8b456e3dd46a12604ca37f7ba/index.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 770, 67, 6408, 12, 2890, 16, 2151, 67, 768, 16, 3270, 67, 768, 16, 2076, 16, 7429, 4672, 3536, 3340, 281, 326, 864, 7429, 364, 326, 864, 4511, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 770, 67, 6408, 12, 2890, 16, 2151, 67, 768, 16, 3270, 67, 768, 16, 2076, 16, 7429, 4672, 3536, 3340, 281, 326, 864, 7429, 364, 326, 864, 4511, 18, 2, -100, -100, -100, -100, -100, -1...
out.append((CHARSET, data))
outappend((CHARSET, data))
def _optimize_charset(charset, fixup): # internal: optimize character set out = [] charmap = [False]*256 try: for op, av in charset: if op is NEGATE: out.append((op, av)) elif op is LITERAL: charmap[fixup(av)] = True elif op is RANGE: for i in range(fixup(av[0]), fixup(av[1])+1): charmap[i] = True elif op is CATEGORY: # XXX: could append to charmap tail return charset # cannot compress except IndexError: # character set contains unicode characters return _optimize_unicode(charset, fixup) # compress character map i = p = n = 0 runs = [] for c in charmap: if c: if n == 0: p = i n = n + 1 elif n: runs.append((p, n)) n = 0 i = i + 1 if n: runs.append((p, n)) if len(runs) <= 2: # use literal/range for p, n in runs: if n == 1: out.append((LITERAL, p)) else: out.append((RANGE, (p, p+n-1))) if len(out) < len(charset): return out else: # use bitmap data = _mk_bitmap(charmap) out.append((CHARSET, data)) return out return charset
01c9f8c35f583338e3638906ceeef9d2f29a0254 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/01c9f8c35f583338e3638906ceeef9d2f29a0254/sre_compile.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 29155, 67, 9999, 12, 9999, 16, 2917, 416, 4672, 468, 2713, 30, 10979, 3351, 444, 596, 273, 5378, 1149, 1458, 273, 306, 8381, 5772, 5034, 775, 30, 364, 1061, 16, 1712, 316, 4856, 3...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 29155, 67, 9999, 12, 9999, 16, 2917, 416, 4672, 468, 2713, 30, 10979, 3351, 444, 596, 273, 5378, 1149, 1458, 273, 306, 8381, 5772, 5034, 775, 30, 364, 1061, 16, 1712, 316, 4856, 3...
self.megapov_checkbox.setText(QString.null) QToolTip.add(self.megapov_checkbox,self.__tr("Enable MegaPOV"))
def languageChange(self): self.setCaption(self.__tr("Preferences")) self.ok_btn.setText(self.__tr("OK")) self.groupBox7_2.setTitle(self.__tr("Axes")) self.display_origin_axis_checkbox.setText(self.__tr("Display Origin Axis")) QToolTip.add(self.display_origin_axis_checkbox,self.__tr("Show/Hide Origin Axis")) self.display_pov_axis_checkbox.setText(self.__tr("Display Point of View Axis")) QToolTip.add(self.display_pov_axis_checkbox,self.__tr("Show/Hide Point of View Axis")) self.groupBox17_2.setTitle(self.__tr("Compass")) self.display_compass_checkbox.setText(self.__tr("Display Compass")) QToolTip.add(self.display_compass_checkbox,self.__tr("Show/Hide Display Compass")) self.display_compass_labels_checkbox.setText(self.__tr("Display Compass Labels")) QToolTip.add(self.display_compass_labels_checkbox,self.__tr("Show/Hide Display Compass")) self.textLabel1_4.setText(self.__tr("Location :")) self.compass_position_combox.clear() self.compass_position_combox.insertItem(self.__tr("Upper Right")) self.compass_position_combox.insertItem(self.__tr("Upper Left")) self.compass_position_combox.insertItem(self.__tr("Lower Left")) self.compass_position_combox.insertItem(self.__tr("Lower Right")) self.default_projection_btngrp.setTitle(self.__tr("Default Projection")) QToolTip.add(self.default_projection_btngrp,QString.null) self.radioButton12.setText(self.__tr("Perspective")) self.radioButton13.setText(self.__tr("Orthographic")) self.groupBox14.setTitle(self.__tr("Settings for Adjust")) self.watch_min_in_realtime_checkbox.setText(self.__tr("Watch motion in real time")) QToolTip.add(self.watch_min_in_realtime_checkbox,self.__tr("Enable/disable real time graphical updates for <b>Adjust All</b> or <b>Adjust Selection</b>")) self.update_btngrp.setTitle(QString.null) QToolTip.add(self.update_number_spinbox,self.__tr("Specify how often to update the screen during adjustments")) self.update_units_combobox.clear() self.update_units_combobox.insertItem(self.__tr("frames")) self.update_units_combobox.insertItem(self.__tr("seconds")) self.update_units_combobox.insertItem(self.__tr("minutes")) self.update_units_combobox.insertItem(self.__tr("hours")) QToolTip.add(self.update_units_combobox,self.__tr("Specify how often to update the screen during adjustments")) self.update_every_rbtn.setText(self.__tr("Update every")) QToolTip.add(self.update_every_rbtn,self.__tr("Specify how often to update the screen during adjustments")) self.update_asap_rbtn.setText(self.__tr("Update as fast as possible")) QToolTip.add(self.update_asap_rbtn,self.__tr("Update every 2 seconds, or faster if it doesn't slow adjustments by more than 20%")) self.groupBox20.setTitle(self.__tr("Convergence Criteria")) self.endrms_lbl.setText(self.__tr("EndRMS :")) QToolTip.add(self.endrms_lbl,self.__tr("Target RMS force (pN)")) self.endmax_lbl.setText(self.__tr("EndMax :")) QToolTip.add(self.endmax_lbl,self.__tr("Target max force (pN)")) self.endmax_linedit.setText(self.__tr("10.0")) QToolTip.add(self.endmax_linedit,self.__tr("Target max force (pN)")) self.endrms_linedit.setText(self.__tr("1.0")) QToolTip.add(self.endrms_linedit,self.__tr("Target RMS force (pN)")) self.cutovermax_lbl.setText(self.__tr("CutoverMax :")) QToolTip.add(self.cutovermax_lbl,self.__tr("Cutover max force (pN)")) self.cutovermax_linedit.setText(self.__tr("300.0")) QToolTip.add(self.cutovermax_linedit,self.__tr("Cutover max force (pN)")) self.cutoverrms_linedit.setText(self.__tr("50.0")) QToolTip.add(self.cutoverrms_linedit,self.__tr("Cutover RMS force (pN)")) self.cutoverrms_lbl.setText(self.__tr("CutoverRMS :")) QToolTip.add(self.cutoverrms_lbl,self.__tr("Cutover RMS force (pN)")) self.minimize_warning_lbl.setText(QString.null) self.groupBox8.setTitle(self.__tr("View Animation")) self.animate_views_checkbox.setText(self.__tr("Animate between views")) QToolTip.add(self.animate_views_checkbox,self.__tr("Enable/disable animation between current view and a new view")) self.textLabel1_5.setText(self.__tr("Speed :")) self.textLabel2_3.setText(self.__tr("Slow")) self.textLabel3_4.setText(self.__tr("Fast")) QToolTip.add(self.animation_speed_slider,self.__tr("View Animation Speed")) self.high_quality_graphics_checkbox.setText(self.__tr("High quality graphics")) self.prefs_tab.changeTab(self.tab,self.__tr("General")) self.textLabel1_7.setText(self.__tr("Level of Detail :")) QToolTip.add(self.textLabel1_7,self.__tr("Level of detail for atoms (and bonds)")) self.level_of_detail_combox.clear() self.level_of_detail_combox.insertItem(self.__tr("Low")) self.level_of_detail_combox.insertItem(self.__tr("Medium")) self.level_of_detail_combox.insertItem(self.__tr("High")) self.level_of_detail_combox.insertItem(self.__tr("Variable")) self.level_of_detail_combox.setCurrentItem(2) QToolTip.add(self.level_of_detail_combox,self.__tr("Sets graphics quality for atoms (and bonds)")) self.atom_colors_grpbox.setTitle(self.__tr("Colors")) self.change_element_colors_btn.setText(self.__tr("Change Element Colors...")) self.groupBox13.setTitle(QString.null) self.textLabel3_2_3.setText(self.__tr("Atom Highlighting :")) self.hotspot_lbl_2.setText(self.__tr("Bondpoint Highlighting :")) self.hotspot_lbl.setText(self.__tr("Bondpoint Hotspot :")) self.bondpoint_hilite_color_btn.setText(self.__tr("Choose...")) self.atom_hilite_color_btn.setText(self.__tr("Choose...")) self.hotspot_color_btn.setText(self.__tr("Choose...")) self.reset_atom_colors_btn.setText(self.__tr("Restore Default Colors")) self.textLabel1_3_2.setText(self.__tr("Ball and Stick Atom Scale :")) QToolTip.add(self.textLabel1_3_2,self.__tr("Set Atom Scale factor for Ball and Stick display mode")) self.cpk_atom_rad_spinbox.setSuffix(self.__tr("%")) QToolTip.add(self.cpk_atom_rad_spinbox,self.__tr("Set Atom Scale factor for Ball and Stick display mode")) self.textLabel1_3_2_2.setText(self.__tr("CPK Atom Scale :")) QToolTip.add(self.textLabel1_3_2_2,self.__tr("CPK Atom Scale factor for CPK display mode")) QToolTip.add(self.cpk_scale_factor_linedit,self.__tr("Displays the value of the CPK Atom Scale")) QToolTip.add(self.cpk_scale_factor_slider,self.__tr("Changes Atom Scale factor for CPK display mode")) self.reset_cpk_scale_factor_btn.setText(QString.null) QToolTip.add(self.reset_cpk_scale_factor_btn,self.__tr("Restore default value")) self.prefs_tab.changeTab(self.TabPage,self.__tr("Atoms")) self.high_order_bond_display_btngrp.setTitle(self.__tr("High Order Bonds")) self.radioButton11.setText(self.__tr("Multiple Cylinders")) QToolTip.add(self.radioButton11,self.__tr("Display high order bonds using multiple cylinders")) self.radioButton11_2.setText(self.__tr("Vanes")) QToolTip.add(self.radioButton11_2,self.__tr("Display pi systems in high order bonds as Vanes")) self.radioButton11_2_2.setText(self.__tr("Ribbons")) QToolTip.add(self.radioButton11_2_2,self.__tr("Display pi systems in high order bonds as Ribbons")) self.show_bond_labels_checkbox.setText(self.__tr("Show Bond Type Letters")) QToolTip.add(self.show_bond_labels_checkbox,self.__tr("Display Bond Type Label")) self.show_valence_errors_checkbox.setText(self.__tr("Show Valence Errors")) QToolTip.add(self.show_valence_errors_checkbox,self.__tr("Enable/Disable Valence Error Checker")) self.groupBox4.setTitle(self.__tr("Colors")) self.reset_bond_colors_btn.setText(self.__tr("Restore Default Colors")) self.textLabel3_2.setText(self.__tr("Bond Highlighting :")) self.bond_stretch_color_btn.setText(self.__tr("Choose...")) self.bond_hilite_color_btn.setText(self.__tr("Choose...")) self.textLabel3_3.setText(self.__tr("Vane/Ribbon :")) self.ballstick_bondcolor_btn.setText(self.__tr("Choose...")) self.textLabel3.setText(self.__tr("Ball and Stick Cylinder :")) self.textLabel3_2_2.setText(self.__tr("Bond Stretch :")) self.bond_vane_color_btn.setText(self.__tr("Choose...")) self.textLabel1_3.setText(self.__tr("Ball and Stick Bond Scale :")) QToolTip.add(self.textLabel1_3,self.__tr("Set scale (size) factor for the cylinder representing bonds in Ball and Stick display mode")) self.textLabel1.setText(self.__tr("Bond Line Thickness :")) QToolTip.add(self.textLabel1,self.__tr("Bond thickness (in pixels) for Lines Display Mode")) self.cpk_cylinder_rad_spinbox.setSuffix(self.__tr("%")) QToolTip.add(self.cpk_cylinder_rad_spinbox,self.__tr("Set scale (size) factor for the cylinder representing bonds in Ball and Stick display mode")) self.bond_line_thickness_spinbox.setSuffix(self.__tr(" pixel")) QToolTip.add(self.bond_line_thickness_spinbox,self.__tr("Bond thickness (in pixels) for Lines Display Mode")) self.prefs_tab.changeTab(self.TabPage_2,self.__tr("Bonds")) self.startup_mode_lbl.setText(self.__tr("Startup Mode :")) QToolTip.add(self.startup_mode_lbl,self.__tr("Startup Mode")) self.startup_mode_combox.clear() self.startup_mode_combox.insertItem(self.__tr("Default Mode")) self.startup_mode_combox.insertItem(self.__tr("Build")) QToolTip.add(self.startup_mode_combox,self.__tr("Startup Mode")) self.default_mode_lbl.setText(self.__tr("Default Mode :")) QToolTip.add(self.default_mode_lbl,self.__tr("Default Mode")) self.default_mode_combox.clear() self.default_mode_combox.insertItem(self.__tr("Select Chunks")) self.default_mode_combox.insertItem(self.__tr("Move Chunks")) self.default_mode_combox.insertItem(self.__tr("Build")) QToolTip.add(self.default_mode_combox,self.__tr("Default Mode")) self.mode_groupbox.setTitle(self.__tr("Mode Settings")) self.mode_lbl.setText(self.__tr("Mode :")) self.display_mode_lbl.setText(self.__tr("Display Mode :")) self.mode_combox.clear() self.mode_combox.insertItem(self.__tr("Select Chunks")) self.mode_combox.insertItem(self.__tr("Move Chunks")) self.mode_combox.insertItem(self.__tr("Build")) self.mode_combox.insertItem(self.__tr("Cookie Cutter")) self.mode_combox.insertItem(self.__tr("Extrude")) self.mode_combox.insertItem(self.__tr("Fuse Chunks")) self.mode_combox.insertItem(self.__tr("Movie Player")) QToolTip.add(self.mode_combox,QString.null) self.display_mode_combox.clear() self.display_mode_combox.insertItem(self.__tr("Default")) self.display_mode_combox.insertItem(self.__tr("Invisible")) self.display_mode_combox.insertItem(self.__tr("CPK")) self.display_mode_combox.insertItem(self.__tr("Lines")) self.display_mode_combox.insertItem(self.__tr("Ball and Stick")) self.display_mode_combox.insertItem(self.__tr("Tubes")) QToolTip.add(self.display_mode_combox,self.__tr("Display Mode for this mode")) self.bg_groupbox.setTitle(self.__tr("Background Color")) self.fill_type_lbl.setText(self.__tr("Fill Type :")) self.bg1_color_lbl.setText(self.__tr("Color :")) self.fill_type_combox.clear() self.fill_type_combox.insertItem(self.__tr("Solid")) self.fill_type_combox.insertItem(self.__tr("Blue Sky")) QToolTip.add(self.fill_type_combox,self.__tr("Background fill type")) self.choose_bg1_color_btn.setText(self.__tr("Choose...")) self.restore_bgcolor_btn.setText(self.__tr("Restore Default Color")) self.default_display_btngrp.setTitle(self.__tr("Default Display Mode")) self.vwd_rbtn.setText(self.__tr("CPK")) QToolTip.add(self.vwd_rbtn,self.__tr("CPK (Space Filling) Display Mode")) self.cpk_rbtn.setText(self.__tr("Ball and Stick")) QToolTip.add(self.cpk_rbtn,self.__tr("Ball and Stick Display Mode")) self.lines_rbtn.setText(self.__tr("Lines")) QToolTip.add(self.lines_rbtn,self.__tr("Lines Display Mode")) self.tubes_rbtn.setText(self.__tr("Tubes")) QToolTip.add(self.tubes_rbtn,self.__tr("Tubes Display Mode")) self.buildmode_groupbox.setTitle(self.__tr("Build Mode Defaults")) self.autobond_checkbox.setText(self.__tr("Autobond")) QToolTip.add(self.autobond_checkbox,self.__tr("Build mode's default setting for Autobonding at startup (enabled/disabled)")) self.water_checkbox.setText(self.__tr("Water")) QToolTip.add(self.water_checkbox,self.__tr("Build mode's default setting for Water at startup (enabled/disabled)")) self.buildmode_select_atoms_checkbox.setText(self.__tr("Select Atoms of Deposited Object")) QToolTip.add(self.buildmode_select_atoms_checkbox,self.__tr("Automatically select atoms when depositing")) self.buildmode_highlighting_checkbox.setText(self.__tr("Highlighting")) QToolTip.add(self.buildmode_highlighting_checkbox,self.__tr("Build mode's default setting for Highlighting at startup (enabled/disabled)")) self.prefs_tab.changeTab(self.TabPage_3,self.__tr("Modes")) self.groupBox8_2.setTitle(self.__tr("Directional Light Properties")) self.light_label.setText(self.__tr("Light :")) self.on_label.setText(self.__tr("On :")) self.color_label.setText(self.__tr("Color :")) self.ambient_label.setText(self.__tr("Ambient :")) self.diffuse_label.setText(self.__tr("Diffuse :")) self.specularity_label.setText(self.__tr("Specular :")) self.x_label.setText(self.__tr("X :")) self.y_label.setText(self.__tr("Y :")) self.z_label.setText(self.__tr("Z :")) self.light_combobox.clear() self.light_combobox.insertItem(self.__tr("1 (On)")) self.light_combobox.insertItem(self.__tr("2 (On)")) self.light_combobox.insertItem(self.__tr("3 (Off)")) self.light_checkbox.setText(QString.null) self.light_color_btn.setText(self.__tr("Choose...")) self.lighting_restore_defaults_btn.setText(self.__tr("Restore Defaults")) self.groupBox9_2.setTitle(self.__tr("Material Specular Properties")) self.ms_on_label.setText(self.__tr("On :")) self.ms_finish_label.setText(self.__tr("Finish :")) self.ms_shininess_label.setText(self.__tr("Shininess :")) self.ms_brightness__label.setText(self.__tr("Brightness :")) self.ms_on_checkbox.setText(QString.null) self.textLabel1_6.setText(self.__tr("Metal")) self.textLabel2_4.setText(self.__tr("Plastic")) self.textLabel1_6_2.setText(self.__tr("Flat")) self.textLabel2_4_2.setText(self.__tr("Glossy")) self.textLabel1_6_3.setText(self.__tr("Low")) self.textLabel2_4_3.setText(self.__tr("High")) self.prefs_tab.changeTab(self.TabPage_4,self.__tr("Lighting")) self.file_locations_grp.setTitle(self.__tr("Location of Executables")) self.povray_checkbox.setText(QString.null) QToolTip.add(self.povray_checkbox,self.__tr("Enable POV-Ray")) self.povray_lbl.setText(self.__tr("POV-Ray :")) QToolTip.add(self.povray_lbl,self.__tr("Enable POV-Ray")) self.nanohive_lbl.setText(self.__tr("Nano-Hive :")) QToolTip.add(self.nanohive_lbl,self.__tr("Enable Nano-Hive.")) self.nanohive_checkbox.setText(QString.null) QToolTip.add(self.nanohive_checkbox,self.__tr("Enable Nano-Hive.")) self.povray_path_linedit.setText(QString.null) QToolTip.add(self.povray_path_linedit,self.__tr("The full path to the POV-Ray executable file.")) self.nanohive_path_linedit.setText(QString.null) QToolTip.add(self.nanohive_path_linedit,self.__tr("The full path to the Nano-Hive executable file.")) self.gamess_lbl.setText(self.__tr("GAMESS :")) QToolTip.add(self.gamess_lbl,self.__tr("Enable GAMESS.")) self.megapov_path_linedit.setText(QString.null) QToolTip.add(self.megapov_path_linedit,self.__tr("The full path to the MegaPOV executable file (megapov.exe).")) self.megapov_checkbox.setText(QString.null) QToolTip.add(self.megapov_checkbox,self.__tr("Enable MegaPOV")) self.gamess_path_linedit.setText(QString.null) QToolTip.add(self.gamess_path_linedit,self.__tr("The gamess executable file. Usually it's called gamess.??.x or ??gamess.exe.")) self.gamess_checkbox.setText(QString.null) QToolTip.add(self.gamess_checkbox,self.__tr("Enable GAMESS.")) QToolTip.add(self.povdir_linedit,self.__tr("Select custom POV include directory")) self.nanohive_choose_btn.setText(self.__tr("Choose...")) QToolTip.add(self.nanohive_choose_btn,self.__tr("Choose location of Nano-Hive executable ")) self.povray_choose_btn.setText(self.__tr("Choose...")) QToolTip.add(self.povray_choose_btn,self.__tr("Choose POV-Ray executable")) self.megapov_choose_btn.setText(self.__tr("Choose...")) QToolTip.add(self.megapov_choose_btn,self.__tr("Choose MegaPOV executable (megapov.exe)")) self.povdir_choose_btn.setText(self.__tr("Choose...")) self.gamess_choose_btn.setText(self.__tr("Choose...")) QToolTip.add(self.gamess_choose_btn,self.__tr("Choose GAMESS executable")) self.megapov_lbl.setText(self.__tr("MegaPOV :")) QToolTip.add(self.megapov_lbl,self.__tr("Enable MegaPOV")) self.povdir_checkbox.setText(QString.null) QToolTip.add(self.povdir_checkbox,self.__tr("User-custom directory for POV libraries")) self.povdir_lbl.setText(self.__tr("POV include dir :")) self.prefs_tab.changeTab(self.TabPage_5,self.__tr("Plug-ins")) self.groupBox17.setTitle(self.__tr("History Preferences")) self.msg_serial_number_checkbox.setText(self.__tr("Include message serial number")) self.msg_timestamp_checkbox.setText(self.__tr("Include message timestamp")) self.undo_automatic_checkpoints_checkbox.setText(self.__tr("Automatic Checkpoints")) QToolTip.add(self.undo_automatic_checkpoints_checkbox,self.__tr("Specify Automatic or Manual Checkpoints at program startup.")) self.undo_stack_memory_limit_spinbox.setSuffix(self.__tr(" MB")) self.undo_restore_view_checkbox.setText(self.__tr("Restore View when Undoing Structural Changes")) QToolTip.add(self.undo_restore_view_checkbox,self.__tr("Undo will switch to the view saved with each structural change.")) self.undo_stack_memory_limit_label.setText(self.__tr("Undo Stack Memory Limit :")) self.prefs_tab.changeTab(self.TabPage_6,self.__tr("Undo")) self.groupBox3.setTitle(self.__tr("Window Caption Format")) QToolTip.add(self.groupBox3,self.__tr("Window Border Caption Format")) self.textLabel2.setText(self.__tr("Caption Prefix for Modified File :")) self.textLabel2_2.setText(self.__tr("Caption Suffix for Modified File :")) self.caption_suffix_linedit.setText(QString.null) self.caption_fullpath_checkbox.setText(self.__tr("Display full path of part")) self.groupBox10.setTitle(self.__tr("Window Position and Size")) self.textLabel1_2.setText(self.__tr("Current Size :")) self.textLabel1_2_2.setText(self.__tr("Saved Size :")) self.save_current_btn.setText(self.__tr("Save Current Size")) QToolTip.add(self.save_current_btn,self.__tr("Save current window position and size for next startup")) self.restore_saved_size_btn.setText(self.__tr("Restore Saved Size")) QToolTip.add(self.restore_saved_size_btn,self.__tr("Save current window position and size for next startup")) self.current_width_spinbox.setSuffix(self.__tr(" pixels")) self.current_height_spinbox.setSuffix(self.__tr(" pixels")) self.textLabel1_2_2_2.setText(self.__tr("x")) self.textLabel1_2_2_2_2.setText(self.__tr("x")) self.remember_win_pos_and_size_checkbox.setText(self.__tr("Always save current window position and size when quitting")) self.prefs_tab.changeTab(self.TabPage_7,self.__tr("Window"))
7cc18611a6df93ebc9707a8878e31fb8451f8ff6 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/7cc18611a6df93ebc9707a8878e31fb8451f8ff6/UserPrefsDialog.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2653, 3043, 12, 2890, 4672, 365, 18, 542, 21158, 12, 2890, 16186, 313, 2932, 12377, 6, 3719, 365, 18, 601, 67, 11898, 18, 542, 1528, 12, 2890, 16186, 313, 2932, 3141, 6, 3719, 365, 18,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2653, 3043, 12, 2890, 4672, 365, 18, 542, 21158, 12, 2890, 16186, 313, 2932, 12377, 6, 3719, 365, 18, 601, 67, 11898, 18, 542, 1528, 12, 2890, 16186, 313, 2932, 3141, 6, 3719, 365, 18,...
cur.execute('insert into booze values (%s)',("Cooper's",))
cur.execute( 'insert into %sbooze values (%%s)' % self.table_prefix, ("Cooper's",) )
def _paraminsert(self,cur): self.executeDDL1(cur) cur.execute("insert into booze values ('Victoria Bitter')") self.failUnless(cur.rowcount in (-1,1))
cfcc2c93bb2f3afff81cb651f606e164b84e739d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7955/cfcc2c93bb2f3afff81cb651f606e164b84e739d/dbapi20.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 891, 6387, 12, 2890, 16, 1397, 4672, 365, 18, 8837, 21293, 21, 12, 1397, 13, 662, 18, 8837, 2932, 6387, 1368, 800, 83, 8489, 924, 7707, 58, 933, 280, 1155, 605, 6132, 23291, 13, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 891, 6387, 12, 2890, 16, 1397, 4672, 365, 18, 8837, 21293, 21, 12, 1397, 13, 662, 18, 8837, 2932, 6387, 1368, 800, 83, 8489, 924, 7707, 58, 933, 280, 1155, 605, 6132, 23291, 13, ...
} M_NS = {
} M_NS = {
def _child_get(self, cr, name=False, parent_id=False, domain=None): dirobj = self.context._dirobj uid = self.context.uid ctx = self.context.context.copy() ctx.update(self.dctx) if not domain: domain = [] domain2 = domain + [('calendar_collection','=', False)] res = super(node_database, self)._child_get(cr, name=name, parent_id=parent_id, domain=domain2) where = [('parent_id','=',parent_id)] domain2 = domain + [('calendar_collection','=', True)] if name: where.append(('name','=',name)) if domain2: where += domain2
fe0ee384c7b5a64c12f8725027352df173701456 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/fe0ee384c7b5a64c12f8725027352df173701456/caldav_node.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 3624, 67, 588, 12, 2890, 16, 4422, 16, 508, 33, 8381, 16, 982, 67, 350, 33, 8381, 16, 2461, 33, 7036, 4672, 4314, 303, 441, 273, 365, 18, 2472, 6315, 3211, 303, 441, 4555, 273, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 3624, 67, 588, 12, 2890, 16, 4422, 16, 508, 33, 8381, 16, 982, 67, 350, 33, 8381, 16, 2461, 33, 7036, 4672, 4314, 303, 441, 273, 365, 18, 2472, 6315, 3211, 303, 441, 4555, 273, ...
from zope.app.keyreference.persistent import KeyReferenceToPersistent
from zope.keyreference.persistent import KeyReferenceToPersistent
def create_subscriptions(): # Register adapters and handlers: from zope.component import provideUtility, provideAdapter, provideHandler for adapter in (subscribe.catalog_data, subscribe.SubscriptionSearchableText): provideAdapter(adapter) from zope.component.event import objectEventNotify for handler in (subscribe.subscription_modified, subscribe.subscription_removed, objectEventNotify): provideHandler(handler) # Set up an IIntIds utility: from zope.app.intid import IntIds from zope.app.intid.interfaces import IIntIds intids = IntIds() provideUtility(intids, IIntIds) # We'll register a slight variation of the subscription_added # handler that commits the transaction, so that a later lookup of # IKeyReference for our subscription will work: from ZODB.DemoStorage import DemoStorage from ZODB import DB global root db = DB(DemoStorage()) root = db.open().root() subscription_added.__component_adapts__ = ( subscribe.subscription_added.__component_adapts__) provideHandler(subscription_added) # As a last step, we'll register the IKeyReference adapter for all # persistent objects: from zope.app.keyreference.persistent import KeyReferenceToPersistent from persistent.interfaces import IPersistent provideAdapter(KeyReferenceToPersistent, adapts=(IPersistent,)) return subscribe.Subscriptions()
3fb60a10319ab198bd5c7915f3a0a3e936887e2b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9844/3fb60a10319ab198bd5c7915f3a0a3e936887e2b/tests.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 67, 19964, 13332, 468, 5433, 17272, 471, 4919, 30, 628, 998, 1306, 18, 4652, 1930, 5615, 6497, 16, 5615, 4216, 16, 5615, 1503, 364, 4516, 316, 261, 9174, 18, 7199, 67, 892, 16, 91...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 67, 19964, 13332, 468, 5433, 17272, 471, 4919, 30, 628, 998, 1306, 18, 4652, 1930, 5615, 6497, 16, 5615, 4216, 16, 5615, 1503, 364, 4516, 316, 261, 9174, 18, 7199, 67, 892, 16, 91...
print(" Ziel: "+dstendpoint[0][0]+"/"+str(dstendpoint[0][1])+":"+dstendpoint[1]) dstfilter = filter2where("dst", dstendpoint[0][0], dstendpoint[0][1], dstendpoint[1])
print(" Ziel: "+dstendpoint[0][0]+"/"+dstendpoint[0][1]+":"+dstendpoint[1]) dstfilter = filter2where("dst", dstendpoint[0][0], string.atoi(dstendpoint[0][1]), dstendpoint[1])
def printresult(result, colnames, ipcols): rowstr = "" for col in colnames: rowstr += col + "\t" print(rowstr) for row in result: rowstr = "" i = 0 for col in row: if i in ipcols: rowstr += int2ip(col) + "\t" else: rowstr += str(col) + "\t" i += 1 print rowstr
36da5ebce86e3ad89a7e51790f261c3c224af8d8 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/2286/36da5ebce86e3ad89a7e51790f261c3c224af8d8/inspect-anomaly.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1172, 2088, 12, 2088, 16, 645, 1973, 16, 2359, 6842, 4672, 1027, 701, 273, 1408, 364, 645, 316, 645, 1973, 30, 1027, 701, 1011, 645, 397, 1548, 88, 6, 1172, 12, 492, 701, 13, 364, 10...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1172, 2088, 12, 2088, 16, 645, 1973, 16, 2359, 6842, 4672, 1027, 701, 273, 1408, 364, 645, 316, 645, 1973, 30, 1027, 701, 1011, 645, 397, 1548, 88, 6, 1172, 12, 492, 701, 13, 364, 10...
print "Oh yay it worked!"
irccommand("JOIN " + msg[2])
#def isusingprefix(INDATA,COMMAND):
ef8ca3a4a95f2af723d7168bd94db4f6f12db03d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2452/ef8ca3a4a95f2af723d7168bd94db4f6f12db03d/ircbp.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 468, 536, 353, 9940, 3239, 12, 2356, 3706, 16, 19104, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 468, 536, 353, 9940, 3239, 12, 2356, 3706, 16, 19104, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100,...
the lcov format.
the lcov format. This method needs to be called before the TearDown method.
def Upload(self, coverage_file, upload_path, sym_path=None, src_root=None): """Upload the results to the dashboard.
1ad94952bd681814884ac8d6bfb7e25018c6a904 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5060/1ad94952bd681814884ac8d6bfb7e25018c6a904/coverage.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 9414, 12, 2890, 16, 11196, 67, 768, 16, 3617, 67, 803, 16, 5382, 67, 803, 33, 7036, 16, 1705, 67, 3085, 33, 7036, 4672, 3536, 4777, 326, 1686, 358, 326, 11825, 18, 2, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 9414, 12, 2890, 16, 11196, 67, 768, 16, 3617, 67, 803, 16, 5382, 67, 803, 33, 7036, 16, 1705, 67, 3085, 33, 7036, 4672, 3536, 4777, 326, 1686, 358, 326, 11825, 18, 2, -100, -100, -10...
imageDict[sngl.ifo]=fnmatch.filter(wikiFileFinder.get_RDS_R_L1_SEIS(),\
imageDict[sngl.ifo]=fnmatch.filter(filesOmega,\
def prepareChecklist(wikiFilename=None,wikiCoinc=None,wikiTree=None,file2URL=None): """ Method to prepare a checklist where data products are isolated in directory. """ endOfS5=int(875232014) wikiFileFinder=findFileType(wikiTree,wikiCoinc) # Check to see if wiki file with name already exists maxCount=0 while os.path.exists(wikiFilename) and maxCount < 15: sys.stdout.write("File %s already exists.\n"%\ os.path.split(wikiFilename)[1]) wikiFilename=wikiFilename+".wiki" maxCount=maxCount+1 # #Create the wikipage object etc # wikiPage=wiki(wikiFilename) # # Create top two trigger params tables # cTable=wikiPage.wikiTable(2,9) cTable.data=[ ["Trigger Type", "Rank", "FAR", "SNR", "IFOS(Coinc)", "Instruments(Active)", "Coincidence Time (s)", "Total Mass (mSol)", "Chirp Mass (mSol)" ], ["%s"%(wikiCoinc.type), "%s"%(wikiCoinc.rank), "%s"%(wikiCoinc.far), "%s"%(wikiCoinc.snr), "%s"%(wikiCoinc.ifos), "%s"%(wikiCoinc.instruments), "%s"%(wikiCoinc.time), "%s"%(wikiCoinc.mass), "%s"%(wikiCoinc.mchirp) ] ] pTable=wikiPage.wikiTable(len(wikiCoinc.sngls_in_coinc())+1,7) pTable.data[0]=[ "IFO", "GPS Time(s)", "SNR", "CHISQR", "Mass 1", "Mass 2", "Chirp Mass" ] for row,cSngl in enumerate(wikiCoinc.sngls_in_coinc()): pTable.data[row+1]=[ "%s"%(cSngl.ifo), "%s"%(cSngl.time), "%s"%(cSngl.snr), "%s"%(cSngl.chisqr), "%s"%(cSngl.mass1), "%s"%(cSngl.mass2), "%s"%(cSngl.mchirp) ] #Write the tables into the Wiki object wikiPage.putText("Coincident Trigger Event Information: %s\n"\ %(stfu_pipe.gpsTimeToReadableDate(wikiCoinc.time))) wikiPage.insertTable(cTable) wikiPage.putText("Corresponding Coincident Single IFO Trigger Information\n") wikiPage.insertTable(pTable) #Generate a table of contents to appear after candidate params table wikiPage.tableOfContents(3) #Begin including each checklist item as section with subsections wikiPage.section("Follow-up Checklist") #Put each checklist item wikiPage.subsection("Checklist Summary") wikiPage.subsubsection("Does this candidate pass this checklist?") wikiPage.subsubsection("Answer") wikiPage.subsubsection("Relevant Information and Comments") wikiPage.insertHR() # #First real checklist item wikiPage.subsection("#0 False Alarm Probability") wikiPage.subsubsection("Question") wikiPage.putText("What is the false alarm rate associated with this candidate?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") farTable=wikiPage.wikiTable(2,1) farTable.setTableStyle("background-color: yellow; text-align center;") farTable.data[0][0]="False Alarm Rate" farTable.data[1][0]="%s"%(wikiCoinc.far) wikiPage.insertTable(farTable) wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # #Additional Checklist Item #First real checklist item wikiPage.subsection("#1 Data Quality Flags") wikiPage.subsubsection("Question") wikiPage.putText("Can the data quality flags coincident with this candidate be safely disregarded?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPath=os.path.split(wikiFilename)[0] dqFileList=wikiFileFinder.get_findFlags() if len(dqFileList) != 1: sys.stdout.write("Warning: DQ flags data product import problem.\n") print "Found %i files."%len(dqFileList) for mf in dqFileList: print mf for myFile in dqFileList: wikiPage.putText("%s\n"%(file(myFile).read())) wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # #Additional Checklist Item #First real checklist item wikiPage.subsection("#2 Veto Investigations") wikiPage.subsubsection("Question") wikiPage.putText("Does the candidate survive the veto investigations performed at its time?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") vetoFileList=wikiFileFinder.get_findVetos() if len(vetoFileList) != 1: sys.stdout.write("Warning: Veto flags data product import problem.\n") for myFile in vetoFileList:print myFile for myFile in vetoFileList: wikiPage.putText("%s\n"%(file(myFile).read())) wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # #Additional Checklist Item #First real checklist item wikiPage.subsection("#3 IFO Status") wikiPage.subsubsection("Question") wikiPage.putText("Are the interferometers operating normally with a reasonable level of sensitivity around the time of the candidate?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") #Add link to Daily Stats if wikiCoinc.time <= endOfS5: statsLink=wikiPage.makeExternalLink("http://blue.ligo-wa.caltech.edu/scirun/S5/DailyStatistics/",\ "S5 Daily Stats Page") else: statsLink="This should be a link to S6 Daily Stats!\n" wikiPage.putText(statsLink) #Link figures of merit #Get link for all members of wikiCoinc wikiPage.putText("Figures of Merit\n") if wikiCoinc.time > endOfS5: fomLinks=dict() elems=0 for wikiSngl in wikiCoinc.sngls: if not(wikiSngl.ifo.upper().rstrip().lstrip() == 'V1'): fomLinks[wikiSngl.ifo]=stfu_pipe.getFOMLinks(wikiCoinc.time,wikiSngl.ifo) elems=elems+len(fomLinks[wikiSngl.ifo]) else: for myLabel,myLink,myThumb in stfu_pipe.getFOMLinks(wikiCoinc.time,wikiSngl.ifo): wikiPage.putText("%s\n"%(wikiPage.makeExternalLink(myLink,myLabel))) cols=4 rows=(elems/3)+1 fTable=wikiPage.wikiTable(rows,cols) fTable.data[0]=["IFO,Shift","FOM1","FOM2","FOM3"] currentIndex=0 for myIFOKey in fomLinks.keys(): for label,link,thumb in fomLinks[myIFOKey]: myRow=currentIndex/int(3)+1 myCol=currentIndex%int(3)+1 fTable.data[myRow][0]=label thumbURL=thumb fTable.data[myRow][myCol]="%s"%(wikiPage.linkedRemoteImage(thumb,link)) currentIndex=currentIndex+1 wikiPage.insertTable(fTable) else: wikiPage.putText("Can not automatically fetch S5 FOM links.") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # #Additional Checklist Item #First real checklist item wikiPage.subsection("#4 Candidate Appearance") wikiPage.subsubsection("Question") wikiPage.putText("Do the Qscan figures show what we would expect for a gravitational-wave event?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") imageDict=dict() indexDict=dict() thumbDict=dict() for sngl in wikiCoinc.sngls: frametype,channelName=stfu_pipe.figure_out_type(sngl.time,sngl.ifo,'hoft') indexDict[sngl.ifo]=fnmatch.filter(wikiFileFinder.get_hoft_frame(),\ "*/%s/*/%s/*index.html"%(frametype,sngl.time)) imageDict[sngl.ifo]=fnmatch.filter(wikiFileFinder.get_hoft_frame(),\ "*%s*_%s_16.00_spectrogram_whitened.png"\ %(sngl.time,channelName)) thumbDict[sngl.ifo]=fnmatch.filter(wikiFileFinder.get_hoft_frame(),\ "*%s*_%s_16.00_spectrogram_whitened?thumb.png"\ %(sngl.time,channelName)) # #Convert disk locals to URLs imageDict[sngl.ifo]=[file2URL.convert(x) for x in imageDict[sngl.ifo]] indexDict[sngl.ifo]=[file2URL.convert(x) for x in indexDict[sngl.ifo]] thumbDict[sngl.ifo]=[file2URL.convert(x) for x in thumbDict[sngl.ifo]] if len(indexDict[sngl.ifo]) < 1: wikiPage.putText("GW data channel scans for %s not available.\n"%sngl.ifo) enoughImage=[len(imageDict[key])>0 for key in imageDict.keys()].count(True) >= 1 enoughIndex=[len(indexDict[key])>0 for key in indexDict.keys()].count(True) >= 1 if enoughImage and enoughIndex: wikiPage.insertQscanTable(imageDict,\ thumbDict,\ indexDict) else: sys.stdout.write("Warning: Candidate appearance plot import problem.\n") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#5 Seismic Plots") wikiPage.subsubsection("Question") wikiPage.putText("Is the seismic activity insignificant around the time of the candidate?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPage.putText("Plots and pipeline data go here!") imageDict=dict() indexDict=dict() thumbDict=dict() zValueDict=dict() imageDictAQ=dict() indexDictAQ=dict() thumbDictAQ=dict() zValueDictAQ=dict() # for sngl in wikiCoinc.sngls_in_coinc(): indexDict[sngl.ifo]=fnmatch.filter(wikiFileFinder.get_RDS_R_L1_SEIS(),\ "*/%s_RDS_*/%s/*index.html"%(sngl.ifo,sngl.time)) imageDict[sngl.ifo]=fnmatch.filter(wikiFileFinder.get_RDS_R_L1_SEIS(),\ "*/%s_RDS_*/%s/*SEI*_512.00_spectrogram_whitened.png"%\ (sngl.ifo,sngl.time)) thumbDict[sngl.ifo]=fnmatch.filter(wikiFileFinder.get_RDS_R_L1_SEIS(),\ "*/%s_RDS_*/%s/*SEI*_512.00_spectrogram_whitened?thumb.png"%\ (sngl.ifo,sngl.time)) #Search for corresponding Omega summary.txt file zValueFiles=fnmatch.filter(wikiFileFinder.get_RDS_R_L1_SEIS(),\ "*/%s_RDS_*/%s/*summary.txt"%(sngl.ifo,sngl.time)) zValueDict[sngl.ifo]=list() if (len(zValueFiles) > 0): for zFile in zValueFiles: zValueDict[sngl.ifo].extend(wikiFileFinder.__readSummary__(zFile)) #Reparse only keeping SEI channels tmpList=list() for chan in zValueDict[sngl.ifo]: if "SEI" in chan[0]: tmpList.append(chan) zValueDict[sngl.ifo]=tmpList else: sys.stdout.write("Omega scan summary file not for for %s. ...skipping...\n"%sngl.ifo) #Search for analyzeQscan files #/L1-analyseQscan_L1_932797512_687_seis_rds_L1_SEI-ETMX_X_z_scat-unspecified-gpstime.png timeString=str(float(sngl.time)).replace(".","_") zValueFiles=fnmatch.filter(wikiFileFinder.get_analyzeQscan_SEIS(),\ "*_%s_%s_*.txt"%(sngl.ifo,timeString)) indexDictAQ[sngl.ifo]=fnmatch.filter(wikiFileFinder.get_analyzeQscan_SEIS(),\ "*_%s_%s_*.html"%(sngl.ifo,timeString)) thumbDictAQ[sngl.ifo]=fnmatch.filter(wikiFileFinder.get_analyzeQscan_SEIS(),\ "*%s-*_%s_*_SEI*_z_scat-unspecified-gpstime_thumb.png"\ %(sngl.ifo,timeString)) imageDictAQ[sngl.ifo]=fnmatch.filter(wikiFileFinder.get_analyzeQscan_SEIS(),\ "*%s-*_%s_*_SEI*_z_scat-unspecified-gpstime.png"\ %(sngl.ifo,timeString)) #Process zValue ranking file if found for IFO zValueDictAQ[sngl.ifo]=list() if len(zValueFiles)>0: for zFile in zValueFiles: zValueDictAQ[sngl.ifo].extend(wikiFileFinder.__readZranks__(zFile)) #Reparse keeping SEI channels tmpList=list() for chan in zValueDictAQ[sngl.ifo]: if "SEI" in chan[0]: tmpList.append(chan) zValueDictAQ[sngl.ifo]=tmpList else: sys.stdout.write("Analyze Qscan Z ranking file not found for %s. ...skipping...\n"%sngl.ifo) #Convert disk locals to URLs imageDict[sngl.ifo]=[file2URL.convert(x) for x in imageDict[sngl.ifo]] indexDict[sngl.ifo]=[file2URL.convert(x) for x in indexDict[sngl.ifo]] thumbDict[sngl.ifo]=[file2URL.convert(x) for x in thumbDict[sngl.ifo]] imageDictAQ[sngl.ifo]=[file2URL.convert(x) for x in imageDictAQ[sngl.ifo]] indexDictAQ[sngl.ifo]=[file2URL.convert(x) for x in indexDictAQ[sngl.ifo]] thumbDictAQ[sngl.ifo]=[file2URL.convert(x) for x in thumbDictAQ[sngl.ifo]] if len(indexDict[sngl.ifo]) < 1: wikiPage.putText("Seismic scans for %s not available.\n"%sngl.ifo) enoughImage=[len(imageDict[key])>0 for key in imageDict.keys()].count(True) >=1 enoughIndex=[len(indexDict[key])>0 for key in indexDict.keys()].count(True) >=1 if enoughImage and enoughIndex: wikiPage.insertAnalyzeQscanTable(imageDict, thumbDict, indexDict, zValueDict, imageDictAQ, thumbDictAQ, indexDictAQ, zValueDictAQ) else: sys.stdout.write("Warning: Seismic plots product import problem.\n") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#6 Other environmental causes") wikiPage.subsubsection("Question") wikiPage.putText("Were the environmental disturbances (other than seismic) insignificant at the time of the candidate?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPage.putText("Plots and pipeline data go here!") imageDict=dict() indexDict=dict() thumbDict=dict() zValueDict=dict() imageDictAQ=dict() indexDictAQ=dict() thumbDictAQ=dict() zValueDictAQ=dict() #Select only PEM channels for sngl in wikiCoinc.sngls_in_coinc(): imageDict[sngl.ifo]=list() indexDict[sngl.ifo]=list() thumbDict[sngl.ifo]=list() for myFile in fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ "*/%s_RDS_*/%s/*html"%(sngl.ifo,sngl.time)): indexDict[sngl.ifo].append(myFile) for myFile in fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ "*/%s_RDS_*/%s/*_16.00_spectrogram_whitened.png"%\ (sngl.ifo,sngl.time)): if "PEM" in myFile.upper() and not "SEI" in myFile.upper(): imageDict[sngl.ifo].append(myFile) for myFile in fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ "*/%s_RDS_*/%s/*_16.00_spectrogram_whitened?thumb.png"%\ (sngl.ifo,sngl.time)): if "PEM" in myFile.upper() and not "SEI" in myFile.upper(): thumbDict[sngl.ifo].append(myFile) #Search for corresponding Omega summary.txt file zValueFiles=fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ "*/%s_RDS_*/%s/*summary.txt"%(sngl.ifo,sngl.time)) zValueDict[sngl.ifo]=list() if len(zValueFiles)>0: for zFile in zValueFiles: zValueDict[sngl.ifo].extend(wikiFileFinder.__readSummary__(zFile)) #Reparse only keeping PEM and not SEI channels tmpList=list() for chan in zValueDict[sngl.ifo]: if "PEM" in chan[0] and not "SEI" in chan[0]: tmpList.append(chan) zValueDict[sngl.ifo]=tmpList else: sys.stdout.write("Omega scan summary file not for for %s. ...skipping...\n"%sngl.ifo) #Select associated analyzeQscans imageDictAQ[sngl.ifo]=list() indexDictAQ[sngl.ifo]=list() thumbDictAQ[sngl.ifo]=list() timeString=str(float(sngl.time)).replace(".","_") for myFile in fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*html"%(sngl.ifo,timeString)): indexDictAQ[sngl.ifo].append(myFile) zValueFiles=fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*txt"%(sngl.ifo,timeString)) zValueDictAQ[sngl.ifo]=list() if len(zValueFiles)>0: for zFile in zValueFiles: zValueDictAQ[sngl.ifo].extend(wikiFileFinder.__readZranks__(zFile)) for chan in zValueDictAQ[sngl.ifo]: if "PEM" in chan[0] and not "SEI" in chan[0]: tmpList.append(chan) zValueDictAQ[sngl.ifo]=tmpList else: sys.stdout.write("Analyze Qscan Z ranking file not found for %s. ...skipping...\n"%sngl.ifo) #H1-analyseQscan_H1_931176926_116_rds_H0_PEM-MY_SEISX_z_scat-unspecified-gpstime_thumb.png #H1-analyseQscan_H1_931176926_116_rds_H0_PEM-MY_SEISX_z_scat-unspecified-gpstime.png for myFile in fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*_z_scat-unspecified-gpstime.png"%\ (sngl.ifo,timeString)): if "PEM" in myFile.upper() and not "SEI" in myFile.upper(): imageDictAQ[sngl.ifo].append(myFile) for myFile in fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*_z_scat-unspecified-gpstime?thumb.png"%\ (sngl.ifo,timeString)): if "PEM" in myFile.upper() and not "SEI" in myFile.upper(): thumbDictAQ[sngl.ifo].append(myFile) #Convert disk locals to URLs imageDict[sngl.ifo]=[file2URL.convert(x) for x in imageDict[sngl.ifo]] indexDict[sngl.ifo]=[file2URL.convert(x) for x in indexDict[sngl.ifo]] thumbDict[sngl.ifo]=[file2URL.convert(x) for x in thumbDict[sngl.ifo]] imageDictAQ[sngl.ifo]=[file2URL.convert(x) for x in imageDictAQ[sngl.ifo]] indexDictAQ[sngl.ifo]=[file2URL.convert(x) for x in indexDictAQ[sngl.ifo]] thumbDictAQ[sngl.ifo]=[file2URL.convert(x) for x in thumbDictAQ[sngl.ifo]] if len(imageDict[sngl.ifo]) < 1: wikiPage.putText("PEM scans for %s not available.\n"%sngl.ifo) enoughImage=[len(imageDict[key])>0 for key in imageDict.keys()].count(True) >=1 enoughIndex=[len(indexDict[key])>0 for key in indexDict.keys()].count(True) >=1 if enoughImage and enoughIndex: wikiPage.insertAnalyzeQscanTable(imageDict, thumbDict, indexDict, zValueDict, imageDictAQ, thumbDictAQ, indexDictAQ, zValueDictAQ) else: sys.stdout.write("Warning: PEM plots import trouble.\n") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#7 Auxiliary degree of freedom") wikiPage.subsubsection("Question") wikiPage.putText("Were the auxiliary channel transients coincident with the candidate insignificant?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") imageDict=dict() indexDict=dict() thumbDict=dict() zValueDict=dict() imageDictAQ=dict() indexDictAQ=dict() thumbDictAQ=dict() zValueDictAQ=dict() #Select only AUX channels for sngl in wikiCoinc.sngls: imageDict[sngl.ifo]=list() indexDict[sngl.ifo]=list() thumbDict[sngl.ifo]=list() for myFile in fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ "*/%s_RDS_*/%s/*html"%(sngl.ifo,sngl.time)): indexDict[sngl.ifo].append(myFile) for myFile in fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ "*/%s_RDS_*/%s/*_16.00_spectrogram_whitened.png"%\ (sngl.ifo,sngl.time)): if not "PEM" in myFile.upper() or not "SEI" in myFile.upper(): imageDict[sngl.ifo].append(myFile) for myFile in fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ "*/%s_RDS_*/%s/*_16.00_spectrogram_whitened?thumb.png"%\ (sngl.ifo,sngl.time)): if not "PEM" in myFile.upper() or not "SEI" in myFile.upper(): thumbDict[sngl.ifo].append(myFile) zValueFiles=fnmatch.filter(wikiFileFinder.get_RDS_R_L1(),\ "*/%s_RDS_*/%s/*summary.txt"%(sngl.ifo,sngl.time)) zValueDict[sngl.ifo]=list() if len(zValueFiles)>0: for zFile in zValueFiles: zValueDict[sngl.ifo].extend(wikiFileFinder.__readSummary__(zFile)) #Reparse NOT keeping PEM or SEI channels tmpList=list() for chan in zValueDict[sngl.ifo]: if not "PEM" in chan[0] or not "SEI" in chan[0]: tmpList.append(chan) zValueDict[sngl.ifo]=tmpList else: sys.stdout.write("Omega scan summary file not for for %s. ...skipping...\n"%sngl.ifo) #Select associated analyzeQscans imageDictAQ[sngl.ifo]=list() indexDictAQ[sngl.ifo]=list() thumbDictAQ[sngl.ifo]=list() timeString=str(float(sngl.time)).replace(".","_") #H1-analyseQscan_H1_931176926_116_rds-unspecified-gpstime.html for myFile in fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*html"%(sngl.ifo,timeString)): indexDictAQ[sngl.ifo].append(myFile) zValueFiles=fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*txt"%(sngl.ifo,timeString)) #Process zValue ranking file if found for IFO zValueDictAQ[sngl.ifo]=list() if len(zValueFiles)>0: for zFile in zValueFiles: zValueDictAQ[sngl.ifo].extend(wikiFileFinder.__readZranks__(zFile)) #Reparse NOT keeping PEM or SEI channels tmpList=list() for chan in zValueDictAQ[sngl.ifo]: if not "PEM" in chan[0] or not "SEI" in chan[0]: tmpList.append(chan) zValueDictAQ[sngl.ifo]=tmpList else: sys.stdout.write("Z ranking file not found for %s. ...skipping...\n"%sngl.ifo) #H1-analyseQscan_H1_931176926_116_rds_H0_PEM-MY_SEISX_z_scat-unspecified-gpstime_thumb.png #H1-analyseQscan_H1_931176926_116_rds_H0_PEM-MY_SEISX_z_scat-unspecified-gpstime.png for myFile in fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*_z_scat-unspecified-gpstime.png"%\ (sngl.ifo,timeString)): if not "PEM" in myFile.upper() or not "SEI" in myFile.upper(): imageDictAQ[sngl.ifo].append(myFile) for myFile in fnmatch.filter(wikiFileFinder.get_analyzeQscan_RDS(),\ "*%s-*_%s_*_z_scat-unspecified-gpstime?thumb.png"%\ (sngl.ifo,timeString)): if not "PEM" in myFile.upper() or not "SEI" in myFile.upper(): thumbDictAQ[sngl.ifo].append(myFile) #Convert disk locals to URLs imageDict[sngl.ifo]=[file2URL.convert(x) for x in imageDict[sngl.ifo]] indexDict[sngl.ifo]=[file2URL.convert(x) for x in indexDict[sngl.ifo]] thumbDict[sngl.ifo]=[file2URL.convert(x) for x in thumbDict[sngl.ifo]] imageDictAQ[sngl.ifo]=[file2URL.convert(x) for x in imageDictAQ[sngl.ifo]] indexDictAQ[sngl.ifo]=[file2URL.convert(x) for x in indexDictAQ[sngl.ifo]] thumbDictAQ[sngl.ifo]=[file2URL.convert(x) for x in thumbDictAQ[sngl.ifo]] if len(indexDict[sngl.ifo]) < 1: wikiPage.putText("Other scans for %s not available.\n"%sngl.ifo) enoughImage=[len(imageDict[key])>0 for key in imageDict.keys()].count(True) >=1 enoughIndex=[len(indexDict[key])>0 for key in indexDict.keys()].count(True) >=1 if enoughImage and enoughIndex: wikiPage.insertAnalyzeQscanTable(imageDict, thumbDict, indexDict, zValueDict, imageDictAQ, thumbDictAQ, indexDictAQ, zValueDictAQ) else: sys.stdout.write("Warning: AUX plots import trouble.\n") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#8 Electronic Log Book") wikiPage.subsubsection("Question") wikiPage.putText("Were the instruments behaving normally according to the comments posted by the sci-mons or the operators in the e-log?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPage.putText("Plots and pipeline data go here!") wikiLinkLHOlog=wikiPage.makeExternalLink(stfu_pipe.getiLogURL(myCoinc.time,"H1"), "Hanford eLog") wikiLinkLLOlog=wikiPage.makeExternalLink(stfu_pipe.getiLogURL(myCoinc.time,"L1"), "Livingston eLog") wikiPage.putText("%s\n\n%s\n\n"%(wikiLinkLHOlog,wikiLinkLLOlog)) wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#9 Glitch Report") wikiPage.subsubsection("Question") wikiPage.putText("Were the instruments behaving normally according to the weekly glitch report?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPage.putText("Plots and pipeline data go here!") if int(wikiCoinc.time) >= endOfS5: wikiLinkGlitch=wikiPage.makeExternalLink( "https://www.lsc-group.phys.uwm.edu/twiki/bin/view/DetChar/GlitchStudies", "Glitch Reports for S6" ) else: wikiLinkGlitch=wikiPage.makeExternalLink( "http://www.lsc-group.phys.uwm.edu/glitch/investigations/s5index.html#shift", "Glitch Reports for S5" ) wikiPage.putText("%s\n"%(wikiLinkGlitch)) wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#10 Snr versus time") wikiPage.subsubsection("Question") wikiPage.putText("Is this trigger significant in a SNR versus time plot of all triggers in its analysis chunk?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPage.putText("Plots and pipeline data go here!") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#11 Parameters of the candidate") wikiPage.subsubsection("Question") wikiPage.putText("Does the candidate have a high likelihood of being a gravitational-wave according to its parameters?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPage.putText("Effective Distance Ratio Test\n") effDList=wikiFileFinder.get_effDRatio() if len(effDList) != 1: sys.stdout.write("Warning: Effective Distance Test import problem.\n") for myFile in effDList: wikiPage.putText("%s\n"%(file(myFile).read())) wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#12 Snr and Chisq") wikiPage.subsubsection("Question") wikiPage.putText("Are the SNR and CHISQ time series consistent with our expectations for a gravitational wave?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") # #Put plots SNR and Chi sqr # indexList=fnmatch.filter(wikiFileFinder.get_plotsnrchisq(),"*.html") thumbList=fnmatch.filter(wikiFileFinder.get_plotsnrchisq(),"*_snr-*thumb.png") thumbList.extend(fnmatch.filter(wikiFileFinder.get_plotsnrchisq(),"*_chisq-*thumb.png")) thumbList.sort() indexList=[file2URL.convert(x) for x in indexList] thumbList=[file2URL.convert(x) for x in thumbList] #Two thumb types possible "_thumb.png" or ".thumb.png" imageList=[x.replace("_thumb.png",".png").replace(".thumb.png",".png") for x in thumbList] ifoCount=len(wikiCoinc.sngls) rowLabel={"SNR":1,"CHISQ":2} rowCount=len(rowLabel) colCount=ifoCount if len(indexList) >= 1: snrTable=wikiPage.wikiTable(rowCount+1,colCount+1) for i,sngl in enumerate(wikiCoinc.sngls): myIndex="" for indexFile in indexList: if indexFile.__contains__("_pipe_%s_FOLLOWUP_"%sngl.ifo): myIndex=indexFile if myIndex=="": snrTable.data[0][i+1]=" %s "%sngl.ifo else: snrTable.data[0][i+1]=wikiPage.makeExternalLink(myIndex,sngl.ifo) for col,sngl in enumerate(wikiCoinc.sngls): for row,label in enumerate(rowLabel.keys()): snrTable.data[row+1][0]=label for k,image in enumerate(imageList): if (image.__contains__("_%s-"%label.lower()) \ and image.__contains__("pipe_%s_FOLLOWUP"%sngl.ifo)): snrTable.data[row+1][col+1]=" %s "%(wikiPage.linkedRemoteImage(thumbList[k],thumbList[k])) wikiPage.insertTable(snrTable) else: sys.stdout.write("Warning: SNR and CHISQ plots not found.\n") wikiPage.putText("SNR and CHISQ plots not found.\n") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#13 Template bank veto") wikiPage.subsubsection("Question") wikiPage.putText("Is the bank veto value consistent with our expectations for a gravitational wave?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPage.putText("Plots and pipeline data go here!") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#14 Coherent studies") wikiPage.subsubsection("Question") wikiPage.putText("Are the triggers found in multiple interferometers coherent with each other?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPage.putText("Plots and pipeline data go here!") indexList=fnmatch.filter(wikiFileFinder.get_plotchiatimeseries(),"*.html") if len(indexList) >= 1: myIndex=file2URL.convert(indexList[0]) wikiPage.putText(wikiPage.makeExternalLink(myIndex,\ "%s Coherence Study Results"%(wikiCoinc.ifos))) thumbList=fnmatch.filter(wikiFileFinder.get_plotchiatimeseries(),\ "PLOT_CHIA_%s_snr-squared*thumb.png"%(wikiCoinc.time)) imageList=[x.replace("_thumb.png",".png").replace(".thumb.png",".png") for x in thumbList] rowCount=len(imageList) colCount=1 cohSnrTimeTable=wikiPage.wikiTable(rowCount+1,colCount) cohSnrTimeTable.data[0][0]="%s Coherent SNR Squared Times Series"%(wikiCoinc.ifos) for i,image in enumerate(imageList): cohSnrTimeTable.data[i+1][0]=wikiPage.linkedRemoteImage(image,thumbList[i]) wikiPage.insertTable(cohSnrTimeTable) else: sys.stdout.write("Warning: Coherent plotting jobs not found.\n") wikiPage.putText("Coherent Studies plots not found.\n") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#15 Segmentation Stability") wikiPage.subsubsection("Question") wikiPage.putText("Is the candidate stable against changes in segmentation?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPage.putText("Plots and pipeline data go here!") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # # #Additional Checklist Item wikiPage.subsection("#16 Calibration Stability") wikiPage.subsubsection("Question") wikiPage.putText("Is the candidate stable against changes in calibration that are consistent with systematic uncertainties?") wikiPage.subsubsection("Answer") wikiPage.putText("Edit Here") wikiPage.subsubsection("Relevant Information") wikiPage.putText("Plots and pipeline data go here!") wikiPage.subsubsection("Investigator Comments") wikiPage.putText("Edit Here") wikiPage.insertHR() # #
fb84ac33aba836b8fcb61075cfaec34ced2846c0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5758/fb84ac33aba836b8fcb61075cfaec34ced2846c0/makeCheckListWiki.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2911, 1564, 1098, 12, 13044, 5359, 33, 7036, 16, 13044, 27055, 71, 33, 7036, 16, 13044, 2471, 33, 7036, 16, 768, 22, 1785, 33, 7036, 4672, 3536, 2985, 358, 2911, 279, 866, 1098, 1625, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2911, 1564, 1098, 12, 13044, 5359, 33, 7036, 16, 13044, 27055, 71, 33, 7036, 16, 13044, 2471, 33, 7036, 16, 768, 22, 1785, 33, 7036, 4672, 3536, 2985, 358, 2911, 279, 866, 1098, 1625, ...
sage: sr = mq.SR(1,1,1,4,gf2=True,polybori=True) sage: P = sr.vector([0,0,1,0]) sage: K = sr.vector([1,0,0,1]) sage: F,s = sr.polynomial_system(P, K)
sage: sr = mq.SR(1, 1, 1, 4, gf2=True, polybori=True) sage: P = sr.vector([0, 0, 1, 0]) sage: K = sr.vector([1, 0, 0, 1]) sage: F, s = sr.polynomial_system(P, K)
def polynomial_system(self, P=None, K=None): """ Return a MPolynomialSystem for self for a given plaintext-key pair.
62424369e932ac59629cb4d40b7e47ae2a712293 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/62424369e932ac59629cb4d40b7e47ae2a712293/sr.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 16991, 67, 4299, 12, 2890, 16, 453, 33, 7036, 16, 1475, 33, 7036, 4672, 3536, 2000, 279, 490, 5850, 13602, 3163, 364, 365, 364, 279, 864, 11917, 17, 856, 3082, 18, 2, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 16991, 67, 4299, 12, 2890, 16, 453, 33, 7036, 16, 1475, 33, 7036, 4672, 3536, 2000, 279, 490, 5850, 13602, 3163, 364, 365, 364, 279, 864, 11917, 17, 856, 3082, 18, 2, -100, -100, -100,...
PangoFontset = struct__PangoFontset
PangoFontset = struct__PangoFontset
def load_lib(name): libname = ctypes.util.find_library(name) if not libname: raise OSError("Could not find library '%s'" % name) else: return CDLL(libname)
92bb4862df49194567fad21ea00d921ffc62dd6a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10761/92bb4862df49194567fad21ea00d921ffc62dd6a/pango.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1262, 67, 2941, 12, 529, 4672, 2561, 529, 273, 6983, 18, 1367, 18, 4720, 67, 12083, 12, 529, 13, 309, 486, 2561, 529, 30, 1002, 10002, 2932, 4445, 486, 1104, 5313, 1995, 87, 4970, 738,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1262, 67, 2941, 12, 529, 4672, 2561, 529, 273, 6983, 18, 1367, 18, 4720, 67, 12083, 12, 529, 13, 309, 486, 2561, 529, 30, 1002, 10002, 2932, 4445, 486, 1104, 5313, 1995, 87, 4970, 738,...
cm('test.test_pyclbr', ignore=('defined_in',))
cm('test.test_pyclbr')
def test_others(self): cm = self.checkModule
72a925f8cd88b62bf84ae9701ad9e59a7e7d85d9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/72a925f8cd88b62bf84ae9701ad9e59a7e7d85d9/test_pyclbr.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 28197, 12, 2890, 4672, 5003, 273, 365, 18, 1893, 3120, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 28197, 12, 2890, 4672, 5003, 273, 365, 18, 1893, 3120, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
cslist.append(self._cset_from_rev(anc, r)) anc=r
chtmp = MonotoneChangeset(anc, r) self.logparser.convertLog(chtmp) if self.branch in chtmp.branches: cslist.append(MonotoneChangeset(anc, r)) anc=r if onlyFirst: break
def getCset(self, revlist): # receives a revlist, already toposorted (i.e. ordered by # ancestry) and outputs a list of changesets cslist=[] anc=revlist[0] for r in revlist[1:]: cslist.append(self._cset_from_rev(anc, r)) anc=r return cslist
99161ee430dbc4a866be3eb80981544ead94434b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5981/99161ee430dbc4a866be3eb80981544ead94434b/monotone.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1927, 542, 12, 2890, 16, 5588, 1098, 4672, 468, 17024, 279, 5588, 1098, 16, 1818, 358, 917, 24726, 261, 77, 18, 73, 18, 5901, 635, 468, 392, 27626, 13, 471, 6729, 279, 666, 434, 3478, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1927, 542, 12, 2890, 16, 5588, 1098, 4672, 468, 17024, 279, 5588, 1098, 16, 1818, 358, 917, 24726, 261, 77, 18, 73, 18, 5901, 635, 468, 392, 27626, 13, 471, 6729, 279, 666, 434, 3478, ...
with io.open(support.TESTFN, "rb", buffering=0) as f:
with self.open(support.TESTFN, "rb", buffering=0) as f:
def test_truncate(self): # Truncate implicitly flushes the buffer. with io.open(support.TESTFN, self.write_mode, buffering=0) as raw: bufio = self.tp(raw, 8) bufio.write(b"abcdef") self.assertEqual(bufio.truncate(3), 3) self.assertEqual(bufio.tell(), 3) with io.open(support.TESTFN, "rb", buffering=0) as f: self.assertEqual(f.read(), b"abc")
1eb8e6fd9a20dd970536c19c924ca697ad648652 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3187/1eb8e6fd9a20dd970536c19c924ca697ad648652/test_io.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 27201, 12, 2890, 4672, 468, 25869, 24682, 3663, 281, 326, 1613, 18, 598, 2527, 18, 3190, 12, 13261, 18, 16961, 19793, 16, 365, 18, 2626, 67, 3188, 16, 25056, 33, 20, 13, 487,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 27201, 12, 2890, 4672, 468, 25869, 24682, 3663, 281, 326, 1613, 18, 598, 2527, 18, 3190, 12, 13261, 18, 16961, 19793, 16, 365, 18, 2626, 67, 3188, 16, 25056, 33, 20, 13, 487,...
'SAGE External System Code (e.g., PARI, MAGMA, etc.)',
'Sage External System Code (e.g., PARI, MAGMA, etc.)',
def bundle(self, filename, options='', url=None, base=None, to=None): r""" Create an hg changeset bundle with the given filename against the repository at the given url (which is by default the 'official' SAGE repository, unless push_url() is changed in a setup file).
ae2ebdf61ef0e4abf79e5e38de819254d5ad978b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9417/ae2ebdf61ef0e4abf79e5e38de819254d5ad978b/hg.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3440, 12, 2890, 16, 1544, 16, 702, 2218, 2187, 880, 33, 7036, 16, 1026, 33, 7036, 16, 358, 33, 7036, 4672, 436, 8395, 1788, 392, 22576, 22463, 3440, 598, 326, 864, 1544, 5314, 326, 335...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3440, 12, 2890, 16, 1544, 16, 702, 2218, 2187, 880, 33, 7036, 16, 1026, 33, 7036, 16, 358, 33, 7036, 4672, 436, 8395, 1788, 392, 22576, 22463, 3440, 598, 326, 864, 1544, 5314, 326, 335...
self.xorg_conf.addOption('Screen', 'DefaultDepth', '24', position=0, prefix='')
self.xorg_conf.addOption('Screen', 'Device', 'Configured Video Device', position=0)
def customiseConfig(self): # set DefaultDepth to 24; X.org does not work otherwise self.xorg_conf.makeSection('Screen', identifier='Default Screen') self.xorg_conf.addOption('Screen', 'DefaultDepth', '24', position=0, prefix='')
6b1dc8a89445b666d185f48bc208d573fc88650e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/13004/6b1dc8a89445b666d185f48bc208d573fc88650e/config-xorg-vboxguest.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1679, 784, 809, 12, 2890, 4672, 468, 444, 2989, 6148, 358, 4248, 31, 1139, 18, 3341, 1552, 486, 1440, 3541, 365, 18, 92, 3341, 67, 3923, 18, 6540, 5285, 2668, 7956, 2187, 2756, 2218, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1679, 784, 809, 12, 2890, 4672, 468, 444, 2989, 6148, 358, 4248, 31, 1139, 18, 3341, 1552, 486, 1440, 3541, 365, 18, 92, 3341, 67, 3923, 18, 6540, 5285, 2668, 7956, 2187, 2756, 2218, 1...
print event
self.log.debug(event)
def _process(self, event): for processor in ibid.processors: try: processor.process(event) except Exception: self.log.exception(u"Exception occured in %s processor of %s plugin", processor.__class__.__name__, processor.name)
891bf388f99cc1f33520dd3511934dec3fae630f /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12048/891bf388f99cc1f33520dd3511934dec3fae630f/core.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2567, 12, 2890, 16, 871, 4672, 364, 6659, 316, 9834, 350, 18, 22962, 30, 775, 30, 6659, 18, 2567, 12, 2575, 13, 1335, 1185, 30, 365, 18, 1330, 18, 4064, 12, 89, 6, 503, 16206, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2567, 12, 2890, 16, 871, 4672, 364, 6659, 316, 9834, 350, 18, 22962, 30, 775, 30, 6659, 18, 2567, 12, 2575, 13, 1335, 1185, 30, 365, 18, 1330, 18, 4064, 12, 89, 6, 503, 16206, ...
result += get_source(env, '../rts/System/creg', ignore_builddir=True)
cwdPrev = os.getcwd() os.chdir('..') result += get_source(env, 'rts/System/creg', ignore_builddir=False) os.chdir(cwdPrev)
def get_shared_skirmishAI_source_Creg(env): result = [] if env.has_key('builddir') and env['builddir']: result += get_source(env, '../rts/System/creg', ignore_builddir=True) return result
553bbc8f87003da2c4935f1146c4ae9d8818885c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1030/553bbc8f87003da2c4935f1146c4ae9d8818885c/filelist.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 11574, 67, 7771, 3985, 1468, 18194, 67, 3168, 67, 39, 1574, 12, 3074, 4672, 563, 273, 5378, 309, 1550, 18, 5332, 67, 856, 2668, 70, 89, 330, 449, 481, 6134, 471, 1550, 3292, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 11574, 67, 7771, 3985, 1468, 18194, 67, 3168, 67, 39, 1574, 12, 3074, 4672, 563, 273, 5378, 309, 1550, 18, 5332, 67, 856, 2668, 70, 89, 330, 449, 481, 6134, 471, 1550, 3292, ...
return [self[id] for (id,) in CoincDatabaseConnection.cursor().execute("SELECT DISTINCT time_slide_id FROM time_slide")]
return [self[id] for (id,) in self.connection.cursor().execute("SELECT DISTINCT time_slide_id FROM time_slide")]
def all_offsets(self): return [self[id] for (id,) in CoincDatabaseConnection.cursor().execute("SELECT DISTINCT time_slide_id FROM time_slide")]
d54fb732022835e478a56be4372e71bb33d3cede /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3592/d54fb732022835e478a56be4372e71bb33d3cede/SnglBurstUtils.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 777, 67, 16542, 12, 2890, 4672, 327, 306, 2890, 63, 350, 65, 364, 261, 350, 16, 13, 316, 28932, 71, 4254, 1952, 18, 9216, 7675, 8837, 2932, 4803, 23286, 813, 67, 26371, 67, 350, 4571, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 777, 67, 16542, 12, 2890, 4672, 327, 306, 2890, 63, 350, 65, 364, 261, 350, 16, 13, 316, 28932, 71, 4254, 1952, 18, 9216, 7675, 8837, 2932, 4803, 23286, 813, 67, 26371, 67, 350, 4571, ...
xmlrpc_tests.append(KeepaliveServerTestCase)
xmlrpc_tests.append(KeepaliveServerTestCase1) xmlrpc_tests.append(KeepaliveServerTestCase2)
def test_main(): xmlrpc_tests = [XMLRPCTestCase, HelperTestCase, DateTimeTestCase, BinaryTestCase, FaultTestCase] xmlrpc_tests.append(SimpleServerTestCase) xmlrpc_tests.append(KeepaliveServerTestCase) xmlrpc_tests.append(GzipServerTestCase) xmlrpc_tests.append(ServerProxyTestCase) xmlrpc_tests.append(FailingServerTestCase) xmlrpc_tests.append(CGIHandlerTestCase) support.run_unittest(*xmlrpc_tests)
9a96ed69551dade73e041c5d7528afed89a0c37e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3187/9a96ed69551dade73e041c5d7528afed89a0c37e/test_xmlrpc.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 5254, 13332, 31811, 67, 16341, 273, 306, 4201, 54, 52, 1268, 395, 2449, 16, 9705, 4709, 2449, 16, 3716, 4709, 2449, 16, 7896, 4709, 2449, 16, 25609, 4709, 2449, 65, 31811, 67, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 5254, 13332, 31811, 67, 16341, 273, 306, 4201, 54, 52, 1268, 395, 2449, 16, 9705, 4709, 2449, 16, 3716, 4709, 2449, 16, 7896, 4709, 2449, 16, 25609, 4709, 2449, 65, 31811, 67, ...
result = jobMonitor.setJobParameters(self.jobID,parameters)
def sendStoredJobParameters(self): """ Send the job parameters stored in the internal cache """
51caaa7e74aea392fd6679301a6ed710987614fd /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12864/51caaa7e74aea392fd6679301a6ed710987614fd/JobReport.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1366, 18005, 2278, 2402, 12, 2890, 4672, 3536, 2479, 326, 1719, 1472, 4041, 316, 326, 2713, 1247, 3536, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1366, 18005, 2278, 2402, 12, 2890, 4672, 3536, 2479, 326, 1719, 1472, 4041, 316, 326, 2713, 1247, 3536, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10...
"""If there's nothing nearby, do nothing If cursor is on a
"""If there's nothing nearby, do nothing. If cursor is on a
def leftShiftDown(self, event): """If there's nothing nearby, do nothing If cursor is on a singlet, drag it around, rotating the atom it's bonded to if possible. If it is a real atom, drag it around. """ a = self.o.selatom if not a: return # now, if something was "lit up" self.modified = 1 self.o.assy.modified = 1 if a.element == Singlet: pivatom = a.neighbors()[0] neigh = pivatom.realNeighbors() self.baggage = pivatom.singNeighbors() self.baggage.remove(a) if neigh: if len(neigh)==2: self.pivot = pivatom.posn() self.pivax = norm(neigh[0].posn()-neigh[1].posn()) self.baggage = [] elif len(neigh)>2: self.pivot = None self.pivax = None self.baggage = [] else: # atom on a single stalk self.pivot = pivatom.posn() self.pivax = norm(self.pivot-neigh[0].posn()) else: # no real neighbors self.pivot = pivatom.posn() self.pivax = None else: # we've grabbed an atom self.pivot = None self.pivax = None self.baggage = a.singNeighbors() self.dragatom = a self.w.update()
9bb0cf1da79ef56a2947c3b48ffbb6563ab8c0eb /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/9bb0cf1da79ef56a2947c3b48ffbb6563ab8c0eb/depositMode.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2002, 10544, 4164, 12, 2890, 16, 871, 4672, 3536, 2047, 1915, 1807, 5083, 13378, 1637, 16, 741, 5083, 18, 971, 3347, 353, 603, 279, 4864, 1810, 16, 8823, 518, 6740, 16, 4168, 1776, 326, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2002, 10544, 4164, 12, 2890, 16, 871, 4672, 3536, 2047, 1915, 1807, 5083, 13378, 1637, 16, 741, 5083, 18, 971, 3347, 353, 603, 279, 4864, 1810, 16, 8823, 518, 6740, 16, 4168, 1776, 326, ...
def main(input, output): p = Doxy2SWIG(input)
def convert(input, output, include_function_definition=True): p = Doxy2SWIG(input, include_function_definition)
def main(input, output): p = Doxy2SWIG(input) p.generate() p.write(output)
5406ea81f76a42a2764112ef6fe5ea7842c15201 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/1130/5406ea81f76a42a2764112ef6fe5ea7842c15201/doxy2swig.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1765, 12, 2630, 16, 876, 16, 2341, 67, 915, 67, 6907, 33, 5510, 4672, 293, 273, 2256, 1698, 22, 18746, 3047, 12, 2630, 16, 2341, 67, 915, 67, 6907, 13, 293, 18, 7163, 1435, 293, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1765, 12, 2630, 16, 876, 16, 2341, 67, 915, 67, 6907, 33, 5510, 4672, 293, 273, 2256, 1698, 22, 18746, 3047, 12, 2630, 16, 2341, 67, 915, 67, 6907, 13, 293, 18, 7163, 1435, 293, 18, ...
self.cr.execute("SELECT DISTINCT res_partner.id AS id, " \ "res_partner.name AS name " \ "FROM res_partner,account_move_line AS line, account_account " \ "WHERE (line.account_id=account_account.id) " \ "AND ((reconcile_id IS NULL) " \ "OR (reconcile_id IN (SELECT recon.id FROM account_move_reconcile AS recon WHERE recon.create_date > %s ))) " \ "AND (line.partner_id=res_partner.id) " \ "AND (account_account.company_id = %s) " \ "ORDER BY res_partner.name", (form['date1'],form['company_id']))
self.cr.execute("""SELECT DISTINCT res_partner.id AS id, res_partner.name AS name FROM res_partner,account_move_line AS line, account_account WHERE (line.account_id=account_account.id) AND ((reconcile_id IS NULL) OR (reconcile_id IN (SELECT recon.id FROM account_move_reconcile AS recon WHERE recon.create_date > '%s' ))) AND (line.partner_id=res_partner.id) AND (account_account.company_id = %s) ORDER BY res_partner.name""" % (form['date1'],form['company_id']))
def _get_lines(self, form):
643490cad73510344d6ed02d798ee204cbf1bfb7 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7397/643490cad73510344d6ed02d798ee204cbf1bfb7/aged_trial_balance.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 588, 67, 3548, 12, 2890, 16, 646, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 588, 67, 3548, 12, 2890, 16, 646, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
dart.edge().setFlag(contourProtection)
dart.edge().setFlag(CONTOUR_SEGMENT)
def fakedConstrainedDelaunayMap(points, jumpPoints, imageSize, contourProtection = CONTOUR_PROTECTION, onlyInner = True): """See constrainedDelaunayMap, this calculates a DT and throws away outer edges retroactively. This may fail when the DT does not contain all constrained segments, which is checked in this function and leads to an AssertionError.""" print "- performing Delaunay Triangulation (%d points)..." % len(points) nodePositions, edges, sigma = hourglass.delaunay(points) print "- storing result in a GeoMap..." result = _delaunayMapFromData(nodePositions, edges, imageSize, sigma) if not contourProtection: return result print "- ex-post marking of contour edges for faked CDT..." print " (keep your fingers crossed that no segment is missing!)" edgeSourceDarts = [None] * result.maxEdgeLabel() i = 0 while i < len(jumpPoints) - 1: contourStartLabel = jumpPoints[i] + 1 # +1 for conversion of contourEndLabel = jumpPoints[i+1] + 1 # point indices -> node labels dart = result.node(contourStartLabel).anchor() for nodeLabel in range(contourStartLabel+1, contourEndLabel): j = dart.startNode().degree() + 2 while dart.endNodeLabel() != nodeLabel and j > 0: dart.nextSigma() j -= 1 assert j > 0, """Original contour fragment missing in Delauny map! (This is a problem of the fakedConstrainedDelaunayMap, try compiling the triangle module and using the real constrainedDelaunayMap instead.)""" dart.edge().setFlag(contourProtection) edgeSourceDarts[dart.edgeLabel()] = dart.label() dart.nextAlpha() j = dart.startNode().degree() + 1 while dart.endNodeLabel() != contourStartLabel and j > 0: dart.nextSigma() j -= 1 assert j > 0, "Original contour fragment missing in Delauny map!" dart.edge().setFlag(contourProtection) edgeSourceDarts[dart.edgeLabel()] = dart.label() i += 1 if onlyInner: sys.stdout.write("- reducing delaunay triangulation to inner part...") outerFace = [False] * result.maxFaceLabel() for edge in result.edgeIter(): if edge.flags() & contourProtection: dart = result.dart(edgeSourceDarts[edge.label()]) assert dart.edgeLabel() == edge.label(), str(edge) while True: mergeDart = dart.clone().prevSigma() if mergeDart.edge().flags() & contourProtection: break outerFace[result.mergeFaces(mergeDart).label()] = True sys.stdout.write(".") sys.stdout.flush() sys.stdout.write("\n") return result
9780fdb920d074c0946e5e96a8fed5fb752733ff /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/10394/9780fdb920d074c0946e5e96a8fed5fb752733ff/delaunay.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 284, 9477, 442, 31142, 2837, 69, 318, 528, 863, 12, 4139, 16, 11833, 5636, 16, 29344, 16, 15332, 16938, 273, 3492, 4296, 1099, 67, 3373, 1448, 3106, 16, 1338, 2857, 273, 1053, 4672, 225,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 284, 9477, 442, 31142, 2837, 69, 318, 528, 863, 12, 4139, 16, 11833, 5636, 16, 29344, 16, 15332, 16938, 273, 3492, 4296, 1099, 67, 3373, 1448, 3106, 16, 1338, 2857, 273, 1053, 4672, 225,...
if target in self.targets: self.targets.remove(target)
if target in self.targets: self.targets.remove(target)
def _remove_target_addr(self, target): if target in self.targets: self.targets.remove(target)
66eda46415e31c1e17a158bdfb92fc82dc56f58c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3762/66eda46415e31c1e17a158bdfb92fc82dc56f58c/soat.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 4479, 67, 3299, 67, 4793, 12, 2890, 16, 1018, 4672, 309, 1018, 316, 365, 18, 11358, 30, 365, 18, 11358, 18, 4479, 12, 3299, 13, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 4479, 67, 3299, 67, 4793, 12, 2890, 16, 1018, 4672, 309, 1018, 316, 365, 18, 11358, 30, 365, 18, 11358, 18, 4479, 12, 3299, 13, 2, -100, -100, -100, -100, -100, -100, -100, -100, ...
result += "| " + format(res[i][1]) + "\n"
result += "| " + format(i[1]) + "\n"
def create_window_overview_table(data): """Creates a table for all available windows.""" #matches a line like # Addon_connect The dialog to connect to the addon server variable = "(?:[a-z]|[A-Z])(?:[a-z]|[A-Z]|[0-9]|_)*" regex = re.compile(" *(" + variable + ") +(.*)\n") res = regex.findall(data)
03dd1fa87e7ed6a1aae10ff4e922ba7db9cd3ff8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9355/03dd1fa87e7ed6a1aae10ff4e922ba7db9cd3ff8/wiki_grabber.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 67, 5668, 67, 26081, 67, 2121, 12, 892, 4672, 3536, 2729, 279, 1014, 364, 777, 2319, 9965, 12123, 468, 8436, 279, 980, 3007, 468, 1436, 265, 67, 3612, 1171, 1021, 6176, 358, 3077, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 67, 5668, 67, 26081, 67, 2121, 12, 892, 4672, 3536, 2729, 279, 1014, 364, 777, 2319, 9965, 12123, 468, 8436, 279, 980, 3007, 468, 1436, 265, 67, 3612, 1171, 1021, 6176, 358, 3077, ...
if self.dictOrigin['Program'] == "hyp2000":
if self.dictOrigin['Program'] == "hyp2000" and \ 'SsynthWeight' in d:
def dicts2XML(self): """ Returns information of all dictionaries as xml file """ xml = Element("event") Sub(Sub(xml, "event_id"), "value").text = self.dictEvent['xmlEventID'] event_type = Sub(xml, "event_type") Sub(event_type, "value").text = "manual" Sub(event_type, "user").text = self.username Sub(event_type, "public").text = "%s" % self.flagPublicEvent # XXX standard values for unset keys!!!???!!!??? epidists = [] # go through all stream-dictionaries and look for picks for i in range(len(self.streams)): d = self.dicts[i] st = self.streams[i]
b822b77814749ba0fa0b0a638d2c2e8ccf514f6c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10470/b822b77814749ba0fa0b0a638d2c2e8ccf514f6c/obspyck.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 15838, 22, 4201, 12, 2890, 4672, 3536, 2860, 1779, 434, 777, 16176, 487, 2025, 585, 3536, 2025, 273, 225, 3010, 2932, 2575, 7923, 2592, 12, 1676, 12, 2902, 16, 315, 2575, 67, 350, 6, 3...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 15838, 22, 4201, 12, 2890, 4672, 3536, 2860, 1779, 434, 777, 16176, 487, 2025, 585, 3536, 2025, 273, 225, 3010, 2932, 2575, 7923, 2592, 12, 1676, 12, 2902, 16, 315, 2575, 67, 350, 6, 3...
start=[]
def main(): docorrections=True start=[] gen = None genFactory = pagegenerators.GeneratorFactory() for arg in wikipedia.handleArgs(): if arg == '-onlynew': docorrections=False else: generator = genFactory.handleArg(arg) if generator: gen = generator if gen: try: for p in generator: try: text=p.get() cats=p.categories() if cats == []: wikipedia.output(u"========== %s ==========" % p.title()) print "No categories" print "----------------------------------------" newcats=choosecats(text) if newcats != [] and newcats is not None: make_categories(p, newcats, mysite) else: if docorrections: wikipedia.output(u"========== %s ==========" % p.title()) for c in cats: print c.title() print "----------------------------------------" newcats=choosecats(text) if newcats == None: make_categories(p, [], mysite) elif newcats != []: make_categories(p, newcats, mysite) except wikipedia.IsRedirectPage: pass finally: pass else: wikipedia.showHelp('catall')
8f6f6f483ed9be711e2173b5e4718262d618f4ab /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/4404/8f6f6f483ed9be711e2173b5e4718262d618f4ab/catall.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 997, 280, 9329, 87, 33, 5510, 3157, 273, 599, 225, 3157, 1733, 273, 1363, 25959, 18, 3908, 1733, 1435, 225, 364, 1501, 316, 21137, 18, 4110, 2615, 13332, 309, 1501, 422, 240...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 997, 280, 9329, 87, 33, 5510, 3157, 273, 599, 225, 3157, 1733, 273, 1363, 25959, 18, 3908, 1733, 1435, 225, 364, 1501, 316, 21137, 18, 4110, 2615, 13332, 309, 1501, 422, 240...
s.appendExten("9","Goto(%s,%s,1)" % (pbx,self.phone_9), self.pbx)
s.appendExten("9","Goto(%s,%s,1)" % (pbx,self.phone_9), context)
def createAsteriskConfig(self): s = AstConf("extensions.conf") s.setSection(self.name) if self.pbx: pbx = self.pbx else: pbx = "phones" s.append("include=%s" % pbx) s.appendExten("s","Set(TIMEOUT(digit)=%d)" % self.digittimeout, self.pbx) if self.moh: s.appendExten("s","Setmusiconhold(%s)" % self.moh, self.pbx) s.appendExten("s","Set(DIAL_OPTIONS=m)", self.pbx) else: s.appendExten("s","Set(DIAL_OPTIONS=r)", self.pbx) if self.ivrtime: if self.times: times=self.times.split(',') for t in times: s.appendExten("s","GotoIfTime(%s?%s,s,1)" % (t,self.ivrtime), self.pbx) if self.timeout: s.appendExten("s","Set(TIMEOUT(absolute)=%s)" % self.timeout, self.pbx) for i in range(self.repeat): s.appendExten("s","Background(%s)" % self.backgroundfile, self.pbx) if self.pause: s.appendExten("s","WaitExten(%s)" % self.pause, self.pbx) s.appendExten("s","WaitExten(%d)" % self.waittime, self.pbx) s.appendExten("s","Goto(%s,%s,1)" % (pbx,self.operator), self.pbx) if self.ivr_1: s.appendExten("1","Goto(%s,s,1)" % self.ivr_1, self.pbx) if self.ivr_2: s.appendExten("2","Goto(%s,s,1)" % self.ivr_2, self.pbx) if self.ivr_3: s.appendExten("3","Goto(%s,s,1)" % self.ivr_3, self.pbx) if self.ivr_4: s.appendExten("4","Goto(%s,s,1)" % self.ivr_4, self.pbx) if self.ivr_5: s.appendExten("5","Goto(%s,s,1)" % self.ivr_5, self.pbx) if self.ivr_6: s.appendExten("6","Goto(%s,s,1)" % self.ivr_6, self.pbx) if self.ivr_7: s.appendExten("7","Goto(%s,s,1)" % self.ivr_7, self.pbx) if self.ivr_8: s.appendExten("8","Goto(%s,s,1)" % self.ivr_8, self.pbx) if self.ivr_9: s.appendExten("9","Goto(%s,s,1)" % self.ivr_9, self.pbx) if self.ivr_ast: s.appendExten("*","Goto(%s,s,1)" % self.ivr_ast, self.pbx) if self.phone_1: s.appendExten("1","Goto(%s,%s,1)" % (pbx,self.phone_1), self.pbx) if self.phone_2: s.appendExten("2","Goto(%s,%s,1)" % (pbx,self.phone_2), self.pbx) if self.phone_3: s.appendExten("3","Goto(%s,%s,1)" % (pbx,self.phone_3), self.pbx) if self.phone_4: s.appendExten("4","Goto(%s,%s,1)" % (pbx,self.phone_4), self.pbx) if self.phone_5: s.appendExten("5","Goto(%s,%s,1)" % (pbx,self.phone_5), self.pbx) if self.phone_6: s.appendExten("6","Goto(%s,%s,1)" % (pbx,self.phone_6), self.pbx) if self.phone_7: s.appendExten("7","Goto(%s,%s,1)" % (pbx,self.phone_7), self.pbx) if self.phone_8: s.appendExten("8","Goto(%s,%s,1)" % (pbx,self.phone_8), self.pbx) if self.phone_9: s.appendExten("9","Goto(%s,%s,1)" % (pbx,self.phone_9), self.pbx) if self.phone_ast: s.appendExten("*","Goto(%s,%s,1)" % (pbx,self.phone_ast), self.pbx) s.appendExten("0", "Goto(%s,%s,1)" % (pbx,self.operator), self.pbx) s.appendExten("i","Playback(invalid)", self.pbx) s.appendExten("i","Goto(%s,%s,1)" % (pbx,self.operator), self.pbx) s.appendExten("t","ResetCDR(w)", self.pbx) s.appendExten("t","NoCDR", self.pbx) s.appendExten("t","Hangup", self.pbx) s.appendExten("T","ResetCDR(w)", self.pbx) s.appendExten("T","NoCDR", self.pbx) s.appendExten("T","Hangup", self.pbx) s.appendExten("#","ResetCDR(w)", self.pbx) s.appendExten("#","NoCDR", self.pbx) s.appendExten("#","Hangup", self.pbx)
e581660927aebaacc03692e1a06a4d11542ff6a3 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2627/e581660927aebaacc03692e1a06a4d11542ff6a3/cfg_ivr_treenode.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 37, 8190, 10175, 809, 12, 2890, 4672, 272, 273, 16614, 3976, 2932, 9489, 18, 3923, 7923, 272, 18, 542, 5285, 12, 2890, 18, 529, 13, 309, 365, 18, 5733, 92, 30, 6386, 92, 273, 36...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 37, 8190, 10175, 809, 12, 2890, 4672, 272, 273, 16614, 3976, 2932, 9489, 18, 3923, 7923, 272, 18, 542, 5285, 12, 2890, 18, 529, 13, 309, 365, 18, 5733, 92, 30, 6386, 92, 273, 36...
REFERENCES: Howard M. Heys, A Tutorial on Linear and Differential Cryptanalysis, Cryptologia, v.XXVI n.3, p.189-221, July 2002
def difference_distribution_matrix(self): """ Return difference distribution matrix $A$ for this S-box.
dd81b4a97db8b232434d8623d237dba45e418834 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/dd81b4a97db8b232434d8623d237dba45e418834/sbox.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7114, 67, 16279, 67, 5667, 12, 2890, 4672, 3536, 2000, 7114, 7006, 3148, 271, 37, 8, 364, 333, 348, 17, 2147, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7114, 67, 16279, 67, 5667, 12, 2890, 4672, 3536, 2000, 7114, 7006, 3148, 271, 37, 8, 364, 333, 348, 17, 2147, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
f90 = self.executables['compiler_f90']
f90 = self.executables.get('compiler_f90')
def get_version_cmd(self): """ Compiler command to print out version information. """ f77 = self.executables['compiler_f77'] if f77 is not None: f77 = f77[0] cmd = self.executables['version_cmd'] if cmd is not None: cmd = cmd[0] if cmd==f77: cmd = self.compiler_f77[0] else: f90 = self.executables['compiler_f90'] if f90 is not None: f90 = f90[0] if cmd==f90: cmd = self.compiler_f90[0] return cmd
9ddd860b31ea1d4517eb3fff6ab4c280ebb14dec /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/14925/9ddd860b31ea1d4517eb3fff6ab4c280ebb14dec/__init__.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 1589, 67, 4172, 12, 2890, 4672, 3536, 12972, 1296, 358, 1172, 596, 1177, 1779, 18, 3536, 284, 4700, 273, 365, 18, 4177, 322, 1538, 3292, 9576, 67, 74, 4700, 3546, 309, 284, 47...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 1589, 67, 4172, 12, 2890, 4672, 3536, 12972, 1296, 358, 1172, 596, 1177, 1779, 18, 3536, 284, 4700, 273, 365, 18, 4177, 322, 1538, 3292, 9576, 67, 74, 4700, 3546, 309, 284, 47...
self._finalize_connect(prefinished)
def _changed(self, action): """ Called when a callback connects or disconnects from us. """ if len(self) == 1 and action == Signal.SIGNAL_CONNECTED and not self.finished: # Someone wants to know when we finish, so now we connect to the # underlying InProgress objects to find out when they finish. prefinished = [] for n, ip in enumerate(self._objects): if ip.finished: # This one is finished already, no need to connect to it. prefinished.append(n) continue args = self._get_connect_args(ip, n) ip.connect(self.finish, False, *args).user_args_first = True ip.exception.connect(self.finish, True, *args).user_args_first = True
0d9f6bc38cba8464d9f7e2cdc2c065c63763827b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11722/0d9f6bc38cba8464d9f7e2cdc2c065c63763827b/async.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 6703, 12, 2890, 16, 1301, 4672, 3536, 11782, 1347, 279, 1348, 23441, 578, 9479, 87, 628, 584, 18, 3536, 309, 562, 12, 2890, 13, 422, 404, 471, 1301, 422, 12032, 18, 11260, 1013, 6...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 6703, 12, 2890, 16, 1301, 4672, 3536, 11782, 1347, 279, 1348, 23441, 578, 9479, 87, 628, 584, 18, 3536, 309, 562, 12, 2890, 13, 422, 404, 471, 1301, 422, 12032, 18, 11260, 1013, 6...
cannot_set_attr(self.b, '__name__', 7, TypeError) cannot_set_attr(self.b, 'func_name', 7, TypeError)
self.cannot_set_attr(self.b, '__name__', 7, TypeError) self.cannot_set_attr(self.b, 'func_name', 7, TypeError)
def test_func_name(self): self.assertEqual(self.b.__name__, 'b') self.assertEqual(self.b.func_name, 'b') self.b.__name__ = 'c' self.assertEqual(self.b.__name__, 'c') self.assertEqual(self.b.func_name, 'c') self.b.func_name = 'd' self.assertEqual(self.b.__name__, 'd') self.assertEqual(self.b.func_name, 'd') # __name__ and func_name must be a string cannot_set_attr(self.b, '__name__', 7, TypeError) cannot_set_attr(self.b, 'func_name', 7, TypeError) # __name__ must be available when in restricted mode. Exec will raise # AttributeError if __name__ is not available on f. s = """def f(): pass\nf.__name__""" exec s in {'__builtins__': {}} # Test on methods, too self.assertEqual(self.f.a.__name__, 'a') self.assertEqual(self.fi.a.__name__, 'a') cannot_set_attr(self.f.a, "__name__", 'a', AttributeError) cannot_set_attr(self.fi.a, "__name__", 'a', AttributeError)
8ada17742baaa36357dbdf70009d247c1863c291 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8546/8ada17742baaa36357dbdf70009d247c1863c291/test_funcattrs.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 644, 67, 529, 12, 2890, 4672, 365, 18, 11231, 5812, 12, 2890, 18, 70, 16186, 529, 972, 16, 296, 70, 6134, 365, 18, 11231, 5812, 12, 2890, 18, 70, 18, 644, 67, 529, 16, 29...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 644, 67, 529, 12, 2890, 4672, 365, 18, 11231, 5812, 12, 2890, 18, 70, 16186, 529, 972, 16, 296, 70, 6134, 365, 18, 11231, 5812, 12, 2890, 18, 70, 18, 644, 67, 529, 16, 29...
raise OSError(0, 'No such directory', path)
raise OSError(0, 'No such directory', path)
def listdir(path): l = File(path).list() if l is None: raise OSError(0, 'No such directory', path) return list(l)
5983a691b16eecde718073ce68a1624ff2a162aa /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6753/5983a691b16eecde718073ce68a1624ff2a162aa/javaos.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13529, 12, 803, 4672, 328, 273, 1387, 12, 803, 2934, 1098, 1435, 309, 328, 353, 599, 30, 1002, 10002, 12, 20, 16, 296, 2279, 4123, 1867, 2187, 589, 13, 327, 666, 12, 80, 13, 225, 2, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13529, 12, 803, 4672, 328, 273, 1387, 12, 803, 2934, 1098, 1435, 309, 328, 353, 599, 30, 1002, 10002, 12, 20, 16, 296, 2279, 4123, 1867, 2187, 589, 13, 327, 666, 12, 80, 13, 225, 2, ...
if not subitem in result and not subitem in excludeWithDeps and not subitem in deps["warn"]:
if not subitem in result and not subitem in excludeWithDeps and not subitem in skipList:
def classlistFromClassRecursive(item, excludeWithDeps, variants, result): # support blocking if item in excludeWithDeps: return
779574821bd05657d9aed278a2d44657763d63b8 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5718/779574821bd05657d9aed278a2d44657763d63b8/DependencyLoader.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 667, 1098, 1265, 797, 10483, 12, 1726, 16, 4433, 1190, 14430, 16, 12935, 16, 563, 4672, 468, 2865, 9445, 309, 761, 316, 4433, 1190, 14430, 30, 327, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 667, 1098, 1265, 797, 10483, 12, 1726, 16, 4433, 1190, 14430, 16, 12935, 16, 563, 4672, 468, 2865, 9445, 309, 761, 316, 4433, 1190, 14430, 30, 327, 2, -100, -100, -100, -100, -100, -100,...