rem
stringlengths
2
226k
add
stringlengths
0
227k
context
stringlengths
8
228k
meta
stringlengths
156
215
input_ids
list
attention_mask
list
labels
list
H = subgraph_search(self, G, induced = True)
from sage.graphs.graph_generators import GraphGenerators if G.order() == 0: return GraphGenerators().EmptyGraph() H = subgraph_search(self, G, induced=induced)
def induced_subgraph_search(self, G): r""" Returns an induced copy of `G` in self.
ff7673c4297965681e16ef36d204d5f5d4c2273d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9890/ff7673c4297965681e16ef36d204d5f5d4c2273d/generic_graph.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1547, 26372, 67, 1717, 4660, 67, 3072, 12, 2890, 16, 611, 4672, 436, 8395, 2860, 392, 1547, 26372, 1610, 434, 1375, 43, 68, 316, 365, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1547, 26372, 67, 1717, 4660, 67, 3072, 12, 2890, 16, 611, 4672, 436, 8395, 2860, 392, 1547, 26372, 1610, 434, 1375, 43, 68, 316, 365, 18, 2, -100, -100, -100, -100, -100, -100, -100, -...
):
)):
def mime_collector_finalize (self): # wake up the pipeline if there are request pipelined # and that the connection is kept alive if ( self.pipeline_requests and self.mime_collector_headers.get ( 'connection' ) == 'keep-alive' ): self.pipeline_wake_up () # reset the channel state to collect the next request ... self.set_terminator ('\r\n\r\n') self.mime_collector_headers = \ self.mime_collector_lines = \ self.mime_collector_body = None # pop the reactor finalized and count the response self.pipeline_responses.popleft () self.http_responses += 1 return self.closing # stall the collector if closing
85b502d28f914926bc29a97c9ab6cd9f4d496329 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2577/85b502d28f914926bc29a97c9ab6cd9f4d496329/http_client.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4892, 67, 21356, 67, 30343, 261, 2890, 3719, 30, 468, 19891, 731, 326, 5873, 309, 1915, 854, 590, 8341, 292, 1920, 468, 471, 716, 326, 1459, 353, 16555, 13714, 309, 261, 365, 18, 14511, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4892, 67, 21356, 67, 30343, 261, 2890, 3719, 30, 468, 19891, 731, 326, 5873, 309, 1915, 854, 590, 8341, 292, 1920, 468, 471, 716, 326, 1459, 353, 16555, 13714, 309, 261, 365, 18, 14511, ...
selection = schema.One(schema.Item, initialValue = None) schema.addClouds( copying = schema.Cloud(byRef=[selection]) )
def installParcel(parcel, oldVersion=None): # Declare our attribute editors at repository-init time # # If you modify this list, please keep it in alphabetical order by type string. # Also, note that there are more attribute editors custom to the detail # view; they're declared in its installParcel method. aeList = { 'DateTime+calendarDateOnly': 'CalendarDateAttributeEditor', 'DateTime+calendarTimeOnly': 'CalendarTimeAttributeEditor', 'EmailAddress+outgoing': 'OutgoingEmailAddressAttributeEditor', 'RecurrenceRuleSet+custom': 'RecurrenceCustomAttributeEditor', 'RecurrenceRuleSet+ends': 'RecurrenceEndsAttributeEditor', 'RecurrenceRuleSet+occurs': 'RecurrenceAttributeEditor', 'TimeDelta+reminderPopup': 'ReminderDeltaAttributeEditor', } for typeName, className in aeList.items(): AttributeEditorMapping.update(parcel, typeName, className=\ __name__ + '.' + className)
4ccd0e6c804f648ff0b37b18090947c30b4b7e2f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/4ccd0e6c804f648ff0b37b18090947c30b4b7e2f/Detail.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3799, 1553, 2183, 12, 1065, 2183, 16, 1592, 1444, 33, 7036, 4672, 468, 16110, 834, 3134, 1566, 29431, 622, 3352, 17, 2738, 813, 468, 468, 971, 1846, 5612, 333, 666, 16, 9582, 3455, 518, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3799, 1553, 2183, 12, 1065, 2183, 16, 1592, 1444, 33, 7036, 4672, 468, 16110, 834, 3134, 1566, 29431, 622, 3352, 17, 2738, 813, 468, 468, 971, 1846, 5612, 333, 666, 16, 9582, 3455, 518, ...
c += SemistandardTableaux(part).cardinality()
c += SemistandardTableaux_p(part, self.max_entry).cardinality()
def cardinality(self): """ EXAMPLES:: sage: SemistandardTableaux(3).cardinality() 19 sage: SemistandardTableaux(4).cardinality() 116 sage: ns = range(1, 6) sage: ssts = [ SemistandardTableaux(n) for n in ns ] sage: all([sst.cardinality() == len(sst.list()) for sst in ssts]) True """ c = 0 for part in partition.Partitions(self.n): c += SemistandardTableaux(part).cardinality() return c
6ec7d5a4cd0b83ae5413f0375d51a0339f12b4b7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9890/6ec7d5a4cd0b83ae5413f0375d51a0339f12b4b7/tableau.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 14379, 12, 2890, 4672, 3536, 5675, 8900, 11386, 2866, 225, 272, 410, 30, 15661, 376, 2958, 1388, 18196, 12, 23, 2934, 3327, 9232, 1435, 5342, 272, 410, 30, 15661, 376, 2958, 1388, 18196, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 14379, 12, 2890, 4672, 3536, 5675, 8900, 11386, 2866, 225, 272, 410, 30, 15661, 376, 2958, 1388, 18196, 12, 23, 2934, 3327, 9232, 1435, 5342, 272, 410, 30, 15661, 376, 2958, 1388, 18196, ...
'<a href="http://localhost:8888/echo">back to referring page</a></div>'
'<a href="/echo">back to referring page</a></div>'
def EchoAllHandler(self): """This handler yields a (more) human-readable page listing information about the request header & contents."""
df0401f579455dcc6e70ad63ffae1704549cf148 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5060/df0401f579455dcc6e70ad63ffae1704549cf148/testserver.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 28995, 1595, 1503, 12, 2890, 4672, 3536, 2503, 1838, 16932, 279, 261, 10161, 13, 8672, 17, 11018, 1363, 11591, 1779, 2973, 326, 590, 1446, 473, 2939, 12123, 2, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 28995, 1595, 1503, 12, 2890, 4672, 3536, 2503, 1838, 16932, 279, 261, 10161, 13, 8672, 17, 11018, 1363, 11591, 1779, 2973, 326, 590, 1446, 473, 2939, 12123, 2, -100, -100, -100, -100, -100...
del md5Cache[tmpPath]
del md5Cache[tmpFileName]
def run(self): log.log(2,"info",_("New XiboDownloadManager instance started.")) while (self.running): self.interval = 300
66a075cab3d68c237f51e86004d5840adb9c1f9c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5464/66a075cab3d68c237f51e86004d5840adb9c1f9c/XiboClient.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 4672, 613, 18, 1330, 12, 22, 10837, 1376, 3113, 67, 2932, 1908, 1139, 495, 83, 7109, 1318, 791, 5746, 1199, 3719, 1323, 261, 2890, 18, 8704, 4672, 365, 18, 6624, 273, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 4672, 613, 18, 1330, 12, 22, 10837, 1376, 3113, 67, 2932, 1908, 1139, 495, 83, 7109, 1318, 791, 5746, 1199, 3719, 1323, 261, 2890, 18, 8704, 4672, 365, 18, 6624, 273, 1...
db_generation = row and row[0] or -1
db_generation = not row and -1 or row[0]
def get(self, id, retriever, db=None): """Get cached or fresh data for the given id.""" # Get cache metadata try: local_meta = self._local.meta local_cache = self._local.cache except AttributeError: # First cache usage in this request, retrieve cache metadata # from the database and make a thread-local copy of the cache if db is None: db = self.env.get_db_cnx() cursor = db.cursor() cursor.execute("SELECT id,generation FROM cache") self._local.meta = local_meta = dict(cursor) self._local.cache = local_cache = self._cache.copy() db_generation = local_meta.get(id, -1) # Try the thread-local copy first try: (data, generation) = local_cache[id] if generation == db_generation: return data except KeyError: pass self._lock.acquire() try: # Get data from the process cache try: (data, generation) = local_cache[id] = self._cache[id] if generation == db_generation: return data except KeyError: generation = None # Force retrieval from the database # Check if the process cache has the newest version, as it may # have been updated after the metadata retrieval if db is None: db = self.env.get_db_cnx() cursor = db.cursor() cursor.execute("SELECT generation FROM cache WHERE id=%s", (id,)) row = cursor.fetchone() db_generation = row and row[0] or -1 if db_generation == generation: return data # Retrieve data from the database data = retriever(db) local_cache[id] = self._cache[id] = (data, db_generation) local_meta[id] = db_generation return data finally: self._lock.release()
1dc36aec1545d9484eed74126f9bf114844941de /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9317/1dc36aec1545d9484eed74126f9bf114844941de/cache.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 12, 2890, 16, 612, 16, 3227, 502, 16, 1319, 33, 7036, 4672, 3536, 967, 3472, 578, 12186, 501, 364, 326, 864, 612, 12123, 468, 968, 1247, 1982, 775, 30, 1191, 67, 3901, 273, 365, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 12, 2890, 16, 612, 16, 3227, 502, 16, 1319, 33, 7036, 4672, 3536, 967, 3472, 578, 12186, 501, 364, 326, 864, 612, 12123, 468, 968, 1247, 1982, 775, 30, 1191, 67, 3901, 273, 365, ...
debug(" cookie expired")
_debug(" cookie expired")
def return_ok_expires(self, cookie, request): if cookie.is_expired(self._now): debug(" cookie expired") return False return True
7fde00696ae6c872310a407df8b8923ab907f247 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/7fde00696ae6c872310a407df8b8923ab907f247/cookielib.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 327, 67, 601, 67, 12431, 12, 2890, 16, 3878, 16, 590, 4672, 309, 3878, 18, 291, 67, 15820, 12, 2890, 6315, 3338, 4672, 389, 4148, 2932, 282, 3878, 7708, 7923, 327, 1083, 327, 1053, 2, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 327, 67, 601, 67, 12431, 12, 2890, 16, 3878, 16, 590, 4672, 309, 3878, 18, 291, 67, 15820, 12, 2890, 6315, 3338, 4672, 389, 4148, 2932, 282, 3878, 7708, 7923, 327, 1083, 327, 1053, 2, ...
self.encodeddata = urlencode(self.data, 1) self.headers = { "Content-type": "application/x-www-form-urlencoded", "User-agent": wiki.useragent, "Content-length": len(self.encodeddata) }
self.multipart = multipart if self.multipart: (datagen, self.headers) = multipart_encode(self.data) self.encodeddata = '' for singledata in datagen: self.encodeddata = self.encodeddata + singledata else: self.encodeddata = urlencode(self.data, 1) self.headers = { "Content-type": "application/x-www-form-urlencoded", "Content-length": len(self.encodeddata) } self.headers["User-agent"] = wiki.useragent,
def __init__(self, wiki, data, write=False): """ wiki - A Wiki object data - API parameters in the form of a dict write - set to True if doing a write query, so it won't try again on error maxlag is set by default to 5 but can be changed format is always set to json """ self.sleep = 5 self.data = data.copy() self.data['format'] = "json" self.iswrite = write if not 'maxlag' in self.data and not wiki.maxlag < 0: self.data['maxlag'] = wiki.maxlag self.encodeddata = urlencode(self.data, 1) self.headers = { "Content-type": "application/x-www-form-urlencoded", "User-agent": wiki.useragent, "Content-length": len(self.encodeddata) } if gzip: self.headers['Accept-Encoding'] = 'gzip' self.wiki = wiki self.response = False self.opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(wiki.cookies)) self.request = urllib2.Request(wiki.apibase, self.encodeddata, self.headers)
17fc397db9db6702301626ec8cd08dee4eeb473c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10544/17fc397db9db6702301626ec8cd08dee4eeb473c/api.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 9050, 16, 501, 16, 1045, 33, 8381, 4672, 3536, 9050, 300, 432, 28268, 733, 501, 300, 1491, 1472, 316, 326, 646, 434, 279, 2065, 1045, 300, 444, 358, 1053...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 9050, 16, 501, 16, 1045, 33, 8381, 4672, 3536, 9050, 300, 432, 28268, 733, 501, 300, 1491, 1472, 316, 326, 646, 434, 279, 2065, 1045, 300, 444, 358, 1053...
def openBouquetEPG(self, bouquet):
def openBouquetEPG(self, bouquet, withCallback=True):
def openBouquetEPG(self, bouquet): ptr=eEPGCache.getInstance() services = [ ] servicelist = eServiceCenter.getInstance().list(bouquet) if not servicelist is None: while True: service = servicelist.getNext() if not service.valid(): #check if end of list break if service.flags: #ignore non playable services continue services.append(ServiceReference(service)) if len(services): self.epg_bouquet = bouquet self.session.openWithCallback(self.closed, EPGSelection, services, self.zapToService)
2828914fb949a9f120285e7473512937ea184694 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6652/2828914fb949a9f120285e7473512937ea184694/InfoBarGenerics.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1696, 38, 1395, 28357, 10541, 43, 12, 2890, 16, 800, 89, 28357, 16, 598, 2428, 33, 5510, 4672, 6571, 33, 73, 10541, 43, 1649, 18, 588, 1442, 1435, 4028, 273, 306, 308, 24658, 5449, 273...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1696, 38, 1395, 28357, 10541, 43, 12, 2890, 16, 800, 89, 28357, 16, 598, 2428, 33, 5510, 4672, 6571, 33, 73, 10541, 43, 1649, 18, 588, 1442, 1435, 4028, 273, 306, 308, 24658, 5449, 273...
dim_ethinca = True
dim_ethinca = False
def getTriggersWithinEpsilon(self, candidate, epsilon, epsilon_ball_type = "version2"):
f423289bd30e1207ae4a4aef9ba4c2fba9f047b1 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/3592/f423289bd30e1207ae4a4aef9ba4c2fba9f047b1/CoincInspiralUtils.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 15114, 18949, 29889, 10327, 12, 2890, 16, 5500, 16, 12263, 16, 12263, 67, 19067, 67, 723, 273, 315, 1589, 22, 6, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 15114, 18949, 29889, 10327, 12, 2890, 16, 5500, 16, 12263, 16, 12263, 67, 19067, 67, 723, 273, 315, 1589, 22, 6, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100,...
if isinstance(file, unicode): file = file.encode('utf-8') epub.write(path.join(outdir, file), file, zipfile.ZIP_DEFLATED)
fp = path.join(outdir, file) if isinstance(fp, unicode): fp = fp.encode(sys.getfilesystemencoding()) epub.write(fp, file, zipfile.ZIP_DEFLATED)
def build_epub(self, outdir, outname): """Write the epub file.
6149c462a3da07f61ec9796790f46cfe9130be08 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5532/6149c462a3da07f61ec9796790f46cfe9130be08/epub.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1361, 67, 881, 373, 12, 2890, 16, 15398, 16, 596, 529, 4672, 3536, 3067, 326, 5529, 373, 585, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1361, 67, 881, 373, 12, 2890, 16, 15398, 16, 596, 529, 4672, 3536, 3067, 326, 5529, 373, 585, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
def getDocumentObjectModel(self): domPointer=ctypes.POINTER(comtypes.automation.IDispatch)() wm=winUser.registerWindowMessage(u'WM_HTML_GETOBJECT') lresult=winUser.sendMessage(self.windowHandle,wm,0,0) res=ctypes.windll.oleacc.ObjectFromLresult(lresult,ctypes.byref(domPointer._iid_),0,ctypes.byref(domPointer))
def __init__(self,*args,**kwargs): super(MSHTML,self).__init__(*args,**kwargs) self.domElement=self.getDOMElementFromIAccessible() def getDOMElementFromIAccessible(self): s=self.IAccessibleObject.QueryInterface(IServiceProvider) interfaceAddress=s.QueryService(ctypes.byref(IID_IHTMLElement),ctypes.byref(IID_IHTMLElement))
def getDocumentObjectModel(self): domPointer=ctypes.POINTER(comtypes.automation.IDispatch)() wm=winUser.registerWindowMessage(u'WM_HTML_GETOBJECT') lresult=winUser.sendMessage(self.windowHandle,wm,0,0) res=ctypes.windll.oleacc.ObjectFromLresult(lresult,ctypes.byref(domPointer._iid_),0,ctypes.byref(domPointer)) #We use pywin32 for large IDispatch interfaces since it handles them much better than comtypes o=pythoncom._univgw.interface(ctypes.cast(domPointer,ctypes.c_void_p).value,pythoncom.IID_IDispatch) t=o.GetTypeInfo() a=t.GetTypeAttr() oleRepr=win32com.client.build.DispatchItem(attr=a) return win32com.client.CDispatch(o,oleRepr)
e414733012e267c4ac0a4ed4c2b245743e7a1b97 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9340/e414733012e267c4ac0a4ed4c2b245743e7a1b97/MSHTML.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 14, 1968, 16, 636, 4333, 4672, 2240, 12, 49, 2664, 2838, 16, 2890, 2934, 972, 2738, 972, 30857, 1968, 16, 636, 4333, 13, 365, 18, 9859, 1046, 33, 2890, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 14, 1968, 16, 636, 4333, 4672, 2240, 12, 49, 2664, 2838, 16, 2890, 2934, 972, 2738, 972, 30857, 1968, 16, 636, 4333, 13, 365, 18, 9859, 1046, 33, 2890, ...
def validate(self):
def validate(self, duration=False):
def validate(self): """ Validate filter condition (template method). """ super(TimeFilter, self).validate() timestamp = time.time()
66ffaac957103fefa1835e46aac39a897a44e302 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4865/66ffaac957103fefa1835e46aac39a897a44e302/matching.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1954, 12, 2890, 16, 3734, 33, 8381, 4672, 3536, 3554, 1034, 2269, 261, 3202, 707, 2934, 3536, 2240, 12, 950, 1586, 16, 365, 2934, 5662, 1435, 2858, 273, 813, 18, 957, 1435, 2, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1954, 12, 2890, 16, 3734, 33, 8381, 4672, 3536, 3554, 1034, 2269, 261, 3202, 707, 2934, 3536, 2240, 12, 950, 1586, 16, 365, 2934, 5662, 1435, 2858, 273, 813, 18, 957, 1435, 2, -100, -1...
skiplist = string.split(opts.skip, ",") optpkgs = string.split(opts.addpackages, ',')
global mainclass global basepath skiplist = opts.skip.split(',') optpkgs = opts.addpackages.split(',')
def doCompile(opts): skiplist = string.split(opts.skip, ",") optpkgs = string.split(opts.addpackages, ',') addpackages = ['org.python.modules', 'com.oroinc.text.regex'] + optpkgs comp = Compiler(javapackage=opts.package, deep=opts.deep, include=addpackages, skip=skiplist, options=opts) global mainclass global basepath for target in opts.args: if target.endswith('.py'): classname = os.path.splitext(os.path.basename(target))[0] filename = target if basepath is None: basepath = os.path.split(target)[0] sys.path.insert(0, basepath) else: classname = target import ImportName m = ImportName.lookupName(classname) if not m: print 'Could not find class:', classname sys.exit(1) filename = m.file if mainclass is None: mainclass = classname if opts.package is not None: mainclass = opts.package+'.'+mainclass comp.compilefile(filename, classname) comp.dump(opts.workdir) return comp
eb21a7b702c912ae33844f4b5e0e4c21af19d89f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6753/eb21a7b702c912ae33844f4b5e0e4c21af19d89f/main.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 741, 9937, 12, 4952, 4672, 2552, 2774, 1106, 2552, 23853, 225, 4343, 77, 17842, 273, 1500, 18, 7457, 18, 4939, 12, 2187, 6134, 2153, 25376, 273, 1500, 18, 1289, 10308, 18, 4939, 12, 2187...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 741, 9937, 12, 4952, 4672, 2552, 2774, 1106, 2552, 23853, 225, 4343, 77, 17842, 273, 1500, 18, 7457, 18, 4939, 12, 2187, 6134, 2153, 25376, 273, 1500, 18, 1289, 10308, 18, 4939, 12, 2187...
safesymlink('../../devel/sage/sage/dsage/web', 'local/dsage/web')
os.system('cp -r %s %s' % (tempdir+'/devel/sage/sage/dsage/web',tempdir+'/local/dsage/'))
def buildsage(env, gccc): """ This function actually builds sage """ TM = taskmanager.TM #acquire a list of all pyx files in the tree efw = extfilewalker() efw.addcallback('.pyx',lambda x: True) pyx_list = efw.walk('devel/sage/sage') #set the absolute devel directory if necessary develdir = os.path.realpath('devel/sage/') if verbose>100: print 'Devel Dir: ' + develdir #Create the cython compiler cyc = Cython_compiler() #We want -I. so that it correctly finds include files by default cyc.set_option_val('-I'+'.',None) #Create the build extension dict for all cython extensions pyx_pextn_dict = cyc.get_build_extensions(env, pyx_list, cwd = 'devel/sage') ################################################################### #Configure all extensions ################################################################### config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/flint/fmpz_poly.pyx",include_dirs = [real_sage_local(env, 'include/FLINT/')], libraries = ["csage", "flint", "gmp", "gmpxx", "m", "stdc++"],prop_options={ str(GCC_extension_object): {'-std':'c99' } } ) ntllibs = ["csage", "ntl", "gmp", "gmpxx", "m", "stdc++"] config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_ZZ.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_ZZX.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_ZZ_pContext.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_ZZ_p.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_ZZ_pX.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_ZZ_pEContext.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_ZZ_pE.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_ZZ_pEX.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_lzz_pContext.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_lzz_p.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_lzz_pX.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_GF2.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_GF2X.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_GF2EContext.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_GF2E.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_GF2EX.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_mat_ZZ.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_mat_GF2E.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/ntl/ntl_GF2EContext.pyx",language="C++", libraries=ntllibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/mwrank/mwrank.pyx", language='C++', define_macros=[("NTL_ALL",None)], libraries=["curvesntl", "g0nntl", "jcntl", "rankntl", "ntl", "gmp", "gmpxx", "stdc++", "m", "pari"] ) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/pari/gen.pyx",libraries=['pari', 'gmp']) cremonalibs = ['g0nntl', 'jcntl', 'gmpxx', 'ntl', 'gmp', 'm', 'stdc++', 'pari', 'curvesntl'] config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/cremona/mat.pyx",language="C++", define_macros=[("NTL_ALL",None)], libraries=cremonalibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/cremona/homspace.pyx",language="C++", define_macros=[("NTL_ALL",None)], libraries=cremonalibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/libs/cremona/newforms.pyx",language="C++", define_macros=[("NTL_ALL",None)], libraries=cremonalibs) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/rings/finite_field_givaro.pyx",language="C++", libraries=['givaro', 'gmpxx', 'gmp', 'm', 'stdc++']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/finite_field_ntl_gf2e.pyx',language="C++", libraries=['ntl', 'gmp']) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/rings/real_rqdf.pyx",language="C++", libraries=['qd', 'm', 'stdc++','gmp','mpfr','qd', 'csage' ]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/misc.pyx',libraries=['mpfr','gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/matrix_integer_2x2.pyx',libraries=['gmp']) BLAS = env.options['BLAS'] BLAS2 = env.options['BLAS2'] config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/libs/linbox/linbox.pyx',language="C++", libraries=['linboxwrap', 'ntl', 'linbox', 'gmp', 'gmpxx', 'stdc++', 'givaro', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/libs/singular/singular.pyx',language='C++', include_dirs=[real_sage_local(env,'include/singular')], libraries=['m', 'readline', 'singular', 'singfac', 'singcf', 'omalloc', 'givaro', 'gmpxx', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/libs/fplll/fplll.pyx',language='C++', include_dirs = [real_sage_local(env, 'include/fplll')], libraries=['gmp', 'mpfr', 'stdc++', 'fplll']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/matrix_modn_dense.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/matrix_mod2_dense.pyx',libraries = ['gmp','m4ri']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/matrix_rational_dense.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/matrix_integer_sparse.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/matrix_rational_sparse.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/matrix_integer_dense.pyx',libraries = ['iml', 'gmp', 'm', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/matrix_real_double_dense.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], include_dirs = [real_sage_local(env,'lib/python2.5/site-packages/numpy/core/include/numpy')], libraries=[BLAS, BLAS2, 'gsl']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/change_ring.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=[BLAS, BLAS2, 'gsl', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/matrix_complex_double_dense.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], include_dirs=[real_sage_local(env,'lib/python2.5/site-packages/numpy/core/include/numpy')], libraries=['gsl', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/solve.pyx',define_macros = [('GSL_DISABLE_DEPRECATED','1')], libraries = ['gsl', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/matrix/matrix_mpolynomial_dense.pyx',language="C++", include_dirs = [real_sage_local(env,'include/singular')], libraries = ['m', 'readline', 'singular', 'singcf', 'singfac', 'omalloc', 'givaro', 'gmpxx', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/complex_number.pyx',libraries = ['mpfr', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/gsl/probability_distribution.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=['gsl', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/gsl/integration.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=['gsl', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/gsl/ode.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=['gsl', BLAS]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/gsl/fft.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=['gsl', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/gsl/interpolation.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=['gsl', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/gsl/callback.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=['gsl', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/real_double.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=['gsl', 'gmp', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/complex_double.pyx',libraries = ['gsl', BLAS, BLAS2, 'pari', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/modules/real_double_vector.pyx',define_macros = [('GSL_DISABLE_DEPRECAED','1')], include_dirs=[real_sage_local(env,'lib/python2.5/site-packages/numpy/core/include/numpy')], libraries = ['gsl', BLAS, BLAS2, 'pari','gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/modules/complex_double_vector.pyx',define_macros = [('GSL_DISABLE_DEPRECAED','1')], include_dirs=[real_sage_local(env,'lib/python2.5/site-packages/numpy/core/include/numpy')], libraries = ['gsl', BLAS, BLAS2, 'pari','gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/modules/vector_integer_dense.pyx',libraries=['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/modules/vector_rational_dense.pyx',libraries=['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/gsl/gsl_array.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=['gsl', BLAS, BLAS2]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/gsl/ode.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=['gsl',BLAS]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/gsl/dwt.pyx',define_macros=[('GSL_DISABLE_DEPRECATED','1')], libraries=['gsl',BLAS]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/libs/symmetrica/symmetrica.pyx',libraries=['symmetrica']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/ext/arith_gmp.pyx',libraries=['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/ext/multi_modular.pyx',libraries=['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/polynomial/multi_polynomial_libsingular.pyx',language='C++', include_dirs = [real_sage_local(env,'include/singular')], libraries = ['m', 'readline', 'singular', 'singcf', 'singfac', 'omalloc', 'givaro', 'gmpxx', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/polynomial/multi_polynomial_ideal_libsingular.pyx',language='C++', include_dirs = [real_sage_local(env,'include/singular')], libraries = ['m', 'readline', 'singular', 'singcf', 'singfac', 'omalloc', 'givaro', 'gmpxx', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/structure/sage_object.pyx',libraries=['ntl']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/real_mpfr.pyx',libraries = ['mpfi', 'mpfr', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/real_mpfi.pyx',libraries = ['mpfi', 'mpfr', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/complex_interval.pyx',libraries = ['mpfi', 'mpfr', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/integer.pyx',libraries = ['ntl', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/integer_ring.pyx',libraries = ['ntl', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/libs/libecm.pyx',libraries = ['ecm', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/pow_computer.pyx',language='C++', libraries = ["csage", "ntl", "gmp", "gmpxx", "m", "stdc++"]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/pow_computer_ext.pyx',language='C++', libraries = ["csage", "ntl", "gmp", "gmpxx", "m", "stdc++"]) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/padic_fixed_mod_element.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/padic_capped_absolute_element.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/padic_capped_relative_element.pyx',libraries = ['gmp', 'csage']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/padic_ext_element.pyx',language = 'C++', libraries = ['gmp', 'ntl', 'csage', 'gmpxx', 'm', 'stdc++']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/padic_ZZ_pX_element.pyx',language = 'C++', libraries = ['gmp', 'ntl', 'csage', 'gmpxx', 'm', 'stdc++']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/padic_ZZ_pX_FM_element.pyx',language = 'C++', libraries = ['gmp', 'ntl', 'csage', 'gmpxx', 'm', 'stdc++']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/padic_ZZ_pX_CR_element.pyx',language = 'C++', libraries = ['gmp', 'ntl', 'csage', 'gmpxx', 'm', 'stdc++']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/padic_ZZ_pX_CA_element.pyx',language = 'C++', libraries = ['gmp', 'ntl', 'csage', 'gmpxx', 'm', 'stdc++']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/padics/padic_printing.pyx',language = 'C++', libraries = ['gmp', 'ntl', 'csage', 'gmpxx', 'm', 'stdc++']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/memory.pyx',libraries = ['gmp','stdc++']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/bernoulli_mod_p.pyx',language = 'C++', include_dirs = [abs_sage_path(env,'devel/sage/libs/ntl/')], libraries=['ntl','stdc++']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/schemes/hyperelliptic_curves/hypellfrob.pyx',language = 'C++', include_dirs = [abs_sage_path(env,'devel/sage/libs/ntl/'), abs_sage_path(env,'sage/schemes/hyperelliptic_curves/hypellfrob/')], libraries = ['ntl','stdc++', 'gmp', 'zn_poly']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/polynomial/polynomial_integer_dense_ntl.pyx',language = 'C++', include_dirs = [abs_sage_path(env,'devel/sage/libs/ntl/')], libraries = ['ntl','stdc++', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/polynomial/polynomial_modn_dense_ntl.pyx',language = 'C++', include_dirs = [abs_sage_path(env,'devel/sage/libs/ntl/')], libraries = ['ntl','stdc++', 'gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/rational.pyx',libraries = ['ntl','gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/sparse_poly.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/polynomial/real_roots.pyx',libraries = ['mpfr', 'qd']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/number_field/number_field_element.pyx',language = 'C++', libraries=['ntl','gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/number_field/number_field_element_quadratic.pyx',language = 'C++', libraries=['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/number_field/totallyreal_data.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/modular/modsym/heilbronn.pyx',libraries = ['gmp','m']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/modular/modsym/p1list.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/integer_mod.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/combinat/expnums.pyx',libraries = ['gmp']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/combinat/partitions.pyx',language='C++', libraries = ['qd', 'gmp', 'mpfr']) config_gcc_file(env,pyx_pextn_dict,"devel/sage/sage/combinat/matrices/dancing_links.pyx",language="C++") config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/graphs/graph_fast.pyx',libraries = ['gmp']) #config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/symbolics/symbolicarithmetic.pyx',libraries = ['glite']) #config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/symbolics/symbolicmularithmetic.pyx',libraries = ['glite']) config_gcc_file(env,pyx_pextn_dict,'devel/sage/sage/rings/polynomial/pbori.pyx',language='C++', include_dirs = [real_sage_local(env,'include/cudd'),real_sage_local(env,'include/polybori'), real_sage_local(env,'include/polybori/groebner')], libraries=['polybori','pboriCudd','groebner']) ################################################################### #End standard configurations ################################################################### tempdir = os.path.realpath(env.options['SAGE_ROOT']) #A dictionary of gcc objects extensions gcceo_dict = { } #A dictionary of gcc shared object extensions gcceso_dict = { } #Manually create the mwrank C extension mwrankcc = GCC_extension_object(gccc, env, ["devel/sage/sage/libs/mwrank/wrap.cc"], tempdir+"/devel/sage/build/temp/sage/libs/mwrank", language='C++', define_macros=[("NTL_ALL",None)], options = { '-fPIC':None } ) #Create the Hypellfrob extensions hypellfrob_cpp = GCC_extension_object(gccc, env,["devel/sage/sage/schemes/hyperelliptic_curves/hypellfrob/hypellfrob.cpp"], tempdir+"/devel/sage/build/temp/sage/schemes/hyperelliptic_curves/hypellfrob/", language='C++', include_dirs=[abs_sage_path(env, 'sage/libs/ntl'), abs_sage_path(env,'sage/schemes/hyperelliptic_curves/hypellfrob/') ], options = { '-fPIC':None }) recurrences_zn_poly = GCC_extension_object(gccc, env,[abs_sage_path(env,"sage/schemes/hyperelliptic_curves/hypellfrob/recurrences_zn_poly.cpp")], tempdir+"/devel/sage/build/temp/sage/schemes/hyperelliptic_curves/hypellfrob/", language='C++', include_dirs=[abs_sage_path(env, 'sage/libs/ntl'), abs_sage_path(env,'sage/schemes/hyperelliptic_curves/hypellfrob/') ], options = { '-fPIC':None } ) recurrences_ntl = GCC_extension_object(gccc, env,["devel/sage/sage/schemes/hyperelliptic_curves/hypellfrob/recurrences_ntl.cpp"], tempdir+"/devel/sage/build/temp/sage/schemes/hyperelliptic_curves/hypellfrob/", language='C++', include_dirs=[abs_sage_path(env, 'sage/libs/ntl'), abs_sage_path(env,'sage/schemes/hyperelliptic_curves/hypellfrob/') ], options = { '-fPIC':None } ) #Manually create the C paritions extension partitions_c = GCC_extension_object(gccc, env,["devel/sage/sage/combinat/partitions_c.cc"], tempdir+"/devel/sage/build/temp/sage/combinat", language='C++', options = { '-fPIC':None } ) #If we don't already have a temporary build directory for planarity, make it safemkdirs(tempdir+"/devel/sage/build/temp/sage/graphs/planarity") #Locate all files that can possibley have depfw = extfilewalker() depfw.addcallback('.pyx',lambda x: True) depfw.addcallback('.pxd',lambda x: True) depfw.addcallback('.pxi',lambda x: True) depfw.addcallback('.py',lambda x: x.find(".doctest") ==-1 ) dep_list = depfw.walk('devel/sage/sage') #Set the dependency locater function funcdict = { '.pyx':get_cython_file_deps, '.pxd':get_cython_file_deps, 'pxi':get_cython_file_deps } #Set possible include directories for various file extensions includedict = { '.pyx':[abs_sage_path(env,'')], '.pxd':[abs_sage_path(env,'')], '.pxi':[abs_sage_path(env,'')] } #Set cacheinfo (used for saving the state at the end of the build) global cacheinfo cacheinfo = get_compile_list(dep_list, funcdict, includedict, "sagebuild.chc", env.options['SAGE_ROOT']) #Add the task that runs on exit to write out the build cache TM.addfinishtask(sagebuild_exit) pyx_ext_dict = { } for x in cacheinfo[0]: if os.path.splitext(x)[1]==".pyx": pyx_ext_dict[x]=pyx_pextn_dict[x] if verbose>100: print '---------------PYX_EXT_DICT------------' print pyx_ext_dict print '----------------------------------------' #If we are building planarity if pyx_ext_dict.has_key('devel/sage/sage/graphs/planarity.pyx'): #Locate all planarity C files efwplan = extfilewalker() efwplan.addcallback('.c',lambda x: True) planarity_list = efwplan.walk('devel/sage/sage/graphs/planarity') planarity_ext = list() for x in planarity_list: #Create their extensions and enable them ext = GCC_extension_object(gccc, env, [x], tempdir+"/devel/sage/build/temp/sage/graphs/planarity", options = { '-fPIC':None } ) ext.generate_action(env).enable() planarity_ext.append(ext) #If we are building mwrank.pyx, build the mwrank c extensions if pyx_ext_dict.has_key("devel/sage/sage/libs/mwrank/mwrank.pyx"): mwrankcc.generate_action(env).enable() else: pass #If we are building hypellfrob, make sure the temporary directories exist and enable the C extensions if pyx_ext_dict.has_key('devel/sage/sage/schemes/hyperelliptic_curves/hypellfrob.pyx'): try: os.makedirs(hypellfrob_cpp.get_out_dir()) except: pass hypellfrob_cpp.generate_action(env).enable() recurrences_ntl.generate_action(env).enable() recurrences_zn_poly.generate_action(env).enable() #If we are building partions.pyx, we should build the partitions C extension if pyx_ext_dict.has_key('devel/sage/sage/combinat/partitions.pyx'): partitions_c.generate_action(env).enable() #For all Cython extensions, create gcc object file extensions for x in pyx_ext_dict.values(): q = x.generate_action(env) gcc_temp_dir = tempdir + '/devel/sage/build/temp/sage/' + (os.path.split(x.sources[0])[0])[len('devel/sage/sage/'):] gcceo = GCC_extension_object(gccc, env, [x], gcc_temp_dir, options = { '-fPIC':None } ) gcceo_dict[x.sources[0]] = gcceo gccaction = gcceo.generate_action(env) gccaction.dep_register(q) q.enable() #Create temporary directories for all object file extensions and enable them for gcceo in gcceo_dict.values(): try: os.makedirs(gcceo.get_out_dir()) except: pass gccaction = gcceo.generate_action(env) gccaction.enable() #For all object file extensions, create shared object file extensions and enable them for gcceo in gcceo_dict.values(): gcc_build_dir = develdir + '/build' + (os.path.split(gcceo.sources[0])[0])[len('devel/sage'):] gcceso = GCC_extension_shared_object(gccc,env, [gcceo], gcc_build_dir) gcceso_dict[gcceso.sources[0] ] = gcceso gcceso.generate_action(env).dep_register(gcceo.generate_action(env)) #Link mwrank C files if needed dict_insert_src(gcceso_dict, tempdir+"/devel/sage/build/temp/sage/libs/mwrank/mwrank.o", mwrankcc.outfile) #Setup dependencies for mwrank if needed try: gcceso_dict[tempdir + "/devel/sage/build/temp/sage/libs/mwrank/mwrank.o"].generate_action(env).dep_register(mwrankcc.generate_action(env)) except KeyError: pass #Link hypellfrob C files if needed dict_insert_src(gcceso_dict, tempdir+"/devel/sage/build/temp/sage/schemes/hyperelliptic_curves/hypellfrob.o", hypellfrob_cpp.outfile) dict_insert_src(gcceso_dict, tempdir+"/devel/sage/build/temp/sage/schemes/hyperelliptic_curves/hypellfrob.o", recurrences_ntl.outfile) dict_insert_src(gcceso_dict, tempdir+"/devel/sage/build/temp/sage/schemes/hyperelliptic_curves/hypellfrob.o", recurrences_zn_poly.outfile) #Setup dependencies for hypellfrob shared object if needed try: gcceso_dict[tempdir+"/devel/sage/build/temp/sage/schemes/hyperelliptic_curves/hypellfrob.o"].generate_action(env).dep_register(hypellfrob_cpp.generate_action(env)) gcceso_dict[tempdir+"/devel/sage/build/temp/sage/schemes/hyperelliptic_curves/hypellfrob.o"].generate_action(env).dep_register(recurrences_ntl.generate_action(env)) gcceso_dict[tempdir+"/devel/sage/build/temp/sage/schemes/hyperelliptic_curves/hypellfrob.o"].generate_action(env).dep_register(recurrences_zn_poly.generate_action(env)) except KeyError: pass #Link parition C file if needed dict_insert_src(gcceso_dict, tempdir+"/devel/sage/build/temp/sage/combinat/partitions.o", partitions_c.outfile) #Setup dependencies for parition if needed try: gcceso_dict[tempdir+"/devel/sage/build/temp/sage/combinat/partitions.o"].generate_action(env).dep_register(mwrankcc.generate_action(env)) except KeyError: pass #If we are building planarity, then we should link the C extensions into the planarity shared object if pyx_ext_dict.has_key('devel/sage/sage/graphs/planarity.pyx'): for x in planarity_ext: dict_insert_src(gcceso_dict, tempdir+"/devel/sage/build/temp/sage/graphs/planarity.o",x.outfile) #Make output directorys in site-packages if they don't already exist for gcceso in gcceso_dict.values(): try: os.makedirs(gcceso.get_out_dir()) except: pass gccesoaction = gcceso.generate_action(env) gccesoaction.enable() if verbose>10: print 'Copying *.py files' #Copy all python files to site-packages for filenm in cacheinfo[0]: if os.path.splitext(filenm)[1]==".py": filepart = (os.path.split(filenm)[0]) newdir = env.options['SAGE_ROOT'] + '/devel/sage/build/sage/' + filepart.replace('devel/sage/sage','') filename = os.path.split(filenm)[1] if filename[0]=='.': continue safemkdirs(newdir) cmd = 'cp %s %s/%s' % (filenm, newdir, filename) if verbose>30: print cmd os.system( cmd ) #Setup the site-packages symlink if it doesn't already exist safesymlink('../../../../devel/sage/build/sage','local/lib/python/site-packages/sage') #Handle DSage safemkdirs('local/dsage') safesymlink('../../devel/sage/sage/dsage/web', 'local/dsage/web')
5911148c031026159378916a59511ce4458c3ae2 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9417/5911148c031026159378916a59511ce4458c3ae2/sagebuild.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 10736, 410, 12, 3074, 16, 314, 952, 71, 4672, 3536, 1220, 445, 6013, 10736, 272, 410, 3536, 27435, 273, 1562, 4181, 18, 22903, 468, 1077, 1039, 279, 666, 434, 777, 2395, 92, 1390, 316, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 10736, 410, 12, 3074, 16, 314, 952, 71, 4672, 3536, 1220, 445, 6013, 10736, 272, 410, 3536, 27435, 273, 1562, 4181, 18, 22903, 468, 1077, 1039, 279, 666, 434, 777, 2395, 92, 1390, 316, ...
if e.args[0] not in (EBADF, ECONNRESET, ENOTCONN, ESHUTDOWN, ECONNABORTED):
if e.args[0] not in (EBADF, ECONNRESET, ENOTCONN, ESHUTDOWN, ECONNABORTED):
def readwrite(obj, flags): try: if flags & select.POLLIN: obj.handle_read_event() if flags & select.POLLOUT: obj.handle_write_event() if flags & select.POLLPRI: obj.handle_expt_event() if flags & (select.POLLHUP | select.POLLERR | select.POLLNVAL): obj.handle_close() except socket.error, e: if e.args[0] not in (EBADF, ECONNRESET, ENOTCONN, ESHUTDOWN,
96b79c851af6faac8b41383ab9549aaa7b84fd04 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8546/96b79c851af6faac8b41383ab9549aaa7b84fd04/asyncore.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 855, 2626, 12, 2603, 16, 2943, 4672, 775, 30, 309, 2943, 473, 2027, 18, 14232, 20663, 30, 1081, 18, 4110, 67, 896, 67, 2575, 1435, 309, 2943, 473, 2027, 18, 14232, 1502, 1693, 30, 1081...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 855, 2626, 12, 2603, 16, 2943, 4672, 775, 30, 309, 2943, 473, 2027, 18, 14232, 20663, 30, 1081, 18, 4110, 67, 896, 67, 2575, 1435, 309, 2943, 473, 2027, 18, 14232, 1502, 1693, 30, 1081...
print 'Color and hilshade must be the same size in pixels.'
print('Color and hilshade must be the same size in pixels.')
def Usage(): print("""Usage: hsv_merge.py [-q] [-of format] src_color src_greyscale dst_color
4c13748fd6384faf3cf36102a5c201c9df3d536a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10290/4c13748fd6384faf3cf36102a5c201c9df3d536a/hsv_merge.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 10858, 13332, 1172, 2932, 3660, 5357, 30, 30785, 67, 2702, 18, 2074, 23059, 85, 65, 23059, 792, 740, 65, 1705, 67, 3266, 1705, 67, 11556, 1900, 1869, 3046, 67, 3266, 225, 2, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 10858, 13332, 1172, 2932, 3660, 5357, 30, 30785, 67, 2702, 18, 2074, 23059, 85, 65, 23059, 792, 740, 65, 1705, 67, 3266, 1705, 67, 11556, 1900, 1869, 3046, 67, 3266, 225, 2, -100, -100, ...
text = self._format_raw(*message) if len(text) >= 3 and text[-3:] == "...": self._log.info(self._format(*message), extra={"continued": True}) else: self._log.info(self._format(*message))
self.log(INFO, *message)
def info(self, *message): "Write info message." text = self._format_raw(*message) if len(text) >= 3 and text[-3:] == "...": self._log.info(self._format(*message), extra={"continued": True}) else: self._log.info(self._format(*message))
981348947d3df4d70d1652b8c703ae5d671912b4 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/624/981348947d3df4d70d1652b8c703ae5d671912b4/log.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1123, 12, 2890, 16, 380, 2150, 4672, 315, 3067, 1123, 883, 1199, 977, 273, 365, 6315, 2139, 67, 1899, 30857, 2150, 13, 309, 562, 12, 955, 13, 1545, 890, 471, 977, 18919, 23, 26894, 422...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1123, 12, 2890, 16, 380, 2150, 4672, 315, 3067, 1123, 883, 1199, 977, 273, 365, 6315, 2139, 67, 1899, 30857, 2150, 13, 309, 562, 12, 955, 13, 1545, 890, 471, 977, 18919, 23, 26894, 422...
notifications = Notification.objects.filter(text=None, active=True)
notifications = Notification.objects.filter(text=None, active=True).exclude(type='own')
def notify(sender, **kwargs): from cm.security import get_viewable_comments, has_perm allready_notified = set() # avoid sending multiple notifications to same user activity = kwargs['instance'] if activity.type in Activity.VIEWABLE_ACTIVITIES.get('view_users'): # user activity: only viewed by managers notifications = Notification.objects.filter(text=None, active=True) for notification in notifications: if notification.user: from cm.security import user_has_perm # import here! if user_has_perm(notification.user, 'can_manage_workspace'): send_notification(activity, notification) allready_notified.add(notification.user) elif activity.type in Activity.VIEWABLE_ACTIVITIES.get('view_comments'): notifications = Notification.objects.filter(Q(text=activity.text) | Q(text=None), active=True) for notification in notifications: viewable = get_viewable_comments(FakeRequest(notification.user), Comment.objects.filter(id__in = [activity.comment.id]), text=activity.text) if viewable and \ ((notification.type == 'own' and activity.comment.user != notification.user and activity.comment.top_comment().user == notification.user) or (notification.type != 'own')): if not notification.user in allready_notified: send_notification(activity, notification) allready_notified.add(notification.user) elif activity.type in Activity.VIEWABLE_ACTIVITIES.get('view_texts'): notifications = Notification.objects.filter(Q(text=activity.text) | Q(text=None), active=True) for notification in notifications: if notification.user: from cm.security import user_has_perm # import here! if user_has_perm(notification.user, 'can_view_text', text=activity.text) and not notification.user in allready_notified: send_notification(activity, notification) allready_notified.add(notification.user) else: if has_perm(None, 'can_view_text', text=activity.text) and not notification.email in allready_notified: send_notification(activity, notification) allready_notified.add(notification.email)
c6d424f2da71c7fa5d6e5d7da07e915701db6516 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5622/c6d424f2da71c7fa5d6e5d7da07e915701db6516/notifications.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5066, 12, 15330, 16, 2826, 4333, 4672, 628, 5003, 18, 7462, 1930, 336, 67, 1945, 429, 67, 9231, 16, 711, 67, 12160, 777, 1672, 67, 902, 939, 273, 444, 1435, 468, 4543, 5431, 3229, 9208...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5066, 12, 15330, 16, 2826, 4333, 4672, 628, 5003, 18, 7462, 1930, 336, 67, 1945, 429, 67, 9231, 16, 711, 67, 12160, 777, 1672, 67, 902, 939, 273, 444, 1435, 468, 4543, 5431, 3229, 9208...
and "true" or "false"
and "True" or "False"
def extract_vdi(self, vbd_sxp, document): src = get_child_by_name(vbd_sxp, "uname") name = "vdi" + str(src.__hash__())
f299219ba026e05c81968d43a23b22a8ed9c3b22 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/6195/f299219ba026e05c81968d43a23b22a8ed9c3b22/xenapi_create.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2608, 67, 90, 3211, 12, 2890, 16, 331, 16410, 67, 87, 23829, 16, 1668, 4672, 1705, 273, 336, 67, 3624, 67, 1637, 67, 529, 12, 90, 16410, 67, 87, 23829, 16, 315, 318, 339, 7923, 508, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2608, 67, 90, 3211, 12, 2890, 16, 331, 16410, 67, 87, 23829, 16, 1668, 4672, 1705, 273, 336, 67, 3624, 67, 1637, 67, 529, 12, 90, 16410, 67, 87, 23829, 16, 315, 318, 339, 7923, 508, ...
else
else:
def __init__(data = None) if data == None: quickfix.IntField.__init__(self, 626) else quickfix.IntField.__init__(self, 626, data)
484890147d4b23aac4b9d0e85e84fceab7e137c3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8819/484890147d4b23aac4b9d0e85e84fceab7e137c3/quickfix_fields.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 892, 273, 599, 13, 309, 501, 422, 599, 30, 9549, 904, 18, 1702, 974, 16186, 2738, 972, 12, 2890, 16, 1666, 5558, 13, 469, 30, 9549, 904, 18, 1702, 974, 16186, 27...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 892, 273, 599, 13, 309, 501, 422, 599, 30, 9549, 904, 18, 1702, 974, 16186, 2738, 972, 12, 2890, 16, 1666, 5558, 13, 469, 30, 9549, 904, 18, 1702, 974, 16186, 27...
be overfull if : - `n` is odd - It satisfies `2m > (n-1)\Delta(G)`, where `\Delta(G)` denotes the maximal degree of a vertex in `G`
be overfull if: - `n` is odd - It satisfies `2m > (n-1)\Delta(G)`, where `\Delta(G)` denotes the maximum degree among all vertices in `G`.
def is_overfull(self): r""" Tests whether the current graph is overfull.
a9b829f661d7bc42123ff69ff5975a9a0a8b87de /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9890/a9b829f661d7bc42123ff69ff5975a9a0a8b87de/generic_graph.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 353, 67, 1643, 2854, 12, 2890, 4672, 436, 8395, 7766, 87, 2856, 326, 783, 2667, 353, 1879, 2854, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 353, 67, 1643, 2854, 12, 2890, 4672, 436, 8395, 7766, 87, 2856, 326, 783, 2667, 353, 1879, 2854, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100...
a.set_period(smallest_enclosing_interval(a.get_period(), b.get_period()))
a.set_period(smallest_enclosing_seg(a.get_period(), b.get_period()))
def smallest_enclosing_interval(seg1, seg2): return segments.segment(min(seg1[0], seg2[0]), max(seg1[1], seg2[1]))
9d65c59a41a6b862dbefed419b7f4c44c53f8260 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3592/9d65c59a41a6b862dbefed419b7f4c44c53f8260/SnglBurstUtils.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13541, 67, 1331, 10910, 67, 6624, 12, 5680, 21, 16, 2291, 22, 4672, 327, 5155, 18, 9273, 12, 1154, 12, 5680, 21, 63, 20, 6487, 2291, 22, 63, 20, 65, 3631, 943, 12, 5680, 21, 63, 21...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13541, 67, 1331, 10910, 67, 6624, 12, 5680, 21, 16, 2291, 22, 4672, 327, 5155, 18, 9273, 12, 1154, 12, 5680, 21, 63, 20, 6487, 2291, 22, 63, 20, 65, 3631, 943, 12, 5680, 21, 63, 21...
return "fake", "fake", 80
return address.IPv4Address("TCP", "fake", 80)
def getHost(self): return "fake", "fake", 80
39617348e24d0927ac2e4ee0c0adff192b69f98d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12595/39617348e24d0927ac2e4ee0c0adff192b69f98d/test_woven.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8580, 12, 2890, 4672, 327, 315, 28395, 3113, 315, 28395, 3113, 8958, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8580, 12, 2890, 4672, 327, 315, 28395, 3113, 315, 28395, 3113, 8958, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -1...
for k, v in [('from', from_), ('to', to), ('type', type)]:
for k, v in [('from', from_), ('to', to), ('type', type), ('id', id)]:
def sendIQ(self, type, from_, to, query): id = str(self.lastID) self.lastID += 1 deferred = defer.Deferred() attributes = [] for k, v in [('from', from_), ('to', to), ('type', type)]: if v: attributes.append('%s="%s"' % (k, v)) attributes = " ".join(attributes) self.transport.write('<iq %s>" % attributes) query.writexml(self.transport) self.transport.write('</iq>') self.requests[id] = deferred return id, deferred
c971304756b8307e959a2f79c9b5be5828d47a3b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12595/c971304756b8307e959a2f79c9b5be5828d47a3b/xmlstream.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1366, 45, 53, 12, 2890, 16, 618, 16, 628, 67, 16, 358, 16, 843, 4672, 612, 273, 609, 12, 2890, 18, 2722, 734, 13, 365, 18, 2722, 734, 1011, 404, 7045, 273, 2220, 18, 16886, 1435, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1366, 45, 53, 12, 2890, 16, 618, 16, 628, 67, 16, 358, 16, 843, 4672, 612, 273, 609, 12, 2890, 18, 2722, 734, 13, 365, 18, 2722, 734, 1011, 404, 7045, 273, 2220, 18, 16886, 1435, 1...
self._record_member(tarinfo)
self.members.append(tarinfo)
def next(self): """Return the next member of the archive as a TarInfo object, when TarFile is opened for reading. Return None if there is no more available. """ self._check("ra") if self.firstmember is not None: m = self.firstmember self.firstmember = None return m
d0fb0ac63c5db48d344c6e8e8dab57214f04e854 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/d0fb0ac63c5db48d344c6e8e8dab57214f04e854/tarfile.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1024, 12, 2890, 4672, 3536, 990, 326, 1024, 3140, 434, 326, 5052, 487, 279, 20334, 966, 733, 16, 1347, 20334, 812, 353, 10191, 364, 6453, 18, 2000, 599, 309, 1915, 353, 1158, 1898, 2319,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1024, 12, 2890, 4672, 3536, 990, 326, 1024, 3140, 434, 326, 5052, 487, 279, 20334, 966, 733, 16, 1347, 20334, 812, 353, 10191, 364, 6453, 18, 2000, 599, 309, 1915, 353, 1158, 1898, 2319,...
moves.append([row, col, row, col - 1, row, col - 2])
moves.append((row, col, row, col - 1, row, col - 2))
def generate_moves(self): moves = [] for row in range(len(self.board)): for col in range(len(self.board[row])):
3a3991804fa82cdaedfc119c394eac8ebdbdf688 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/11237/3a3991804fa82cdaedfc119c394eac8ebdbdf688/PegSolitaire.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2103, 67, 81, 10829, 12, 2890, 4672, 13934, 273, 5378, 364, 1027, 316, 1048, 12, 1897, 12, 2890, 18, 3752, 3719, 30, 364, 645, 316, 1048, 12, 1897, 12, 2890, 18, 3752, 63, 492, 22643, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2103, 67, 81, 10829, 12, 2890, 4672, 13934, 273, 5378, 364, 1027, 316, 1048, 12, 1897, 12, 2890, 18, 3752, 3719, 30, 364, 645, 316, 1048, 12, 1897, 12, 2890, 18, 3752, 63, 492, 22643, ...
newSize = self.baseWidth + 40 * self.stepSize
ch = self.ch1 newSize = self.baseWidth1 + 40 * self.stepSize
def OnTestResizeBoundsButton( self, event ): ch = self.ch1 if (self.stepSize == 1): self.stepDir = (-1) else: if (self.stepSize == (-1)): self.stepDir = 1 self.stepSize = self.stepSize + self.stepDir newSize = self.baseWidth + 40 * self.stepSize ch.DoSetSize( self.colStartX, self.colStartY + 20, newSize, 20, 0 ) self.l0.SetLabel( "(%d): resized bounds to %d" %(ch.GetId(), newSize) )
8e00093ae4d0bf1a005a76b9785f4cd179e7b5d1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/8e00093ae4d0bf1a005a76b9785f4cd179e7b5d1/ColumnHeader.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2755, 4709, 12182, 5694, 3616, 12, 365, 16, 871, 262, 30, 462, 273, 365, 18, 343, 21, 309, 261, 2890, 18, 4119, 1225, 422, 404, 4672, 365, 18, 4119, 1621, 273, 24927, 21, 13, 469, 30...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2755, 4709, 12182, 5694, 3616, 12, 365, 16, 871, 262, 30, 462, 273, 365, 18, 343, 21, 309, 261, 2890, 18, 4119, 1225, 422, 404, 4672, 365, 18, 4119, 1621, 273, 24927, 21, 13, 469, 30...
f, buffer, text, model = self.get_current()
f, buff, text, model = self.get_current() curr_mark = buff.get_iter_at_mark(buff.get_insert())
def file_save(self, mi=None, fname=None): if self.new: return self.file_saveas()
10449cb4d34f22fdbee6714299a5514922fc253d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2233/10449cb4d34f22fdbee6714299a5514922fc253d/edit.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 585, 67, 5688, 12, 2890, 16, 12837, 33, 7036, 16, 5299, 33, 7036, 4672, 309, 365, 18, 2704, 30, 327, 365, 18, 768, 67, 5688, 345, 1435, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 585, 67, 5688, 12, 2890, 16, 12837, 33, 7036, 16, 5299, 33, 7036, 4672, 309, 365, 18, 2704, 30, 327, 365, 18, 768, 67, 5688, 345, 1435, 2, -100, -100, -100, -100, -100, -100, -100, -...
package_data={'ropeide': ['docs/COPYING', 'docs/*.txt']},
package_data={'ropeide': ['COPYING', 'README.txt', 'docs/*.txt']},
def get_long_description(): lines = open('README.txt').read().splitlines(False) end = lines.index('Getting Started') return '\n' + '\n'.join(lines[:end]) + '\n'
74dd8c8ba93d7eb9448e116dd8f0a017c22ec0b5 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8274/74dd8c8ba93d7eb9448e116dd8f0a017c22ec0b5/setup.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 5748, 67, 3384, 13332, 2362, 273, 1696, 2668, 6949, 958, 18, 5830, 16063, 896, 7675, 4939, 3548, 12, 8381, 13, 679, 273, 2362, 18, 1615, 2668, 19213, 29386, 6134, 327, 2337, 82,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 5748, 67, 3384, 13332, 2362, 273, 1696, 2668, 6949, 958, 18, 5830, 16063, 896, 7675, 4939, 3548, 12, 8381, 13, 679, 273, 2362, 18, 1615, 2668, 19213, 29386, 6134, 327, 2337, 82,...
replace(SVNCPP_VERSION_HPP_FILE, 'SVNCPP_VERSION .*', \
replace(SVNCPP_VERSION_HPP_FILE, 'SVNCPP_VERSION .*', \
def checkSvncppVersionHpp(self):
7eedaf4bfbd858b768267ab4e35ae8a03dc989e7 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9898/7eedaf4bfbd858b768267ab4e35ae8a03dc989e7/check-version.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 866, 55, 25031, 4057, 84, 1444, 44, 11858, 12, 2890, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 866, 55, 25031, 4057, 84, 1444, 44, 11858, 12, 2890, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
'roa-rup', 'frp', 'as', 'ast', 'gn', 'av', 'ay', 'az', 'bm', 'bnj',
'roa-rup', 'frp', 'as', 'ast', 'gn', 'av', 'ay', 'az', 'bm', 'bjn',
def __init__(self): self.name = None
6e2533c20736225d3ef2c90f7062cc2bd656d77f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4404/6e2533c20736225d3ef2c90f7062cc2bd656d77f/family.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 4672, 365, 18, 529, 273, 599, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 4672, 365, 18, 529, 273, 599, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
milestone = Milestone(self.__env, name)
milestone = Milestone(self.env_open(), name)
def _do_milestone_set_completed(self, name, t): milestone = Milestone(self.__env, name) milestone.completed = self._parse_date(t) milestone.update()
8bfd000da40b8168bc7c6543227d6aa794adcb52 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9317/8bfd000da40b8168bc7c6543227d6aa794adcb52/admin.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2896, 67, 81, 18270, 67, 542, 67, 13615, 12, 2890, 16, 508, 16, 268, 4672, 28664, 273, 490, 18270, 12, 2890, 18, 3074, 67, 3190, 9334, 508, 13, 28664, 18, 13615, 273, 365, 6315, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 2896, 67, 81, 18270, 67, 542, 67, 13615, 12, 2890, 16, 508, 16, 268, 4672, 28664, 273, 490, 18270, 12, 2890, 18, 3074, 67, 3190, 9334, 508, 13, 28664, 18, 13615, 273, 365, 6315, ...
symbols : string (formula) or list of strings
symbols : str (formula) or list of str
def __init__(self, symbols=None, numbers=None, positions=None, tags=None, momenta=None, masses=None, magmoms=None, charges=None, cell=None, pbc=None, constraint=None, calculator=None): """Construct Atoms object.
4f53c765c1e0026359605da5aba1d414cddb5674 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5572/4f53c765c1e0026359605da5aba1d414cddb5674/atoms.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 7963, 33, 7036, 16, 5600, 33, 7036, 16, 6865, 33, 7036, 16, 2342, 33, 7036, 16, 10382, 69, 33, 7036, 16, 8039, 281, 33, 7036, 16, 4447, 81, 25390, 33, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 7963, 33, 7036, 16, 5600, 33, 7036, 16, 6865, 33, 7036, 16, 2342, 33, 7036, 16, 10382, 69, 33, 7036, 16, 8039, 281, 33, 7036, 16, 4447, 81, 25390, 33, ...
plt.ion() fig = plt.figure(1) plt.clf() if opts.plot_spectrum: plt.subplot(311) plt.subplot(312) plt.subplot(313)
def interp_sensor(quantity, default): try: sensor = current_dataset.enviro[quantity] except KeyError: return (lambda times: default) else: interp = scape.fitting.PiecewisePolynomial1DFit(max_degree=0) interp.fit(sensor['timestamp'], sensor['value']) return interp
335d5762a58b0097967c9e481433149498a889f6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5243/335d5762a58b0097967c9e481433149498a889f6/analyse_point_source_scans.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8174, 67, 27426, 12, 16172, 16, 805, 4672, 775, 30, 10583, 273, 783, 67, 8682, 18, 275, 522, 303, 63, 16172, 65, 1335, 4999, 30, 327, 261, 14661, 4124, 30, 805, 13, 469, 30, 8174, 27...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8174, 67, 27426, 12, 16172, 16, 805, 4672, 775, 30, 10583, 273, 783, 67, 8682, 18, 275, 522, 303, 63, 16172, 65, 1335, 4999, 30, 327, 261, 14661, 4124, 30, 805, 13, 469, 30, 8174, 27...
self.prefixlen = 32
self._prefixlen = 32
def __init__(self, ipaddr): """Instantiate a new IPv4 object.
fff47fe8f2324dff8c8e9a139755e429be20efed /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3510/fff47fe8f2324dff8c8e9a139755e429be20efed/ipaddr.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 2359, 4793, 4672, 3536, 22438, 279, 394, 7853, 24, 733, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 2359, 4793, 4672, 3536, 22438, 279, 394, 7853, 24, 733, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -1...
task_env = {}
def parse_recipe_xml(input_xml, hostname): root = minidom.parseString(input_xml) for er in root.getElementsByTagName('recipe'): system = xml_attr(er, 'system') if system == hostname: break else: for er in root.getElementsByTagName('guestrecipe'): system = xml_attr(er, 'system') if system == hostname: break else: log.info("parse_recipe_xml: No recipe for %s." % hostname) return None task_env = {} rs = xml_attr(er, 'status') if rs not in ['Running', 'Waiting']: log.info("parse_recipe_xml: This recipe has finished.") return None dict_update(task_env, ARCH=xml_attr(er, 'arch'), RECIPEID=xml_attr(er, 'id'), JOBID=xml_attr(er, 'job_id'), RECIPESETID=xml_attr(er, 'recipe_set_id'), HOSTNAME=hostname) # The following is necessary for Virtual Workflows: GUEST_ATTRS = ('system', 'mac_address', 'location', 'guestargs', 'guestname') task_env['GUESTS'] = '|'.join([ ';'.join([xml_attr(gr, a, '') for a in GUEST_ATTRS]) for gr in xml_get_nodes(er, 'guestrecipe')]) for job in root.getElementsByTagName('job'): submitter = xml_attr(job, 'owner') if submitter: task_env['SUBMITTER'] = submitter break # FIXME: This will eventually need to be replaced by sth RPM independent... repos = [] repof = '' for r in xml_get_nodes(xml_first_node(er, 'repos'), 'repo'): name = xml_attr(r, 'name') repos.append(name) repof += "[%s]\nname=beaker provided '%s' repo\nbaseurl=%s\nenabled=1\ngpgcheck=0\n\n" \ % (name, name, xml_attr(r, 'url')) task_env['BEAKER_REPOS']=':'.join(repos) task_env['RECIPE_ROLE'] = xml_attr(er, 'role') roles = {} for roles_node in xml_get_nodes(er, 'roles'): proc_roles(roles, roles_node) break test_order = 0 for task in xml_get_nodes(er, 'task'): to = xml_attr(task, 'testorder') if to is not None: test_order = int(to) else: test_order += 1 ts = xml_attr(task, 'status') if ts not in ['Waiting', 'Running']: log.debug("task id: %r status: %r", xml_attr(task, 'id'), ts) continue task_id = xml_attr(task, 'id') task_name = xml_attr(task, 'name') dict_update(task_env, TASKID=str(task_id), RECIPETESTID=str(task_id), TESTID=str(task_id), TASKNAME=task_name, ROLE=xml_attr(task, 'role')) # FIXME: Anything else to save? for p in task.getElementsByTagName('param'): task_env[xml_attr(p, 'name')]=xml_attr(p, 'value') for roles_node in xml_get_nodes(task, 'roles'): proc_roles(roles, roles_node) break for role_str in roles.keys(): task_env[role_str]=' '.join(roles[role_str]) ewd = xml_attr(task, 'avg_time') task_env['KILLTIME'] = ewd executable = '' args = [] while not executable: rpm_tags = task.getElementsByTagName('rpm') log.debug("parse_recipe_xml: rpm tag: %s", rpm_tags) if rpm_tags: rpm_name = xml_attr(rpm_tags[0], 'name') dict_update(task_env, TEST=task_name, TESTRPMNAME=normalize_rpm_name(rpm_name), TESTPATH="/mnt/tests"+task_name , KILLTIME=str(ewd)) executable = mk_rhts_task(task_env, repos, repof) args = [rpm_name] log.info("parse_recipe_xml: RPMTest %s - %s %s", rpm_name, executable, args) break exec_tags = task.getElementsByTagName('executable') log.debug("parse_recipe_xml: executable tag: %s", exec_tags) if exec_tags: if repof: f = open('/etc/yum.repos.d/beaker-tests.repo', 'w+') f.write(repof) f.close() executable = xml_attr(exec_tag[0], 'url') for arg in exec_tag[0].getElementsByTagName('arg'): args.append(xml_attr(arg, 'value')) log.info("parse_recipe_xml: ExecutableTest %s %s", executable, args) break break proto_len = executable.find(':') if proto_len >= 0: proto = executable[:proto_len] if proto == "file" and executable[proto_len+1:proto_len+3] == '//': executable = executable[proto_len+3:] else: # FIXME: retrieve a file and set an executable bit. log.warning("parse_recipe_xml: Feature not implemented yet. proto=%s", proto) continue else: executable = os.path.abspath(executable) if not executable: log.warning("parse_recipe_xml: Task %s(%s) does not have an executable associated!", task_name, task_id) continue if task_env.has_key('TESTORDER'): task_env['TESTORDER'] = str(8*int(task_env['TESTORDER']) + 4) else: task_env['TESTORDER'] = str(8*test_order) return dict(task_env=task_env, executable=executable, args=args, ewd=ewd) return None
07cdd153b03cdcfca4612380e625bd939b203923 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14755/07cdd153b03cdcfca4612380e625bd939b203923/beakerlc.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1109, 67, 3927, 3151, 67, 2902, 12, 2630, 67, 2902, 16, 5199, 4672, 225, 1365, 273, 1131, 350, 362, 18, 2670, 780, 12, 2630, 67, 2902, 13, 364, 6445, 316, 1365, 18, 588, 3471, 10401, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1109, 67, 3927, 3151, 67, 2902, 12, 2630, 67, 2902, 16, 5199, 4672, 225, 1365, 273, 1131, 350, 362, 18, 2670, 780, 12, 2630, 67, 2902, 13, 364, 6445, 316, 1365, 18, 588, 3471, 10401, ...
sys.argv.remove(_opt)
def parse_option(self, short_opt, long_opt, default = None): opt = default if short_opt in sys.argv or long_opt in sys.argv: opt_index = sys.argv.index(long_opt) if long_opt in sys.argv else sys.argv.index(short_opt)
2dbf00cae54826e56522c2d4ff7387410daafe44 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8331/2dbf00cae54826e56522c2d4ff7387410daafe44/__init__.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1109, 67, 3482, 12, 2890, 16, 3025, 67, 3838, 16, 1525, 67, 3838, 16, 805, 273, 599, 4672, 2153, 273, 805, 309, 3025, 67, 3838, 316, 2589, 18, 19485, 578, 1525, 67, 3838, 316, 2589, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1109, 67, 3482, 12, 2890, 16, 3025, 67, 3838, 16, 1525, 67, 3838, 16, 805, 273, 599, 4672, 2153, 273, 805, 309, 3025, 67, 3838, 316, 2589, 18, 19485, 578, 1525, 67, 3838, 316, 2589, ...
return None
raise CmdException, 'unknown sender details'
def __get_maintainer(): """Return the 'authname <authemail>' string as read from the configuration file """ if config.has_option('stgit', 'authname') \ and config.has_option('stgit', 'authemail'): return '%s <%s>' % (config.get('stgit', 'authname'), config.get('stgit', 'authemail')) else: return None
901288c20717c357bfd7b640fbd2e3d24d4b636b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12366/901288c20717c357bfd7b640fbd2e3d24d4b636b/mail.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 588, 67, 81, 1598, 1521, 13332, 3536, 990, 326, 296, 1944, 529, 411, 1944, 3652, 1870, 533, 487, 855, 628, 326, 1664, 585, 3536, 309, 642, 18, 5332, 67, 3482, 2668, 334, 6845, 21...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 588, 67, 81, 1598, 1521, 13332, 3536, 990, 326, 296, 1944, 529, 411, 1944, 3652, 1870, 533, 487, 855, 628, 326, 1664, 585, 3536, 309, 642, 18, 5332, 67, 3482, 2668, 334, 6845, 21...
id = random.randint(0,1000000) obj = ToBeDeleted(id) ids[obj] = 1 t[obj] = 1
id = ToBeDeleted(trandom.randint(0,1000000)) ids[id] = 1 t[id] = 1
def test_LP294788(self): # https://bugs.launchpad.net/bugs/294788 # BTree keeps some deleted objects referenced
4f378e5ca87951659d67123673df32a4c15ec19d /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10048/4f378e5ca87951659d67123673df32a4c15ec19d/testBTrees.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 14461, 5540, 9462, 5482, 12, 2890, 4672, 468, 2333, 2207, 19381, 18, 20738, 6982, 18, 2758, 19, 19381, 19, 5540, 9462, 5482, 468, 605, 2471, 20948, 2690, 4282, 2184, 8042, 2, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 14461, 5540, 9462, 5482, 12, 2890, 4672, 468, 2333, 2207, 19381, 18, 20738, 6982, 18, 2758, 19, 19381, 19, 5540, 9462, 5482, 468, 605, 2471, 20948, 2690, 4282, 2184, 8042, 2, -...
self.gridlayout36 = QtGui.QGridLayout(self.file_locations_grp) self.gridlayout36.setMargin(9) self.gridlayout36.setSpacing(6) self.gridlayout36.setObjectName("gridlayout36")
self.gridlayout32 = QtGui.QGridLayout(self.file_locations_grp) self.gridlayout32.setMargin(9) self.gridlayout32.setSpacing(6) self.gridlayout32.setObjectName("gridlayout32")
def setupUi(self, UserPrefsDialog): UserPrefsDialog.setObjectName("UserPrefsDialog") UserPrefsDialog.resize(QtCore.QSize(QtCore.QRect(0,0,917,661).size()).expandedTo(UserPrefsDialog.minimumSizeHint()))
752b48c67dcf97437d2cb83fa267b09487665da0 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/11221/752b48c67dcf97437d2cb83fa267b09487665da0/UserPrefsDialog.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3875, 13943, 12, 2890, 16, 2177, 1386, 2556, 6353, 4672, 2177, 1386, 2556, 6353, 18, 542, 16707, 2932, 1299, 1386, 2556, 6353, 7923, 2177, 1386, 2556, 6353, 18, 15169, 12, 23310, 4670, 18,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3875, 13943, 12, 2890, 16, 2177, 1386, 2556, 6353, 4672, 2177, 1386, 2556, 6353, 18, 542, 16707, 2932, 1299, 1386, 2556, 6353, 7923, 2177, 1386, 2556, 6353, 18, 15169, 12, 23310, 4670, 18,...
@param commands_in: an inpu command queue (from xmpp)
@param commands_in: an input command queue (from xmpp)
def __init__(self, config, commands_in, commands_out): """A constructor
28d219a6d7e11d121b4c3cd2d2bc0427457b1169 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/888/28d219a6d7e11d121b4c3cd2d2bc0427457b1169/xmlrpcbot.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 642, 16, 4364, 67, 267, 16, 4364, 67, 659, 4672, 3536, 37, 3885, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 642, 16, 4364, 67, 267, 16, 4364, 67, 659, 4672, 3536, 37, 3885, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
e = Variable(e)
e = Variable(e)
def get_basisfunction_deps(x): return basisfunction_deps[formdata.basisfunction_renumbering[x]]
b1b48c45bfffb609a40d14125016d63b2899726b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/624/b1b48c45bfffb609a40d14125016d63b2899726b/dependencies.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 23774, 915, 67, 14877, 12, 92, 4672, 327, 10853, 915, 67, 14877, 63, 687, 892, 18, 23774, 915, 67, 1187, 1226, 310, 63, 92, 13563, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 23774, 915, 67, 14877, 12, 92, 4672, 327, 10853, 915, 67, 14877, 63, 687, 892, 18, 23774, 915, 67, 1187, 1226, 310, 63, 92, 13563, 2, -100, -100, -100, -100, -100, -100, -100,...
'prototype': unicode(task.prototype)}
'prototype': unicode(task.prototype),}
def tasks(request): bzapi = BzAPI() items = [] status = request.GET.get('status', 'open') if status in StatusManager.requestable_for_task: tasks = getattr(Todo.tasks, status)() tasks = tasks.select_related('locale', 'project', 'batch', 'prototype') if request.GET.has_key('bug'): bugs = request.GET.getlist('bug') bugs = [int(bug) for bug in bugs] tasks = tasks.filter(bug__in=bugs) if request.GET.has_key('locale'): locales = request.GET.getlist('locale') locales = Locale.objects.filter(code__in=locales) tasks = tasks.filter(locale__in=locales) if request.GET.has_key('project'): projects = request.GET.getlist('project') projects = Project.objects.filter(slug__in=projects) tasks = tasks.filter(project__in=projects) if request.GET.has_key('batch'): batches = request.GET.getlist('batch') batches = Batch.objects.filter(slug__in=batches) tasks = tasks.filter(batch__in=batches) task_items = [] for task in tasks: task_data = {'type': 'Task', 'id': task.pk, 'pk': task.pk, 'uri': 'http://%s%s' % (request.get_host(), task.get_absolute_url()), 'label': unicode(task), 'status': task.get_status_display(), 'snapshot_ts': task.snapshot_ts_iso(), 'locale': unicode(task.locale), 'locale_code': task.locale.code, 'project': unicode(task.project), 'project_slug': task.project.slug, 'batch': unicode(task.batch) if task.batch is not None else 'Uncategorized tasks', 'prototype': unicode(task.prototype)} if task.bug is not None: task_data.update({'bug': task.bug}) if request.GET.has_key('snapshot'): bug_ts = bzapi.last_modified(task.bug) task_data.update({'snapshot': 'uptodate' if task.is_uptodate(bug_ts) else 'outofdate'}) task_items.append(task_data) items += task_items next_actions = Todo.objects.next().filter(task__in=tasks).select_related('task', 'owner') next_actions_items = [] for action in next_actions: next_actions_items.append({'type': 'Next Action', 'id': action.pk, 'label': unicode(action), 'task': action.task.id, 'owner': unicode(action.owner)}) items += next_actions_items data = {'items': items} data.update(schema) return HttpResponse(simplejson.dumps(data, indent=2), mimetype='application/javascript')
6304c7d96353aee2f79f789c4815eb4009b4a502 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14391/6304c7d96353aee2f79f789c4815eb4009b4a502/api.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4592, 12, 2293, 4672, 24788, 2425, 273, 605, 94, 2557, 1435, 1516, 273, 5378, 1267, 273, 590, 18, 3264, 18, 588, 2668, 2327, 2187, 296, 3190, 6134, 309, 1267, 316, 2685, 1318, 18, 2293, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4592, 12, 2293, 4672, 24788, 2425, 273, 605, 94, 2557, 1435, 1516, 273, 5378, 1267, 273, 590, 18, 3264, 18, 588, 2668, 2327, 2187, 296, 3190, 6134, 309, 1267, 316, 2685, 1318, 18, 2293, ...
debughelper.showParseTree(sys.stdout, bookParseTree)
parser.show(sys.stdout, bookParseTree, verbose=True)
def writeBook(self, bookParseTree, output, removedArticlesFile=None, coverimage=None, status_callback=None): if status_callback: status_callback(status=_('layouting'), progress=0) if self.debug: debughelper.showParseTree(sys.stdout, bookParseTree) pass
6cc24c372963a85d88a5f7e7375e93a533b8ccd5 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12390/6cc24c372963a85d88a5f7e7375e93a533b8ccd5/rlwriter.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1045, 9084, 12, 2890, 16, 6978, 3201, 2471, 16, 876, 16, 3723, 19001, 812, 33, 7036, 16, 5590, 2730, 33, 7036, 16, 1267, 67, 3394, 33, 7036, 4672, 225, 309, 1267, 67, 3394, 30, 1267, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1045, 9084, 12, 2890, 16, 6978, 3201, 2471, 16, 876, 16, 3723, 19001, 812, 33, 7036, 16, 5590, 2730, 33, 7036, 16, 1267, 67, 3394, 33, 7036, 4672, 225, 309, 1267, 67, 3394, 30, 1267, ...
self.assertNotEqual(-1, f.openFile.getvalue().find("C 1: 'here are some bytes'")) self.assertNotEqual(-1, f.openFile.getvalue().find("S 1: 'here are some bytes'"))
v = f.openFile.getvalue() self.assertNotEqual(-1, v.find("C 1: 'here are some bytes'"), "Expected client string not found in %r" % (v,)) self.assertNotEqual(-1, v.find("S 1: 'here are some bytes'"), "Expected server string not found in %r" % (v,))
def testThingsGetLogged(self): t = StringTransportWithDisconnection() f = TestLoggingFactory(Server(), 'test') p = f.buildProtocol(('1.2.3.4', 5678)) t.protocol = p p.makeConnection(t)
d41737c7c99667c5e279f077a259b5f6639c9efc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12595/d41737c7c99667c5e279f077a259b5f6639c9efc/test_policies.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 1315, 899, 967, 19862, 12, 2890, 4672, 268, 273, 514, 6568, 1190, 1669, 4071, 1435, 284, 273, 7766, 7735, 1733, 12, 2081, 9334, 296, 3813, 6134, 293, 273, 284, 18, 3510, 5752, 12, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 1315, 899, 967, 19862, 12, 2890, 4672, 268, 273, 514, 6568, 1190, 1669, 4071, 1435, 284, 273, 7766, 7735, 1733, 12, 2081, 9334, 296, 3813, 6134, 293, 273, 284, 18, 3510, 5752, 12, ...
norepo = ("clone init version help debugconfig debugdata"
norepo = ("clone init version help debugancestor debugconfig debugdata"
def verify(ui, repo): """verify the integrity of the repository""" return repo.verify()
2964ecda77799b7154ad8758634a884bb89753b1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11312/2964ecda77799b7154ad8758634a884bb89753b1/commands.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3929, 12, 4881, 16, 3538, 4672, 3536, 8705, 326, 24425, 434, 326, 3352, 8395, 327, 3538, 18, 8705, 1435, 225, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3929, 12, 4881, 16, 3538, 4672, 3536, 8705, 326, 24425, 434, 326, 3352, 8395, 327, 3538, 18, 8705, 1435, 225, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
raise StreamError, "seeking backwards is not allowed"
raise StreamError("seeking backwards is not allowed")
def seek(self, pos=0): """Set the stream's file pointer to pos. Negative seeking is forbidden. """ if pos - self.pos >= 0: blocks, remainder = divmod(pos - self.pos, self.bufsize) for i in xrange(blocks): self.read(self.bufsize) self.read(remainder) else: raise StreamError, "seeking backwards is not allowed" return self.pos
9d576e732e0c14bd37ecb65d8001d6f596cc8566 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/9d576e732e0c14bd37ecb65d8001d6f596cc8566/tarfile.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6520, 12, 2890, 16, 949, 33, 20, 4672, 3536, 694, 326, 1407, 1807, 585, 4407, 358, 949, 18, 30716, 6520, 310, 353, 17987, 18, 3536, 309, 949, 300, 365, 18, 917, 1545, 374, 30, 4398, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6520, 12, 2890, 16, 949, 33, 20, 4672, 3536, 694, 326, 1407, 1807, 585, 4407, 358, 949, 18, 30716, 6520, 310, 353, 17987, 18, 3536, 309, 949, 300, 365, 18, 917, 1545, 374, 30, 4398, ...
if key == 'channel': key = 'feed' if key == 'items': key = 'entries'
keymap = {'channel': 'feed', 'items': 'entries', 'guid': 'id', 'date': 'modified', 'date_parsed': 'modified_parsed'} key = keymap.get(key, key)
def __getitem__(self, key): if key == 'channel': key = 'feed' if key == 'items': key = 'entries' return UserDict.__getitem__(self, key)
99200d105c483d0627b0c87195d2d3372fe4ba91 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/99200d105c483d0627b0c87195d2d3372fe4ba91/feedparser.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 31571, 972, 12, 2890, 16, 498, 4672, 498, 1458, 273, 13666, 4327, 4278, 296, 7848, 2187, 296, 3319, 4278, 296, 8219, 2187, 296, 14066, 4278, 296, 350, 2187, 296, 712, 4278, 296, 73...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 31571, 972, 12, 2890, 16, 498, 4672, 498, 1458, 273, 13666, 4327, 4278, 296, 7848, 2187, 296, 3319, 4278, 296, 8219, 2187, 296, 14066, 4278, 296, 350, 2187, 296, 712, 4278, 296, 73...
for i in range(len(self.image_list)):
self.num_custom_actions = len(self.image_list) for i in range(self.num_custom_actions): self.curr_custom_action += 1 self.update_statusbar() while gtk.events_pending(): gtk.main_iteration()
def parse_action_command(self, command, batchmode): self.change_cursor(gtk.gdk.Cursor(gtk.gdk.WATCH)) while gtk.events_pending(): gtk.main_iteration() if batchmode == True: for i in range(len(self.image_list)): imagename = self.image_list[i] self.parse_action_command2(command, imagename) else: self.parse_action_command2(command, self.currimg_name) gc.collect() self.change_cursor(None) # Refresh the current image or any preloaded needed if they have changed: if os.path.exists(self.currimg_name) == False: self.currimg_pixbuf_original = None self.image_load_failed(False) else: animtest = gtk.gdk.PixbufAnimation(self.currimg_name) if (animtest.is_static_image() == True and self.currimg_pixbuf_original != animtest.get_static_image()) or (animtest.is_static_image() == False and self.currimg_pixbuf_original != animtest): self.load_new_image(False, False, False, True, False) if os.path.exists(self.preloadimg_prev_name) == False: self.preloadimg_prev_pixbuf_original = None else: animtest = gtk.gdk.PixbufAnimation(self.preloadimg_prev_name) if (animtest.is_static_image() == True and self.preloadimg_prev_pixbuf_original != animtest.get_static_image()) or (animtest.is_static_image() == False and self.preloadimg_prev_pixbuf_original != animtest): self.preloadimg_prev_pixbuf_original = None self.preload_when_idle = gobject.idle_add(self.preload_prev_image, False) if os.path.exists(self.preloadimg_next_name) == False: self.preloadimg_next_pixbuf_original = None else: animtest = gtk.gdk.PixbufAnimation(self.preloadimg_next_name) if (animtest.is_static_image() == True and self.preloadimg_next_pixbuf_original != animtest.get_static_image()) or (animtest.is_static_image() == False and self.preloadimg_next_pixbuf_original != animtest): self.preloadimg_next_pixbuf_original = None self.preload_when_idle = gobject.idle_add(self.preload_next_image, False)
27d2355de4fb441298280d91b3753c51d39b2607 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2291/27d2355de4fb441298280d91b3753c51d39b2607/mirage.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1109, 67, 1128, 67, 3076, 12, 2890, 16, 1296, 16, 2581, 3188, 4672, 365, 18, 3427, 67, 9216, 12, 4521, 79, 18, 75, 2883, 18, 6688, 12, 4521, 79, 18, 75, 2883, 18, 59, 5858, 3719, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1109, 67, 1128, 67, 3076, 12, 2890, 16, 1296, 16, 2581, 3188, 4672, 365, 18, 3427, 67, 9216, 12, 4521, 79, 18, 75, 2883, 18, 6688, 12, 4521, 79, 18, 75, 2883, 18, 59, 5858, 3719, 1...
for i in range(len(res)): for j in range(len(res[i])): sys.stderr.write("Line " + str(i) + " match " + str(j) + " : " + res[i][j] + "\n")
for i, val in res: for j, sub_val in val: sys.stderr.write("Line %s match %s: %s\n" % (i, j, sub_val))
def debug_dump(data, res): """Show the data the regex retrieved from a match.
03dd1fa87e7ed6a1aae10ff4e922ba7db9cd3ff8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9355/03dd1fa87e7ed6a1aae10ff4e922ba7db9cd3ff8/wiki_grabber.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1198, 67, 8481, 12, 892, 16, 400, 4672, 3536, 5706, 326, 501, 326, 3936, 10295, 628, 279, 845, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1198, 67, 8481, 12, 892, 16, 400, 4672, 3536, 5706, 326, 501, 326, 3936, 10295, 628, 279, 845, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
apply(self.tk.call, (self._w, 'selection', 'clear') + self._options(cnf, kw))
apply(self.tk.call, (self._w, 'selection', 'clear') + self._options(cnf, kw))
def selection_clear(self, cnf={}, **kw):
22710823fb554a796dc96c44885d7a9389426824 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/22710823fb554a796dc96c44885d7a9389426824/Tix.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4421, 67, 8507, 12, 2890, 16, 23118, 28793, 2826, 9987, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4421, 67, 8507, 12, 2890, 16, 23118, 28793, 2826, 9987, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10...
terminal_state = move_base_client.get_state() result = move_base_client.get_result() if terminal_state == GoalStatus.PREEMPTED: server.set_preempted(result) elif terminal_state == GoalStatus.SUCCEEDED: server.set_succeeded(result) elif terminal_state == GoalStatus.ABORTED: server.set_aborted(result) else: server.set_aborted(result, "Unknown result from move_base")
terminal_state = move_base_client.get_state() result = move_base_client.get_result() if terminal_state == GoalStatus.PREEMPTED: server.set_preempted(result) elif terminal_state == GoalStatus.SUCCEEDED: server.set_succeeded(result) elif terminal_state == GoalStatus.ABORTED: server.set_aborted(result) else: server.set_aborted(result, "Unknown result from move_base")
def execute_cb(goal): rospy.loginfo("Received a goal") if not set_tilt_profile([1.05, -.7, 1.05], [0.0, 1.8, 2.0125 + .3]): server.set_aborted(MoveBaseResult(), "Couldn't set the profile on the laser") return configure_laser() configure_head() move_base_client.send_goal(goal, None, None, feedback_cb) while not move_base_client.wait_for_result(rospy.Duration(0.1)): if server.is_preempt_requested(): if server.is_new_goal_available(): goal = server.accept_new_goal() move_base_client.send_goal(goal, None, None, feedback_cb) else: move_base_client.cancel_goal() terminal_state = move_base_client.get_state() result = move_base_client.get_result() if terminal_state == GoalStatus.PREEMPTED: server.set_preempted(result) elif terminal_state == GoalStatus.SUCCEEDED: server.set_succeeded(result) elif terminal_state == GoalStatus.ABORTED: server.set_aborted(result) else: server.set_aborted(result, "Unknown result from move_base")
122c6f7455ce682a4e43ae2b05805de7bdb5d781 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9426/122c6f7455ce682a4e43ae2b05805de7bdb5d781/pr2_move_base.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1836, 67, 7358, 12, 27354, 4672, 721, 30884, 18, 1330, 1376, 2932, 8872, 279, 17683, 7923, 309, 486, 444, 67, 29419, 67, 5040, 3816, 21, 18, 6260, 16, 225, 300, 18, 27, 16, 404, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1836, 67, 7358, 12, 27354, 4672, 721, 30884, 18, 1330, 1376, 2932, 8872, 279, 17683, 7923, 309, 486, 444, 67, 29419, 67, 5040, 3816, 21, 18, 6260, 16, 225, 300, 18, 27, 16, 404, 18, ...
raise RuntimeError("thread already started")
raise RuntimeError("threads can only be started once")
def start(self): if not self.__initialized: raise RuntimeError("thread.__init__() not called") if self.__started.is_set(): raise RuntimeError("thread already started") if __debug__: self._note("%s.start(): starting thread", self) with _active_limbo_lock: _limbo[self] = self try: _start_new_thread(self.__bootstrap, ()) except Exception: with _active_limbo_lock: del _limbo[self] raise self.__started.wait()
d94d431eb1d87f0ddc96f708fd9b5ceb4b5862c8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12029/d94d431eb1d87f0ddc96f708fd9b5ceb4b5862c8/threading.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 787, 12, 2890, 4672, 309, 486, 365, 16186, 13227, 30, 1002, 7265, 2932, 5930, 16186, 2738, 972, 1435, 486, 2566, 7923, 309, 365, 16186, 14561, 18, 291, 67, 542, 13332, 1002, 7265, 2932, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 787, 12, 2890, 4672, 309, 486, 365, 16186, 13227, 30, 1002, 7265, 2932, 5930, 16186, 2738, 972, 1435, 486, 2566, 7923, 309, 365, 16186, 14561, 18, 291, 67, 542, 13332, 1002, 7265, 2932, ...
def constant_green_generator(numchips, rgbtuple): green = rgbtuple[1]
def constant_magenta_generator(numchips, rgbtuple): red, green, blue = rgbtuple
def constant_green_generator(numchips, rgbtuple): green = rgbtuple[1] seq = constant(numchips) return map(None, seq, [green] * numchips, seq)
79a787931867fcbd6a6d6a78c25726382cc88bf0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/79a787931867fcbd6a6d6a78c25726382cc88bf0/PyncheWidget.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 225, 1652, 5381, 67, 4527, 24066, 67, 8812, 12, 2107, 343, 7146, 16, 6917, 8052, 4672, 1755, 16, 10004, 16, 8024, 273, 6917, 8052, 3833, 273, 5381, 12, 2107, 343, 7146, 13, 327, 852, 12, 703...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 225, 1652, 5381, 67, 4527, 24066, 67, 8812, 12, 2107, 343, 7146, 16, 6917, 8052, 4672, 1755, 16, 10004, 16, 8024, 273, 6917, 8052, 3833, 273, 5381, 12, 2107, 343, 7146, 13, 327, 852, 12, 703...
// = (my_z ==
// = (my_z == x[image_row][image_col][x_row][x_col]) ? my_gz : 0.0f;
def c_support_code_apply(self, node, nodename): # This code considers every position in the output z, andthen computes the gradient for the # input pixels that were downsampled to that z-position. It does so by running along every # z row (sometimes plus one, to make sure every gx row gets totally filled), and by # running along every x col. This code is not sensitive to the ignore_border flag along # the row dimension (since it runs for every position in the output z), but it is sensitive # along the col dimension. ignore_border = int(self.ignore_border) return """ template<int ds0, int ds1> // ds0 is the downsampling factor in rows, ds1 in columns __global__ void kDownsampleMaxGrad_%(nodename)s( int D0, int D1, int D2, int D3, int xD2, int xD3, const float * x, int xS0, int xS1, int xS2, int xS3, const float * z, int zS0, int zS1, int zS2, int zS3, const float * gz, int gzS0, int gzS1, int gzS2, int gzS3, float *gx) { // D0: number of image rows // D1: number of image cols // D2: number of z rows // D3: number of z cols // xD2: number of x rows // xD3: number of x cols // various .S. variables are strides float cur_max, cur_x, my_z, my_gz; int i0 = blockIdx.x; // image row int i1 = 0; // image col int i2 = blockIdx.y; // row wrt z and/or gz, ranges from 0 to D2 - 1 OR D2 (as needed to cover all x rows) int x_col = threadIdx.x; // col wrt x, ranges from 0 to xD3 - 1 int z_col = x_col/ds1; // z_col corresponding to this x_col
71853b5ef94c55df77674f33b21ba25ca9999ced /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12438/71853b5ef94c55df77674f33b21ba25ca9999ced/blas.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 276, 67, 13261, 67, 710, 67, 9010, 12, 2890, 16, 756, 16, 14003, 1069, 4672, 468, 1220, 981, 1959, 24290, 3614, 1754, 316, 326, 876, 998, 16, 471, 15991, 15881, 326, 10292, 364, 326, 4...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 276, 67, 13261, 67, 710, 67, 9010, 12, 2890, 16, 756, 16, 14003, 1069, 4672, 468, 1220, 981, 1959, 24290, 3614, 1754, 316, 326, 876, 998, 16, 471, 15991, 15881, 326, 10292, 364, 326, 4...
- the *angular position* or *azimuth* (``phi``), - and the *height* or *altitude* (``z``).
- the *height* or *altitude* (``height``) above the `xy`-plane
def gen_transform(self, r=None, theta=None, phi=None): """ EXAMPLE:: sage: T = Spherical('r', ['theta', 'phi']) sage: T.gen_transform(r=var('r'), theta=var('theta'), phi=var('phi')) (r*sin(theta)*cos(phi), r*sin(phi)*sin(theta), r*cos(theta)) """ return (r * sin(theta) * cos(phi), r * sin(theta) * sin(phi), r * cos(theta))
f1a033cac620bf133b1f996c6db511753cab6f3d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9890/f1a033cac620bf133b1f996c6db511753cab6f3d/plot3d.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3157, 67, 6547, 12, 2890, 16, 436, 33, 7036, 16, 7338, 33, 7036, 16, 7706, 33, 7036, 4672, 3536, 5675, 21373, 2866, 225, 272, 410, 30, 399, 273, 348, 21570, 2668, 86, 2187, 10228, 1448...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3157, 67, 6547, 12, 2890, 16, 436, 33, 7036, 16, 7338, 33, 7036, 16, 7706, 33, 7036, 4672, 3536, 5675, 21373, 2866, 225, 272, 410, 30, 399, 273, 348, 21570, 2668, 86, 2187, 10228, 1448...
decoded = self._decoder.decode(input_chunk, eof) self._snapshot = (dec_flags, dec_buffer + input_chunk, len(decoded)) return (input_chunk, decoded) def _encode_tell_cookie(self, position, dec_flags=0,
self._decoded_text = self._decoder.decode(input_chunk, eof) self._decoded_text_offset = 0 self._snapshot = (dec_flags, dec_buffer + input_chunk) return not eof def _pack_cookie(self, position, dec_flags=0,
def _read_chunk(self): """ Read and decode the next chunk of data from the BufferedReader.
2e00b1ed3eaf2ada4c0cb16b188516bd10ad5f0c /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8125/2e00b1ed3eaf2ada4c0cb16b188516bd10ad5f0c/io.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 896, 67, 6551, 12, 2890, 4672, 3536, 2720, 471, 2495, 326, 1024, 2441, 434, 501, 628, 326, 10633, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 896, 67, 6551, 12, 2890, 4672, 3536, 2720, 471, 2495, 326, 1024, 2441, 434, 501, 628, 326, 10633, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100...
def _precompute(self, how_many=150): try: self._b n = self._n except AttributeError: self._b = [] n = 1 self._n = n self._b += [i for i in range(self._n, self._n+how_many) if self.is_powerful(i)] self._n += how_many def _eval(self, n): try: return self._b[n-1] except (AttributeError, IndexError): self._precompute() return self._eval(n)
def _compute_at_least(self, k): r"""Ensure that len(self._b) >= k""" while len(self._b) <= k: if self.is_powerful(self._n): self._b.append(self._n) self._n += 1 def _eval(self, n): self._compute_at_least(n) return self._b[n-1]
def _precompute(self, how_many=150): try: self._b n = self._n except AttributeError: self._b = [] n = 1 self._n = n self._b += [i for i in range(self._n, self._n+how_many) if self.is_powerful(i)] self._n += how_many
d16ad81d36f45de842217962eab38a2bf30ddcdc /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9890/d16ad81d36f45de842217962eab38a2bf30ddcdc/sloane_functions.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 1484, 9200, 12, 2890, 16, 3661, 67, 9353, 33, 23014, 4672, 775, 30, 365, 6315, 70, 290, 273, 365, 6315, 82, 1335, 6394, 30, 365, 6315, 70, 273, 5378, 290, 273, 404, 365, 6315, 8...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 1484, 9200, 12, 2890, 16, 3661, 67, 9353, 33, 23014, 4672, 775, 30, 365, 6315, 70, 290, 273, 365, 6315, 82, 1335, 6394, 30, 365, 6315, 70, 273, 5378, 290, 273, 404, 365, 6315, 8...
'<table border="0" cellpadding="3" width="100%" class="status_table">',
'<table class="status_table">',
def html_completed_job_table(self, job_list): """QueuePage """ completed_list = [] for jdict in job_list: if jdict.get("state") in ["completed", "success", "errors", # completed w/errors "warnings", # completed w/warnings "killed", "died", "syserror", "defunct"]: completed_list.append(jdict)
ebe7911aaa6579796c3b771f64c91ad7ae981d9b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10674/ebe7911aaa6579796c3b771f64c91ad7ae981d9b/webtlsmd.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1729, 67, 13615, 67, 4688, 67, 2121, 12, 2890, 16, 1719, 67, 1098, 4672, 3536, 3183, 1964, 3536, 5951, 67, 1098, 273, 5378, 364, 525, 1576, 316, 1719, 67, 1098, 30, 309, 525, 1576, 18,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1729, 67, 13615, 67, 4688, 67, 2121, 12, 2890, 16, 1719, 67, 1098, 4672, 3536, 3183, 1964, 3536, 5951, 67, 1098, 273, 5378, 364, 525, 1576, 316, 1719, 67, 1098, 30, 309, 525, 1576, 18,...
zip([collator(x) for x in other.sortkey],
zip([collator(x) for x in other.sortkey if isinstance(x, basestring)],
def __cmp__(self, other): result = cmp(zip([collator(x) for x in self.sortkey], [collator(x.textContent) for x in self.key], self.key), zip([collator(x) for x in other.sortkey], [collator(x.textContent) for x in other.key], other.key)) if result == 0 and len(self.key) != len(other.key): return cmp(len(self.key), len(other.key)) return result
19fe0d41c44bfca6371846c42bc4cbde9b9d8e23 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7864/19fe0d41c44bfca6371846c42bc4cbde9b9d8e23/Index.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 9625, 972, 12, 2890, 16, 1308, 4672, 563, 273, 9411, 12, 4450, 3816, 12910, 639, 12, 92, 13, 364, 619, 316, 365, 18, 3804, 856, 6487, 306, 12910, 639, 12, 92, 18, 955, 1350, 13...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 9625, 972, 12, 2890, 16, 1308, 4672, 563, 273, 9411, 12, 4450, 3816, 12910, 639, 12, 92, 13, 364, 619, 316, 365, 18, 3804, 856, 6487, 306, 12910, 639, 12, 92, 18, 955, 1350, 13...
title = '<h1 class="title">%(title)s</h1>\n' % self.headers body = [title,] + self.docinfo + self.body
body = self.docinfo + self.body + self.body_suffix[:-1]
def astext(self): headers = ''.join(['%s: %s\n' % (k,v) \ for (k,v) in self.headers.items()]) title = '<h1 class="title">%(title)s</h1>\n' % self.headers body = [title,] + self.docinfo + self.body
04ba933bd759d6ca12ff70594c7413f2f4b579a3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5620/04ba933bd759d6ca12ff70594c7413f2f4b579a3/hthtml.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3364, 408, 12, 2890, 4672, 1607, 273, 875, 18, 5701, 12, 3292, 9, 87, 30, 738, 87, 64, 82, 11, 738, 261, 79, 16, 90, 13, 521, 364, 261, 79, 16, 90, 13, 316, 365, 18, 2485, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3364, 408, 12, 2890, 4672, 1607, 273, 875, 18, 5701, 12, 3292, 9, 87, 30, 738, 87, 64, 82, 11, 738, 261, 79, 16, 90, 13, 521, 364, 261, 79, 16, 90, 13, 316, 365, 18, 2485, 18, ...
data = "" while len(data) < bytes: d = self.recv(bytes-len(data)) if not d: raise GeneralProxyError("connection closed unexpectedly")
data = bytes("") while len(data) < count: d = self.recv(count-len(data)).decode() if not d: raise GeneralProxyError((0,"connection closed unexpectedly"))
def __recvall(self, bytes): """__recvall(bytes) -> data Receive EXACTLY the number of bytes requested from the socket. Blocks until the required number of bytes have been received. """ data = "" while len(data) < bytes: d = self.recv(bytes-len(data)) if not d: raise GeneralProxyError("connection closed unexpectedly") data = data + d return data
b2ddf409815564ff2aef6a6dcddee52f011d5b06 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9386/b2ddf409815564ff2aef6a6dcddee52f011d5b06/socks.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 18334, 454, 12, 2890, 16, 1731, 4672, 3536, 972, 18334, 454, 12, 3890, 13, 317, 501, 17046, 5675, 6526, 7076, 326, 1300, 434, 1731, 3764, 628, 326, 2987, 18, 22713, 3180, 326, 1931...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 18334, 454, 12, 2890, 16, 1731, 4672, 3536, 972, 18334, 454, 12, 3890, 13, 317, 501, 17046, 5675, 6526, 7076, 326, 1300, 434, 1731, 3764, 628, 326, 2987, 18, 22713, 3180, 326, 1931...
server = Server(sys.stdin, sys.stdout, {})
server = Server(tpool.Proxy(sys.stdin), tpool.Proxy(sys.stdout), {})
def main(): import optparse parser = optparse.OptionParser( usage="usage: %prog [options]", description="Simple saranwrap.Server wrapper") parser.add_option( '-c', '--child', default=False, action='store_true', help='Wrap an object serialized via setattr.') parser.add_option( '-m', '--module', type='string', dest='module', default=None, help='a module to load and export.') parser.add_option( '-l', '--logfile', type='string', dest='logfile', default=None, help='file to log to.') options, args = parser.parse_args() global _g_logfile if options.logfile: _g_logfile = open(options.logfile, 'a') if options.module: export = api.named(options.module) server = Server(sys.stdin, sys.stdout, export) elif options.child: server = Server(sys.stdin, sys.stdout, {}) # *HACK: some modules may emit on stderr, which breaks everything. class NullSTDOut(object): def noop(*args): pass write = noop read = noop flush = noop sys.stderr = NullSTDOut() sys.stdout = NullSTDOut() # Loop until EOF server.loop() if _g_logfile: _g_logfile.close()
d5f17df58ca45e68ce1ad548f1d0ed25f44eae9b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/10078/d5f17df58ca45e68ce1ad548f1d0ed25f44eae9b/saranwrap.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 1930, 2153, 2670, 2082, 273, 2153, 2670, 18, 1895, 2678, 12, 4084, 1546, 9167, 30, 738, 14654, 306, 2116, 65, 3113, 2477, 1546, 5784, 272, 297, 304, 4113, 18, 2081, 4053, 79...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 1930, 2153, 2670, 2082, 273, 2153, 2670, 18, 1895, 2678, 12, 4084, 1546, 9167, 30, 738, 14654, 306, 2116, 65, 3113, 2477, 1546, 5784, 272, 297, 304, 4113, 18, 2081, 4053, 79...
cmd = 'find %s -name %s | sed "s if DEBUG: print ' ', cmd PKGS = os.popen(cmd)
for subdir in ['Base', 'Interfaces', 'Components', 'Systems']: cmd = 'find %s/%s -name %s | sed "s if DEBUG: print ' ', cmd PKGS += os.popen(cmd).readlines()
def CopyDocs(doc_type, doc_loc, find_term): ''' The ASKAPsoft system uses different type of documentation depending on the language used. e.g. java -> javadocs, python -> sphinx, C++ -> doxygen This is the "doc_type" parameter. The documentation is produced in different locations depending on the type. This is the "doc_loc" parameter. To double check we are getting the correct documentation (and avoid any third party documentation) search for a particular file. This is the "find_term" parameter. This function does two things. 1. Creates a document type specific index file that is copied to the remote machine in the top level document type directory. e.g. cmpt/html/[doxygen|javadocs|sphinx] This contains an unordered list <ul> which points to the subdirectories containing documentation. 2. Copies the ASKAPsoft documentation to the remote machine rooted in the top level document type directory, keeping the relative path from the build system down to the package level i.e. the 'doc_loc' is stripped from the end of the path. ''' indent = 0 partial_paths = set() index = INDEX_HEAD % (doc_type.capitalize(), doc_type) index += TAB*2 + "<ul>\n" # Start indented under '<h3> Packages </h3>' # PKGS is a file like object which contains a list of paths to the # package level which should contain documentation of specified type. e.g. # /tmp/ASKAPsoft/Code/Base/py-loghandlers # /tmp/ASKAPsoft/Code/Base/py-parset sed_term = "/%s/%s" % (doc_loc, find_term) cmd = 'find %s -name %s | sed "s#%s##"' % (CODE, find_term, sed_term) if DEBUG: print ' ', cmd PKGS = os.popen(cmd) for pkgpath in PKGS: pkgpath = pkgpath.strip() # Needed for next line to work correctly. pkg = pkgpath.split('/')[-1] # Last bit is package name. subtree = pkgpath.split('/Code/')[-1] # Only stuff to right of Code. source = "%s/%s/" % (pkgpath, doc_loc) # Recreate the full path. target = "%s/%s/%s" % (TARGET, doc_type, subtree) print "info: %s" % pkg if os.path.isdir(source): # It should be a directory. old_indent = indent index += BuildTree(subtree, partial_paths, old_indent) indent = len(subtree.split('/')) + 1 if old_indent > indent: # have moved back up one or more levels. index += TAB*old_indent + "</ul>\n" index += TAB*indent + \ '<li><a href="%s"> %s </a></li>\n' % (subtree, pkg) if DEBUG: print " src = %s" % source print " dst = %s" % target else: os.system("%s && ssh %s mkdir -p %s; %s" % (PRE_SSH, RHOST, target, POST_SSH)) os.system("%s && rsync -av --delete %s %s:%s; %s" % (PRE_SSH, source, RHOST, target, POST_SSH)) # Finish off all outstanding unordered lists <ul>. # The number outstanding can be determined from the current indent level # remember the default indent level for the list is 2. for i in reversed(range(2, indent+1)): index += TAB*i + "</ul>\n" # Add the tail to the index file and write out so that it can # be scp to remote destination. index += INDEX_TAIL % (today.strftime("%d %b %Y"), today.year) f = open('/tmp/index.html', 'w') f.write(index) f.close() if DEBUG: print '\n\n', index else: os.system("%s && scp /tmp/index.html %s:%s/%s; %s" % (PRE_SSH, RHOST, TARGET, doc_type, POST_SSH))
b352652236757cb64a0f6a3ab0f66592b18a8bc5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6216/b352652236757cb64a0f6a3ab0f66592b18a8bc5/generate-embedded-docs.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5631, 12656, 12, 2434, 67, 723, 16, 997, 67, 1829, 16, 1104, 67, 6408, 4672, 9163, 1021, 5355, 47, 2203, 9269, 2619, 4692, 3775, 618, 434, 7323, 8353, 603, 326, 2653, 1399, 18, 425, 18...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5631, 12656, 12, 2434, 67, 723, 16, 997, 67, 1829, 16, 1104, 67, 6408, 4672, 9163, 1021, 5355, 47, 2203, 9269, 2619, 4692, 3775, 618, 434, 7323, 8353, 603, 326, 2653, 1399, 18, 425, 18...
option1.set_active(opt_conv["boolean"][self.param.val('basic_hiragana')]) table.attach(option1,1,2,0,1)
option1.set_active(opt_conv["boolean"][self.param.val( 'basic_hiragana')]) table.attach(option1, 1, 2, 0, 1)
def validedChanges(*args): """Check for a least one selected kana portion (display of a message if not the case), catch parameters, then close the window.
64cfb9e0b60a3a976c72fa9d5d722987641133b9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3073/64cfb9e0b60a3a976c72fa9d5d722987641133b9/gtk_gui.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1244, 13898, 7173, 30857, 1968, 4672, 3536, 1564, 364, 279, 4520, 1245, 3170, 417, 13848, 14769, 261, 5417, 434, 279, 883, 309, 486, 326, 648, 3631, 1044, 1472, 16, 1508, 1746, 326, 2742, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1244, 13898, 7173, 30857, 1968, 4672, 3536, 1564, 364, 279, 4520, 1245, 3170, 417, 13848, 14769, 261, 5417, 434, 279, 883, 309, 486, 326, 648, 3631, 1044, 1472, 16, 1508, 1746, 326, 2742, ...
labelPath = [ x[1].trailingLabel() for x in affinityTroves ]
flavorDict = {} for afTrove in affinityTroves: afVersion, afFlavor = afTrove[1], afTrove[2] if afVersion.isOnLocalHost(): self._addLocalTrove(troveTup) self.map[name] = troveTup else: flavor = troveTup[2] if flavor is None: flavorList = self.overrideFlavors(afFlavor) else: flavorList = self.overrideFlavors(flavor) flavorDict.setdefault(afVersion.trailingLabel(), []).append(flavorList) labelPath = set(flavorDict) if labelPath: self.addQuery(troveTup, labelPath, flavorDict) return
def addQueryWithAffinity(self, troveTup, labelPath, affinityTroves): name = troveTup[0] if labelPath is None: labelPath = [ x[1].trailingLabel() for x in affinityTroves ] self.map[name] = [troveTup, labelPath]
a7b181180c87f3f2da48cfc25d9da64b76052710 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8747/a7b181180c87f3f2da48cfc25d9da64b76052710/findtrove.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24087, 1190, 16310, 12, 2890, 16, 23432, 537, 56, 416, 16, 1433, 743, 16, 21775, 56, 303, 3324, 4672, 508, 273, 23432, 537, 56, 416, 63, 20, 65, 309, 1433, 743, 353, 599, 30, 19496, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24087, 1190, 16310, 12, 2890, 16, 23432, 537, 56, 416, 16, 1433, 743, 16, 21775, 56, 303, 3324, 4672, 508, 273, 23432, 537, 56, 416, 63, 20, 65, 309, 1433, 743, 353, 599, 30, 19496, ...
def __init__(data = None)
def __init__(data = None):
def __init__(data = None) if data == None: quickfix.StringField.__init__(self, 171) else quickfix.StringField.__init__(self, 171, data)
484890147d4b23aac4b9d0e85e84fceab7e137c3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8819/484890147d4b23aac4b9d0e85e84fceab7e137c3/quickfix_fields.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 892, 273, 599, 4672, 309, 501, 422, 599, 30, 9549, 904, 18, 780, 974, 16186, 2738, 972, 12, 2890, 16, 8043, 21, 13, 469, 9549, 904, 18, 780, 974, 16186, 2738, 97...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 892, 273, 599, 4672, 309, 501, 422, 599, 30, 9549, 904, 18, 780, 974, 16186, 2738, 972, 12, 2890, 16, 8043, 21, 13, 469, 9549, 904, 18, 780, 974, 16186, 2738, 97...
hppfiles += ' '+os.path.join(getgx().libdir, 're.hpp')
hppfiles += ' '+connect_paths(getgx().libdir, 're.hpp')
def generate_code(): print '[generating c++ code..]' ident = getgx().main_module.ident if sys.platform == 'win32': pyver = '%d%d' % sys.version_info[:2] else: pyver = sysconfig.get_config_var('VERSION') includes = '-I' + sysconfig.get_python_inc() + ' ' + \ '-I' + sysconfig.get_python_inc(plat_specific=True) ldflags = sysconfig.get_config_var('LIBS') + ' ' + \ sysconfig.get_config_var('SYSLIBS') + ' ' + \ '-lpython'+pyver if not sysconfig.get_config_var('Py_ENABLE_SHARED'): ldflags += ' -L' + sysconfig.get_config_var('LIBPL') if getgx().extension_module: if sys.platform == 'win32': ident += '.pyd' else: ident += '.so' # --- generate C++ files mods = getgx().modules.values() for module in mods: if not module.builtin: # create output directory if necessary if getgx().output_dir: output_dir = os.path.join(getgx().output_dir, module.dir) if not os.path.exists(output_dir): os.makedirs(output_dir) gv = generateVisitor(module) mv = module.mv setmv(mv) gv.func_pointers(False) walk(module.ast, gv) gv.out.close() gv.header_file() gv.out.close() gv.insert_consts(declare=False) gv.insert_consts(declare=True) gv.insert_includes() # --- generate Makefile makefile = file(os.path.join(getgx().output_dir, 'Makefile'), 'w') cppfiles = ' '.join([m.filename[:-3].replace(' ', '\ ')+'.cpp' for m in mods]) hppfiles = ' '.join([m.filename[:-3].replace(' ', '\ ')+'.hpp' for m in mods]) repath = os.path.join(getgx().libdir, 're.cpp') if not repath in cppfiles: cppfiles += ' '+repath hppfiles += ' '+os.path.join(getgx().libdir, 're.hpp') # import flags if getgx().flags: flags = getgx().flags elif os.path.isfile('FLAGS'): flags = 'FLAGS' else: flags = connect_paths(getgx().sysdir, 'FLAGS') for line in file(flags): line = line[:-1] if line[:line.find('=')].strip() == 'CCFLAGS': line += ' -I. -I'+getgx().libdir.replace(' ', '\ ') if sys.platform == 'darwin' and os.path.isdir('/opt/local/include'): line += ' -I/opt/local/include' # macports... and fink? if not getgx().wrap_around_check: line += ' -DNOWRAP' if getgx().bounds_checking: line += ' -DBOUNDS' if getgx().extension_module: if sys.platform == 'win32': line += ' -Ic:/Python%s/include -D__SS_BIND' % pyver else: line += ' -g -fPIC -D__SS_BIND ' + includes elif line[:line.find('=')].strip() == 'LFLAGS': if sys.platform == 'darwin' and os.path.isdir('/opt/local/lib'): line += ' -L/opt/local/lib' if getgx().extension_module: if sys.platform == 'win32': line += ' -shared -Lc:/Python%s/libs -lpython%s' % (pyver, pyver) elif sys.platform == 'darwin': line += ' -bundle -Xlinker -dynamic ' + ldflags elif sys.platform == 'sunos5': line += ' -shared -Xlinker ' + ldflags else: line += ' -shared -Xlinker -export-dynamic ' + ldflags if 'socket' in [m.ident for m in mods]: if sys.platform == 'win32': line += ' -lws2_32' elif sys.platform == 'sunos5': line += ' -lsocket -lnsl' print >>makefile, line print >>makefile print >>makefile, 'all:\t'+ident+'\n' if not getgx().extension_module: print >>makefile, 'run:\tall' print >>makefile, '\t./'+ident+'\n' print >>makefile, 'full:' print >>makefile, '\tshedskin '+ident+'; $(MAKE) run\n' print >>makefile, 'CPPFILES='+cppfiles print >>makefile, 'HPPFILES='+hppfiles+'\n' print >>makefile, ident+':\t$(CPPFILES) $(HPPFILES)' print >>makefile, '\t$(CC) $(CCFLAGS) $(CPPFILES) $(LFLAGS) -o '+ident+'\n' if sys.platform == 'win32': ident += '.exe' print >>makefile, 'clean:' print >>makefile, '\trm '+ident makefile.close()
fb3374fa2bc0145e06a74e93cc64c7f6b1e872a3 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/6661/fb3374fa2bc0145e06a74e93cc64c7f6b1e872a3/ss.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2103, 67, 710, 13332, 1172, 5271, 3441, 1776, 276, 9904, 981, 838, 3864, 225, 3390, 273, 336, 75, 92, 7675, 5254, 67, 2978, 18, 6392, 225, 309, 2589, 18, 9898, 422, 296, 8082, 1578, 42...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2103, 67, 710, 13332, 1172, 5271, 3441, 1776, 276, 9904, 981, 838, 3864, 225, 3390, 273, 336, 75, 92, 7675, 5254, 67, 2978, 18, 6392, 225, 309, 2589, 18, 9898, 422, 296, 8082, 1578, 42...
keys = self.badmodules.keys() keys.sort() for key in keys: if key not in self.excludes: mods = self.badmodules[key].keys()
missing, maybe = self.any_missing_maybe() if missing: print print "Missing modules:" for name in missing: mods = self.badmodules[name].keys()
def report(self): print print " %-25s %s" % ("Name", "File") print " %-25s %s" % ("----", "----") # Print modules found keys = self.modules.keys() keys.sort() for key in keys: m = self.modules[key] if m.__path__: print "P", else: print "m", print "%-25s" % key, m.__file__ or ""
e29310a2b3e50958b00c0e92b142992599cf7ef1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/e29310a2b3e50958b00c0e92b142992599cf7ef1/modulefinder.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2605, 12, 2890, 4672, 1172, 1172, 315, 225, 738, 17, 2947, 87, 738, 87, 6, 738, 7566, 461, 3113, 315, 812, 7923, 1172, 315, 225, 738, 17, 2947, 87, 738, 87, 6, 738, 7566, 553, 3113, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2605, 12, 2890, 4672, 1172, 1172, 315, 225, 738, 17, 2947, 87, 738, 87, 6, 738, 7566, 461, 3113, 315, 812, 7923, 1172, 315, 225, 738, 17, 2947, 87, 738, 87, 6, 738, 7566, 553, 3113, ...
self.m = is_constr.search(line)
self.m = is_constr_factory.search(line)
def __init__(self, line=None, comp_name='', class_name='', params=None): self.comp_name = comp_name self.class_name = class_name if params is None: self.params = {} else: self.params = params self.factory = None
102b791569b78fa95f7f4bd8ae4158ca0037b827 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4325/102b791569b78fa95f7f4bd8ae4158ca0037b827/methodparse.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 980, 33, 7036, 16, 1161, 67, 529, 2218, 2187, 667, 67, 529, 2218, 2187, 859, 33, 7036, 4672, 365, 18, 2919, 67, 529, 273, 1161, 67, 529, 365, 18, 1106,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 980, 33, 7036, 16, 1161, 67, 529, 2218, 2187, 667, 67, 529, 2218, 2187, 859, 33, 7036, 4672, 365, 18, 2919, 67, 529, 273, 1161, 67, 529, 365, 18, 1106,...
def __init__(data = None)
def __init__(data = None):
def __init__(data = None) if data == None: quickfix.StringField.__init__(self, 823) else quickfix.StringField.__init__(self, 823, data)
484890147d4b23aac4b9d0e85e84fceab7e137c3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8819/484890147d4b23aac4b9d0e85e84fceab7e137c3/quickfix_fields.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 892, 273, 599, 4672, 309, 501, 422, 599, 30, 9549, 904, 18, 780, 974, 16186, 2738, 972, 12, 2890, 16, 1725, 4366, 13, 469, 9549, 904, 18, 780, 974, 16186, 2738, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 892, 273, 599, 4672, 309, 501, 422, 599, 30, 9549, 904, 18, 780, 974, 16186, 2738, 972, 12, 2890, 16, 1725, 4366, 13, 469, 9549, 904, 18, 780, 974, 16186, 2738, ...
def tearDown(self): GettextBaseTest.tearDown(self)
def tearDown(self): GettextBaseTest.tearDown(self)
7a617a375bff854094c885c20f14745e1df264f0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/7a617a375bff854094c885c20f14745e1df264f0/test_gettext.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 268, 2091, 4164, 12, 2890, 4672, 968, 955, 2171, 4709, 18, 736, 297, 4164, 12, 2890, 13, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 268, 2091, 4164, 12, 2890, 4672, 968, 955, 2171, 4709, 18, 736, 297, 4164, 12, 2890, 13, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100...
if (self.prefix is not None
if (self.value_str is not None
def rows(self, size, focus=False): if (self.prefix is not None and self.var_label is not None and len(self.prefix) + len(self.var_label) > self.SIZE_LIMIT): return 2 else: return 1
0f9e0509e327cd003cf6b5a39bf27b5e8c5db0e4 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12738/0f9e0509e327cd003cf6b5a39bf27b5e8c5db0e4/debugger.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2595, 12, 2890, 16, 963, 16, 7155, 33, 8381, 4672, 309, 261, 2890, 18, 1132, 67, 701, 353, 486, 599, 471, 365, 18, 1401, 67, 1925, 353, 486, 599, 471, 562, 12, 2890, 18, 3239, 13, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2595, 12, 2890, 16, 963, 16, 7155, 33, 8381, 4672, 309, 261, 2890, 18, 1132, 67, 701, 353, 486, 599, 471, 365, 18, 1401, 67, 1925, 353, 486, 599, 471, 562, 12, 2890, 18, 3239, 13, ...
"""Builds a new Structure object by loading a mmCIF file.
"""Builds a new Structure object by loading an mmCIF file.
def setmapf_cif(smap, skey, dmap, dkey): """For float converisons, treat [?.] as blank. """ if smap.has_key_lower(skey): x = smap.getitem_lower(skey) if x in ('', '?', '.'): return False try: dmap[dkey] = float(x) except ValueError: return False return True return False
f4c16a9f76d0ee479e531eb524c3a7a6365d695e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10674/f4c16a9f76d0ee479e531eb524c3a7a6365d695e/mmCIFBuilder.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 444, 1458, 74, 67, 30674, 12, 87, 1458, 16, 272, 856, 16, 302, 1458, 16, 302, 856, 4672, 3536, 1290, 1431, 14581, 4530, 87, 16, 10116, 306, 35, 18, 65, 487, 7052, 18, 3536, 309, 272,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 444, 1458, 74, 67, 30674, 12, 87, 1458, 16, 272, 856, 16, 302, 1458, 16, 302, 856, 4672, 3536, 1290, 1431, 14581, 4530, 87, 16, 10116, 306, 35, 18, 65, 487, 7052, 18, 3536, 309, 272,...
'linker_so' : ["sage_fortran","-shared"],
'linker_so' : link_command,
def get_libraries(self): opt = GnuFCompiler.get_libraries(self) if sys.platform == 'darwin': opt.remove('cc_dynamic') return opt
705f9e3436b6b583afa5cd12c531540d3e317bf7 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9417/705f9e3436b6b583afa5cd12c531540d3e317bf7/gnu.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 31417, 12, 2890, 4672, 2153, 273, 611, 13053, 42, 9213, 18, 588, 67, 31417, 12, 2890, 13, 309, 2589, 18, 9898, 422, 296, 28076, 4278, 2153, 18, 4479, 2668, 952, 67, 14507, 613...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 31417, 12, 2890, 4672, 2153, 273, 611, 13053, 42, 9213, 18, 588, 67, 31417, 12, 2890, 13, 309, 2589, 18, 9898, 422, 296, 28076, 4278, 2153, 18, 4479, 2668, 952, 67, 14507, 613...
if version:
if version is not None:
def __init__(self, path_str, metadata=None, name=None, version=None, py_version=PY_MAJOR, platform=None, distro_type = EGG_DIST
b9d4dcab4f7f11fd6a89f0e9d680d1db2362848e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/495/b9d4dcab4f7f11fd6a89f0e9d680d1db2362848e/pkg_resources.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 589, 67, 701, 16, 1982, 33, 7036, 16, 508, 33, 7036, 16, 1177, 33, 7036, 16, 2395, 67, 1589, 33, 16235, 67, 26976, 916, 16, 4072, 33, 7036, 16, 27450, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 589, 67, 701, 16, 1982, 33, 7036, 16, 508, 33, 7036, 16, 1177, 33, 7036, 16, 2395, 67, 1589, 33, 16235, 67, 26976, 916, 16, 4072, 33, 7036, 16, 27450, ...
Calendar of nctime.
Calendar of nctime.
def _check_index(indices, dates, nctime, calendar): """Return True if the time indices given correspond to the given dates, False otherwise. Parameters ---------- indices : sequence of integers Positive integers indexing the time variable. dates : sequence of datetime objects Reference dates. nctime : netCDF Variable object NetCDF time object. calendar : string Calendar of nctime. """ if (indices <0).any(): return False if (indices >= nctime.shape[0]).any(): return False t = nctime[indices] return numpy.all( num2date(t, nctime.units, calendar) == dates)
bbd2b412f6583e4d366fbba7cbbd6a40ad4b5276 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8103/bbd2b412f6583e4d366fbba7cbbd6a40ad4b5276/netcdftime.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 1893, 67, 1615, 12, 6836, 16, 7811, 16, 290, 21261, 16, 5686, 4672, 3536, 990, 1053, 309, 326, 813, 4295, 864, 4325, 358, 326, 864, 7811, 16, 1083, 3541, 18, 225, 7012, 12181, 429...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 1893, 67, 1615, 12, 6836, 16, 7811, 16, 290, 21261, 16, 5686, 4672, 3536, 990, 1053, 309, 326, 813, 4295, 864, 4325, 358, 326, 864, 7811, 16, 1083, 3541, 18, 225, 7012, 12181, 429...
changes = self.changed_this_week()
changes = self.get_recent_changes()
def update_pdb(self): """ I guess this is the 'most wanted' function from this module. It gets the weekly lists of new and modified pdb entries and automatically downloads the according PDB files. You can call this module as a weekly cronjob. """ changes = self.changed_this_week() new = changes[0] modified = changes[1] to_download = new + modified
09feacabd5a3d2f3244213b17b5431745b1fa01c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7167/09feacabd5a3d2f3244213b17b5431745b1fa01c/PDBList.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 67, 17414, 12, 2890, 4672, 3536, 467, 7274, 333, 353, 326, 296, 10329, 15504, 11, 445, 628, 333, 1605, 18, 2597, 5571, 326, 4860, 715, 6035, 434, 394, 471, 4358, 10892, 3222, 471, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 67, 17414, 12, 2890, 4672, 3536, 467, 7274, 333, 353, 326, 296, 10329, 15504, 11, 445, 628, 333, 1605, 18, 2597, 5571, 326, 4860, 715, 6035, 434, 394, 471, 4358, 10892, 3222, 471, ...
d = int(brush.get_actual_radius())*2
d = int(brush.get_actual_radius()*self.scale)*2
def update_cursor(self, force=False): #return # OPTIMIZE: looks like big cursors can be a major slowdown with X11 # if everything breaks down, we could only show brush shape shortly after changes if not self.window: return brush = self.doc.brush d = int(brush.get_actual_radius())*2 eraser = brush.is_eraser()
d8d574ac0828f9359eb56aa2710eed464543beb3 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7129/d8d574ac0828f9359eb56aa2710eed464543beb3/tileddrawwidget.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 67, 9216, 12, 2890, 16, 2944, 33, 8381, 4672, 468, 2463, 468, 16456, 3445, 15641, 30, 10192, 3007, 5446, 25326, 1383, 848, 506, 279, 7888, 11816, 2378, 598, 1139, 2499, 468, 282, 3...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 67, 9216, 12, 2890, 16, 2944, 33, 8381, 4672, 468, 2463, 468, 16456, 3445, 15641, 30, 10192, 3007, 5446, 25326, 1383, 848, 506, 279, 7888, 11816, 2378, 598, 1139, 2499, 468, 282, 3...
self.read_secondary_header() self.read_header_pieces()
self.entries = {} self._read_secondary_header() self._read_header_pieces() self._read_section_names() self._read_manifest() self._read_meta() self._read_drm()
def __init__(self, filename_or_stream): if hasattr(filename_or_stream, 'read'): self._stream = filename_or_stream else: self._stream = open(filename_or_stream, 'rb') if self.magic != 'ITOLITLS': raise LitError('Not a valid LIT file') if self.version != 1: raise LitError('Unknown LIT version %d'%(self.version,)) self.read_secondary_header() self.read_header_pieces()
006182e5f46ea8f5da43607ab530dff752a12d94 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9125/006182e5f46ea8f5da43607ab530dff752a12d94/reader.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 1544, 67, 280, 67, 3256, 4672, 309, 3859, 12, 3459, 67, 280, 67, 3256, 16, 296, 896, 11, 4672, 365, 6315, 3256, 273, 1544, 67, 280, 67, 3256, 469, 30, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 1544, 67, 280, 67, 3256, 4672, 309, 3859, 12, 3459, 67, 280, 67, 3256, 16, 296, 896, 11, 4672, 365, 6315, 3256, 273, 1544, 67, 280, 67, 3256, 469, 30, ...
\mu_{3,0}(n) = \left(1 + \left(\frac{-3}{p}\right)\right)
\mu_{3,0}(n) = \prod_{p|n} \left(1 + \left(\frac{-3}{p}\right)\right)
def mu30(n): r""" Return value of the arithmetic function $\mu_{3,0}(n)$, where $\mu_{3,0}(n) = 0$ if $n$ is divisible by 2 or 9, and $$ \mu_{3,0}(n) = \left(1 + \left(\frac{-3}{p}\right)\right) $$ otherwise. INPUT: n -- an integer OUTPUT: Integer EXAMPLES: sage: [sage.modular.dims.mu30(n) for n in [1..19]] [1, 0, 1, 0, 0, 0, 2, 0, 0, 0, 0, 0, 2, 0, 0, 0, 0, 0, 2] """ if n%2==0 or n%9==0: return Integer(0) return mul([1 + kronecker_symbol(-3,p) for p, _ in factor(n)])
6c2ef1c177a1dcbbeb012a2db8ad9ec74a780775 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9417/6c2ef1c177a1dcbbeb012a2db8ad9ec74a780775/dims.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4129, 5082, 12, 82, 4672, 436, 8395, 2000, 460, 434, 326, 30828, 445, 271, 64, 13297, 14147, 23, 16, 20, 97, 12, 82, 21877, 16, 1625, 271, 64, 13297, 14147, 23, 16, 20, 97, 12, 82, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4129, 5082, 12, 82, 4672, 436, 8395, 2000, 460, 434, 326, 30828, 445, 271, 64, 13297, 14147, 23, 16, 20, 97, 12, 82, 21877, 16, 1625, 271, 64, 13297, 14147, 23, 16, 20, 97, 12, 82, ...
downstreamDeps = getMethodDepsR(depsItem, variants, totalDeps.union(localDeps))
downstreamDeps = getForeignDepsR(depsItem, variants, totalDeps.union(localDeps))
def getMethodDepsR(dependencyItem, variants, totalDeps): # We don't add the in-param to the global result classId = dependencyItem.name methodId= dependencyItem.attribute
0f222f89cab0a80a54d89ddc230c9f949a2394bf /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5718/0f222f89cab0a80a54d89ddc230c9f949a2394bf/Class.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6272, 14430, 54, 12, 15896, 1180, 16, 12935, 16, 2078, 14430, 4672, 468, 1660, 2727, 1404, 527, 326, 316, 17, 891, 358, 326, 2552, 563, 31181, 273, 4904, 1180, 18, 529, 707, 548, 33, 4...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6272, 14430, 54, 12, 15896, 1180, 16, 12935, 16, 2078, 14430, 4672, 468, 1660, 2727, 1404, 527, 326, 316, 17, 891, 358, 326, 2552, 563, 31181, 273, 4904, 1180, 18, 529, 707, 548, 33, 4...
def handle_accept(self): conn, addr = self.accept()
def handle_accept(self) try: conn, addr = self.accept() except TypeError: return except socket.error, err: if err[0] != errno.ECONNABORTED: raise return else: if addr == None: return
def handle_accept(self): conn, addr = self.accept() print('Incoming connection from %s' % repr(addr), file=DEBUGSTREAM) channel = self.channel_class(self, conn, addr)
13985d25414a691b0a55a643490a552162577c97 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3187/13985d25414a691b0a55a643490a552162577c97/smtpd.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1640, 67, 9436, 12, 2890, 13, 775, 30, 1487, 16, 3091, 273, 365, 18, 9436, 1435, 1335, 3580, 30, 225, 327, 1335, 2987, 18, 1636, 16, 393, 30, 225, 309, 393, 63, 20, 65, 480, 8402, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1640, 67, 9436, 12, 2890, 13, 775, 30, 1487, 16, 3091, 273, 365, 18, 9436, 1435, 1335, 3580, 30, 225, 327, 1335, 2987, 18, 1636, 16, 393, 30, 225, 309, 393, 63, 20, 65, 480, 8402, ...
'pos':fields.boolean('Point of Sales'),
'point_of_sale':fields.boolean('Point of Sales'),
def _if_account(self, cr, uid, ids, context=None): chart = self.read(cr, uid, ids, ['charts'], context=context)[0]['charts'] self.logger.notifyChannel( 'installer', netsvc.LOG_DEBUG, 'Addon "account" selected, installing chart of accounts %s'%chart) return [chart]
76cb20b5caba757923cc7afd8e8097a87aa11017 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/76cb20b5caba757923cc7afd8e8097a87aa11017/installer.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 430, 67, 4631, 12, 2890, 16, 4422, 16, 4555, 16, 3258, 16, 819, 33, 7036, 4672, 4980, 273, 365, 18, 896, 12, 3353, 16, 4555, 16, 3258, 16, 10228, 11563, 87, 17337, 819, 33, 2472...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 430, 67, 4631, 12, 2890, 16, 4422, 16, 4555, 16, 3258, 16, 819, 33, 7036, 4672, 4980, 273, 365, 18, 896, 12, 3353, 16, 4555, 16, 3258, 16, 10228, 11563, 87, 17337, 819, 33, 2472...
SELECT id, company, ldap_server, ldap_server_port, ldap_binddn, ldap_password, ldap_filter, ldap_base, \"user\", create_user FROM res_company_ldap
SELECT id, company, ldap_server, ldap_server_port, ldap_binddn, ldap_password, ldap_filter, ldap_base, "user", create_user FROM res_company_ldap
def login(self, db, login, password): ret = super(users,self).login(db, login, password) if ret: return ret pool = pooler.get_pool(db) cr = pooler.get_db(db).cursor() action_obj = pool.get('ir.actions.actions') cr.execute(""" SELECT id, company, ldap_server, ldap_server_port, ldap_binddn, ldap_password, ldap_filter, ldap_base, \"user\", create_user FROM res_company_ldap WHERE ldap_server != '' and ldap_binddn != '' ORDER BY sequence""") for res_company_ldap in cr.dictfetchall(): try: l = ldap.open(res_company_ldap['ldap_server'], res_company_ldap['ldap_server_port']) if l.simple_bind_s(res_company_ldap['ldap_binddn'], res_company_ldap['ldap_password']): base = res_company_ldap['ldap_base'] scope = ldap.SCOPE_SUBTREE filter = filter_format(res_company_ldap['ldap_filter'], (login,)) retrieve_attributes = None result_id = l.search(base, scope, filter, retrieve_attributes) timeout = 60 result_type, result_data = l.result(result_id, timeout) if not result_data: continue if result_type == ldap.RES_SEARCH_RESULT and len(result_data) == 1: dn = result_data[0][0] name = result_data[0][1]['cn'][0] if l.bind_s(dn, passwd): l.unbind() cr.execute("SELECT id FROM res_users WHERE login=%s",(tools.ustr(login),)) res = cr.fetchone() if res: cr.close() return res[0] if not res_company_ldap['create_user']: continue action_id = action_obj.search(cr, 1, [('usage', '=', 'menu')])[0] if res_company_ldap['user']: res = self.copy(cr, 1, res_company_ldap['user'], default={'active': True}) self.write(cr, 1, res, { 'name': name, 'login': login.encode('utf-8'), 'company_id': res_company_ldap['company'], }) else: res = self.create(cr, 1, { 'name': name, 'login': login.encode('utf-8'), 'company_id': res_company_ldap['company'], 'action_id': action_id, 'menu_id': action_id, }) cr.commit() cr.close() return res l.unbind() except Exception, e: continue cr.close() return False
d5087fd9736baa7f0ff8e5421ddf76083f0cfde5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/d5087fd9736baa7f0ff8e5421ddf76083f0cfde5/users_ldap.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3925, 12, 2890, 16, 1319, 16, 3925, 16, 2201, 4672, 325, 273, 2240, 12, 5577, 16, 2890, 2934, 5819, 12, 1966, 16, 3925, 16, 2201, 13, 309, 325, 30, 327, 325, 2845, 273, 2845, 264, 18...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3925, 12, 2890, 16, 1319, 16, 3925, 16, 2201, 4672, 325, 273, 2240, 12, 5577, 16, 2890, 2934, 5819, 12, 1966, 16, 3925, 16, 2201, 13, 309, 325, 30, 327, 325, 2845, 273, 2845, 264, 18...
for dep in deps: if not dep in self.queues["add"]: dep_todo_list.append(dep) while dep_todo_list: dep = dep_todo_list.pop(0) if dep.count(">="): dep = dep.split(">=")[0] if not (dep in self.queues["add"]): done, to_do = _req_pac_check(dep, "dep") if done and not done in pacs_queues["add"]: pacs_queues["add"].append(done) for add in to_do: if not add in dep_black_list: dep_todo_list.append(add) dep_black_list.append(add)
for dep in deps: if not dep in self.queues["add"]: dep_todo_list.append(dep) while dep_todo_list: dep = dep_todo_list.pop(0) if dep.count(">="): dep = dep.split(">=")[0] if not (dep in self.queues["add"]): done, to_do = _req_pac_check(dep, "dep") if done and not done in pacs_queues["add"]: pacs_queues["add"].append(done) for add in to_do: if not add in dep_black_list: dep_todo_list.append(add) dep_black_list.append(add)
def _req_pac_check(to_check, flag): to_do = []
4180fce10fefea7e3c84f18c99c63997c3734795 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/2654/4180fce10fefea7e3c84f18c99c63997c3734795/gui.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 3658, 67, 84, 1077, 67, 1893, 12, 869, 67, 1893, 16, 2982, 4672, 358, 67, 2896, 273, 5378, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 3658, 67, 84, 1077, 67, 1893, 12, 869, 67, 1893, 16, 2982, 4672, 358, 67, 2896, 273, 5378, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
def __str__(self): return self.str()
def __repr__(self): P = self._check_valid() return P.get(self._name)
49ba0b76bb983b02b857127dba97a3bab98600be /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9417/49ba0b76bb983b02b857127dba97a3bab98600be/axiom.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 12715, 972, 12, 2890, 4672, 453, 273, 365, 6315, 1893, 67, 877, 1435, 327, 453, 18, 588, 12, 2890, 6315, 529, 13, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 12715, 972, 12, 2890, 4672, 453, 273, 365, 6315, 1893, 67, 877, 1435, 327, 453, 18, 588, 12, 2890, 6315, 529, 13, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
def generate_help (header=None):
def generate_help (self, header=None):
def generate_help (header=None): """Generate help text (a list of strings, one per suggested line of output) from the option table for this FancyGetopt object."""
d67e9c09b331fb3188be89e40f9f74c3676cf33a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/d67e9c09b331fb3188be89e40f9f74c3676cf33a/fancy_getopt.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2103, 67, 5201, 261, 2890, 16, 1446, 33, 7036, 4672, 3536, 4625, 2809, 977, 261, 69, 666, 434, 2064, 16, 1245, 1534, 22168, 980, 434, 876, 13, 628, 326, 1456, 1014, 364, 333, 478, 1251...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2103, 67, 5201, 261, 2890, 16, 1446, 33, 7036, 4672, 3536, 4625, 2809, 977, 261, 69, 666, 434, 2064, 16, 1245, 1534, 22168, 980, 434, 876, 13, 628, 326, 1456, 1014, 364, 333, 478, 1251...
ks = KeyStatistic(self.__spam_label, spam, self.__spam_label) lks.append(ks) ks = KeyStatistic(self.__spam_label, ham, self.__ham_label)
ks = KeyStatistic(self.__spam_label, spam, _('messages')) lks.append(ks) ks = KeyStatistic(self.__ham_label, ham, _('messages'))
def get_graph(self, end_date, report_days, host=None, user=None, email=None): if host or user: return None
7c8714c9665bde328f22f704337c77222195321e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/87/7c8714c9665bde328f22f704337c77222195321e/untangle_base_spam.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 4660, 12, 2890, 16, 679, 67, 712, 16, 2605, 67, 9810, 16, 1479, 33, 7036, 16, 729, 33, 7036, 16, 2699, 33, 7036, 4672, 309, 1479, 578, 729, 30, 327, 599, 2, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 4660, 12, 2890, 16, 679, 67, 712, 16, 2605, 67, 9810, 16, 1479, 33, 7036, 16, 729, 33, 7036, 16, 2699, 33, 7036, 4672, 309, 1479, 578, 729, 30, 327, 599, 2, -100, -100, -1...
OUTPUT_ENCODING = 'ISO-8859-1'
OUTPUT_ENCODING = 'UTF-8'
def extractWordsFromFile(self, filename): outputDir = mkdtemp() outputFile = os.path.join(outputDir, 'out.txt') command_args = {'input' : filename, 'output' : outputFile} cmd = self.COMMAND % command_args print "Executing %r" % cmd errcode = os.system(cmd) try: if errcode == 0: # OK parser = self.getParser() return parser.parseFile(outputFile, self.OUTPUT_ENCODING) else: raise IndexationFailure('Unable to index %r' % filename) finally: if os.path.isfile(outputFile): os.remove(outputFile) os.rmdir(outputDir)
d6147f50e21a179db930d713fe8d62df8ec10b5a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2259/d6147f50e21a179db930d713fe8d62df8ec10b5a/converter.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2608, 7363, 13087, 12, 2890, 16, 1544, 4672, 21412, 273, 31381, 5814, 1435, 15047, 273, 1140, 18, 803, 18, 5701, 12, 2844, 1621, 16, 296, 659, 18, 5830, 6134, 1296, 67, 1968, 273, 13666,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2608, 7363, 13087, 12, 2890, 16, 1544, 4672, 21412, 273, 31381, 5814, 1435, 15047, 273, 1140, 18, 803, 18, 5701, 12, 2844, 1621, 16, 296, 659, 18, 5830, 6134, 1296, 67, 1968, 273, 13666,...
global _old_cpu_times
global _last_cpu_times percentages = []
def cpu_percent(): """Return CPU utilization times (user, kernel, idle) as a percentage calculated from the last call to the function. """ global _old_cpu_times times = cpu_times() delta_cpu = sum(times) - sum(_old_cpu_times) deltas = [m - n for m,n in zip(times, _old_cpu_times)] percent = tuple([(100 * x) / delta_cpu for x in deltas]) _old_cpu_times = times return percent
7a2fac16a92df1733b67d4b6f3db3a6b537660c5 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7296/7a2fac16a92df1733b67d4b6f3db3a6b537660c5/_psutil.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8326, 67, 8849, 13332, 3536, 990, 12154, 1709, 1588, 4124, 261, 1355, 16, 5536, 16, 12088, 13, 487, 279, 11622, 8894, 628, 326, 1142, 745, 358, 326, 445, 18, 3536, 2552, 389, 2722, 67, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8326, 67, 8849, 13332, 3536, 990, 12154, 1709, 1588, 4124, 261, 1355, 16, 5536, 16, 12088, 13, 487, 279, 11622, 8894, 628, 326, 1142, 745, 358, 326, 445, 18, 3536, 2552, 389, 2722, 67, ...
x, y = self.cr.user_to_device (x, y) w, h = self.cr.user_to_device_distance (w, h)
x, y, w, h = self._user_to_device_box (x, y, w, h)
def set_allocation (self, x, y, w, h): x, y = self.cr.user_to_device (x, y) w, h = self.cr.user_to_device_distance (w, h) self.extents = [x, y, w, h]
a1f2567529166c613d13b3c01e4a40cf75dcbc50 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8139/a1f2567529166c613d13b3c01e4a40cf75dcbc50/slippy.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 444, 67, 29299, 261, 2890, 16, 619, 16, 677, 16, 341, 16, 366, 4672, 619, 16, 677, 16, 341, 16, 366, 273, 365, 6315, 1355, 67, 869, 67, 5964, 67, 2147, 261, 92, 16, 677, 16, 341, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 444, 67, 29299, 261, 2890, 16, 619, 16, 677, 16, 341, 16, 366, 4672, 619, 16, 677, 16, 341, 16, 366, 273, 365, 6315, 1355, 67, 869, 67, 5964, 67, 2147, 261, 92, 16, 677, 16, 341, ...
self.addErropr(error)
self.addError(error)
def addErrors(self, errors): """Adds an dict, list or ValidPieErrorSchema of errors""" if isinstance(errors, (ValidPieErrorSchema)): for error in errors.getGlobalErrors(): self.addError(error) for name, error in errors.getNamedErrors().iteritems(): self.addError(error, name) elif isinstance(errors, (dict)): for name, error in errors.iteritems(): self.addError(error, name) elif isinstance(errors, (dict)): for error in errors: self.addErropr(error)
5cb1d6127e3f53091275a3e0550b8c4d3109f97d /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6662/5cb1d6127e3f53091275a3e0550b8c4d3109f97d/ValidPieErrorSchema.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 4229, 12, 2890, 16, 1334, 4672, 3536, 3655, 392, 2065, 16, 666, 578, 2364, 52, 1385, 668, 3078, 434, 1334, 8395, 309, 1549, 12, 4324, 16, 261, 1556, 52, 1385, 668, 3078, 3719, 30,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 4229, 12, 2890, 16, 1334, 4672, 3536, 3655, 392, 2065, 16, 666, 578, 2364, 52, 1385, 668, 3078, 434, 1334, 8395, 309, 1549, 12, 4324, 16, 261, 1556, 52, 1385, 668, 3078, 3719, 30,...
choice = wikipedia.input('WARNING: %s doesn\'t seem to be a disambiguation page, but %s is one. Follow it anyway? [y|N]' % (self.inpl.aslink(), pl.aslink()))
choice = wikipedia.inputChoice('WARNING: %s doesn\'t seem to be a disambiguation page, but %s is one. Follow it anyway?' % (self.inpl.aslink(), pl.aslink()), ['Yes', 'No'], ['y', 'N'], 'N')
def workDone(self, counter): """This is called by a worker to tell us that the promised work was completed as far as possible. The only argument is an instance of a counter class, that has methods minus() and plus() to keep counts of the total work todo.""" # Loop over all the pages that should have been taken care of for pl in self.pending: # Mark the page as done self.done[pl] = pl.site() # Register this fact at the todo-counter. counter.minus(pl.site()) # Assume it's not a redirect isredirect = 0 # Now check whether any interwiki links should be added to the # todo list. if pl.section(): # We have been referred to a part of a page, not the whole page. Do not follow references. pass else: try: iw = pl.interwiki() except wikipedia.IsRedirectPage,arg: pl3 = wikipedia.Page(pl.site(),arg.args[0]) wikipedia.output(u"NOTE: %s is redirect to %s" % (pl.aslink(), pl3.aslink())) if pl == self.inpl: # This is a redirect page itself. We don't need to # follow the redirection. isredirect = 1 # In this case we can also stop all hints! for pl2 in self.todo: counter.minus(pl2.site()) self.todo = {} pass elif not globalvar.followredirect: print "NOTE: not following redirects." else: if self.conditionalAdd(pl3, counter, pl): if globalvar.shownew: wikipedia.output(u"%s: %s gives new redirect %s" % (self.inpl.aslink(), pl.aslink(), pl3.aslink())) except wikipedia.NoPage: wikipedia.output(u"NOTE: %s does not exist" % pl.aslink()) #print "DBG> ",pl.urlname() if pl == self.inpl: # This is the home subject page. # In this case we can stop all hints! for pl2 in self.todo: counter.minus(pl2.site()) self.todo = {} self.done = {} # In some rare cases it might be we already did check some 'automatic' links pass #except wikipedia.SectionError: # wikipedia.output(u"NOTE: section %s does not exist" % pl.aslink()) else: if not globalvar.autonomous: if self.inpl.isDisambig() and not pl.isDisambig(): choice = wikipedia.input('WARNING: %s is a disambiguation page, but %s doesn\'t seem to be one. Follow it anyway? [y|N]' % (self.inpl.aslink(), pl.aslink())) elif not self.inpl.isDisambig() and pl.isDisambig(): choice = wikipedia.input('WARNING: %s doesn\'t seem to be a disambiguation page, but %s is one. Follow it anyway? [y|N]' % (self.inpl.aslink(), pl.aslink())) else: choice = 'y' if choice not in ['y', 'Y']: wikipedia.output(u"NOTE: ignoring %s and its interwiki links" % pl.aslink()) del self.done[pl] iw = () if self.inpl == pl: self.untranslated = (len(iw) == 0) if globalvar.untranslatedonly: # Ignore the interwiki links. iw = () elif pl.isEmpty(): if not pl.isCategory(): wikipedia.output(u"NOTE: %s is empty; ignoring it and its interwiki links" % pl.aslink()) # Ignore the interwiki links iw = () for page2 in iw: if page2.site().language() in globalvar.neverlink: print "Skipping link %s to an ignored language"% page2 continue if globalvar.same=='wiktionary' and page2.linkname().lower()!=self.inpl.linkname().lower(): print "NOTE: Ignoring %s for %s in wiktionary mode"% (page2, self.inpl) continue if not globalvar.autonomous: if self.inpl.namespace() != page2.namespace(): choice = wikipedia.input('WARNING: %s is in namespace %i, but %s is in namespace %i. Follow it anyway? [y|N]' % (self.inpl.aslink(), self.inpl.namespace(), page2.aslink(), page2.namespace())) if choice not in ['y', 'Y']: continue if self.conditionalAdd(page2, counter, pl): if globalvar.shownew: wikipedia.output(u"%s: %s gives new interwiki %s"% (self.inpl.aslink(), pl.aslink(), page2.aslink())) # These pages are no longer 'in progress' del self.pending # Check whether we need hints and the user offered to give them if self.untranslated and not self.hintsasked: wikipedia.output(u"NOTE: %s does not have any interwiki links" % self.inpl.aslink()) if (self.untranslated or globalvar.askhints) and not self.hintsasked and not isredirect: # Only once! self.hintsasked = True if globalvar.untranslated: newhint = None t = globalvar.showtextlink if t: wikipedia.output(pl.get()[:t]) while 1: newhint = wikipedia.input(u'Give a hint (? to see pagetext):') if newhint == '?': t += globalvar.showtextlinkadd wikipedia.output(pl.get()[:t]) elif newhint and not ':' in newhint: print "Please enter a hint like language:pagename" print "or type nothing if you do not have a hint" elif not newhint: break else: arr = {} titletranslate.translate(pl, arr, same = False, hints = [newhint], auto = globalvar.auto) for pl2 in arr.iterkeys(): self.todo[pl2] = pl2.site() counter.plus(pl2.site()) self.foundin[pl2] = [None]
6b8ea9c735b1ac81d36c149cb1ad028fdb959a3b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4404/6b8ea9c735b1ac81d36c149cb1ad028fdb959a3b/interwiki.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1440, 7387, 12, 2890, 16, 3895, 4672, 3536, 2503, 353, 2566, 635, 279, 4322, 358, 9276, 584, 716, 326, 3012, 5918, 1440, 1703, 5951, 487, 10247, 487, 3323, 18, 1021, 1338, 1237, 353, 392...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1440, 7387, 12, 2890, 16, 3895, 4672, 3536, 2503, 353, 2566, 635, 279, 4322, 358, 9276, 584, 716, 326, 3012, 5918, 1440, 1703, 5951, 487, 10247, 487, 3323, 18, 1021, 1338, 1237, 353, 392...
return (cartpos * la.inv(self._lattice)).sum(0)
return (cartpos * numpy.linalg.inv(self._lattice)).sum(0)
def cartesianToFractional(self, cartpos): """Converts a coordinate from cartesian to fractional.""" return (cartpos * la.inv(self._lattice)).sum(0) # should be double-checked
f04f662168003ad199af378621e7dfef9705eccf /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/578/f04f662168003ad199af378621e7dfef9705eccf/Structure.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24549, 774, 13724, 287, 12, 2890, 16, 7035, 917, 4672, 3536, 5692, 279, 7799, 628, 24549, 358, 20462, 12123, 327, 261, 11848, 917, 380, 3972, 18, 80, 11521, 18, 5768, 12, 2890, 6315, 80,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24549, 774, 13724, 287, 12, 2890, 16, 7035, 917, 4672, 3536, 5692, 279, 7799, 628, 24549, 358, 20462, 12123, 327, 261, 11848, 917, 380, 3972, 18, 80, 11521, 18, 5768, 12, 2890, 6315, 80,...
example += '<Language Name="Example Language" Author="arnetheduck" Version=' + version + ' Revision="1" RightToLeft="0">\n'
example += '<Language Name="Example Language" Native="English" Code="en" Author="arnetheduck" Version=' + version + ' Revision="1" RightToLeft="0">\n'
def makename(oldname): name = ""; nextBig = True; for x in oldname: if x == '_': nextBig = True; else: if nextBig: name += x.upper(); nextBig = False; else: name += x.lower(); return name;
d4b507eb52953dec9dac8e84383d3e24ce7dda6f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2638/d4b507eb52953dec9dac8e84383d3e24ce7dda6f/makedefs.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 29796, 1069, 12, 1673, 529, 4672, 508, 273, 1408, 31, 1024, 9901, 273, 1053, 31, 364, 619, 316, 1592, 529, 30, 309, 619, 422, 2070, 4278, 1024, 9901, 273, 1053, 31, 469, 30, 309, 1024,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 29796, 1069, 12, 1673, 529, 4672, 508, 273, 1408, 31, 1024, 9901, 273, 1053, 31, 364, 619, 316, 1592, 529, 30, 309, 619, 422, 2070, 4278, 1024, 9901, 273, 1053, 31, 469, 30, 309, 1024,...
raise error_temp, resp
raise NNTPTemporaryError(resp)
def getresp(self): """Internal: get a response from the server. Raise various errors if the response indicates an error.""" resp = self.getline() if self.debugging: print '*resp*', `resp` c = resp[:1] if c == '4': raise error_temp, resp if c == '5': raise error_perm, resp if c not in '123': raise error_proto, resp return resp
6e5a5b653cf6d01e04eb75ed5d23ba576162ee46 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/6e5a5b653cf6d01e04eb75ed5d23ba576162ee46/nntplib.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 12243, 12, 2890, 4672, 3536, 3061, 30, 336, 279, 766, 628, 326, 1438, 18, 20539, 11191, 1334, 309, 326, 766, 8527, 392, 555, 12123, 1718, 273, 365, 18, 588, 1369, 1435, 309, 365, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 12243, 12, 2890, 4672, 3536, 3061, 30, 336, 279, 766, 628, 326, 1438, 18, 20539, 11191, 1334, 309, 326, 766, 8527, 392, 555, 12123, 1718, 273, 365, 18, 588, 1369, 1435, 309, 365, ...
raise tempalte.TemplateSyntaxError("No such variable: %s", self.cart)
raise template.TemplateSyntaxError("No such variable: %s", self.cart)
def render(self, context): taxer = _get_taxprocessor(context['request']) try: cart = template.resolve_variable(self.cart, context) except template.VariableDoesNotExist: raise tempalte.TemplateSyntaxError("No such variable: %s", self.cart)
e88bf7eb44d0e3ed6d63e3ee5961dc5160e6b466 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/171/e88bf7eb44d0e3ed6d63e3ee5961dc5160e6b466/satchmo_tax.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1743, 12, 2890, 16, 819, 4672, 5320, 264, 273, 389, 588, 67, 8066, 8700, 12, 2472, 3292, 2293, 19486, 775, 30, 7035, 273, 1542, 18, 10828, 67, 6105, 12, 2890, 18, 11848, 16, 819, 13, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1743, 12, 2890, 16, 819, 4672, 5320, 264, 273, 389, 588, 67, 8066, 8700, 12, 2472, 3292, 2293, 19486, 775, 30, 7035, 273, 1542, 18, 10828, 67, 6105, 12, 2890, 18, 11848, 16, 819, 13, ...