rem stringlengths 2 226k | add stringlengths 0 227k | context stringlengths 8 228k | meta stringlengths 156 215 | input_ids list | attention_mask list | labels list |
|---|---|---|---|---|---|---|
def setup_config(self): | def setup_config(self, load_defaults=False): | def setup_config(self): """Load the configuration file.""" self.config = Configuration(os.path.join(self.path, 'conf', 'trac.ini')) | ab363daddb4bbfbf90cb05e9d214467f6d158544 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2831/ab363daddb4bbfbf90cb05e9d214467f6d158544/env.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
67,
1425,
12,
2890,
16,
1262,
67,
7606,
33,
8381,
4672,
3536,
2563,
326,
1664,
585,
12123,
365,
18,
1425,
273,
4659,
12,
538,
18,
803,
18,
5701,
12,
2890,
18,
803,
16,
296,
392... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
67,
1425,
12,
2890,
16,
1262,
67,
7606,
33,
8381,
4672,
3536,
2563,
326,
1664,
585,
12123,
365,
18,
1425,
273,
4659,
12,
538,
18,
803,
18,
5701,
12,
2890,
18,
803,
16,
296,
392... |
cmd = 'select * from cmds where taskrow="%s"' % (taskrow) | cmd = 'select * from cmds where taskrow="%s" LIMIT 200' % (taskrow) | def showTaskCommandsByRow(self, taskrow): cur = self.cursor() cmd = 'select * from cmds where taskrow="%s"' % (taskrow) cur.execute(cmd) for cidtuple in cur.fetchall(): self.showCmdTuple(cidtuple) pass | 5b6a02c82a858db1229fd5d4d621cd3d16fd2060 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5203/5b6a02c82a858db1229fd5d4d621cd3d16fd2060/ssdap_dbutil.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2405,
2174,
9127,
858,
1999,
12,
2890,
16,
1562,
492,
4672,
662,
273,
365,
18,
9216,
1435,
1797,
273,
296,
4025,
380,
628,
13446,
1625,
1562,
492,
11613,
87,
6,
13319,
4044,
11,
738,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2405,
2174,
9127,
858,
1999,
12,
2890,
16,
1562,
492,
4672,
662,
273,
365,
18,
9216,
1435,
1797,
273,
296,
4025,
380,
628,
13446,
1625,
1562,
492,
11613,
87,
6,
13319,
4044,
11,
738,
2... |
"""Checks that the user didn't paste 'Suppression:' into the file""" keyword = 'Suppression:' | """Checks the memcheck suppressions files for bad data.""" errors = [] skip_next_line = False func_re = input_api.re.compile('[a-z_.]+\(.+\)$') | def CheckChange(input_api, output_api): """Checks that the user didn't paste 'Suppression:' into the file""" keyword = 'Suppression:' for f, line_num, line in input_api.RightHandSideLines(lambda x: x.LocalPath().endswith('.txt')): if keyword in line: text = '"%s" must not be included; %s line %s' % ( keyword, f.LocalPath(), line_num) return [output_api.PresubmitError(text)] return [] | 2c2eed3469532e2b0f7c705a858da8cdfecbb609 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9392/2c2eed3469532e2b0f7c705a858da8cdfecbb609/PRESUBMIT.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2073,
3043,
12,
2630,
67,
2425,
16,
876,
67,
2425,
4672,
3536,
4081,
326,
1663,
1893,
1169,
5012,
1390,
364,
5570,
501,
12123,
1334,
273,
5378,
2488,
67,
4285,
67,
1369,
273,
1083,
1326,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2073,
3043,
12,
2630,
67,
2425,
16,
876,
67,
2425,
4672,
3536,
4081,
326,
1663,
1893,
1169,
5012,
1390,
364,
5570,
501,
12123,
1334,
273,
5378,
2488,
67,
4285,
67,
1369,
273,
1083,
1326,... |
context.path_on_end_event() | def def_list_stream(stream, _tag_name, attributes, context): """ stream : parser stream """ INDENT_VALUE = 1 * cm story = [] has_content = False while True: event, value, line_number = stream_next(stream) if event == None: break #### START ELEMENT #### if event == START_ELEMENT: tag_uri, tag_name, attributes = value context.path_on_start_event(tag_name, attributes) if tag_name == 'dt': story.extend(paragraph_stream(stream, tag_name, attributes, context)) elif tag_name == 'dd': story.append(Indenter(left=INDENT_VALUE)) story.extend(paragraph_stream(stream, tag_name, attributes, context)) story.append(Indenter(left=-INDENT_VALUE)) else: print '5', MSG_WARNING_DTD % ('document', line_number, tag_name) #### END ELEMENT #### elif event == END_ELEMENT: tag_uri, tag_name = value if tag_name == _tag_name: return story else: print '4', MSG_WARNING_DTD % ('document', line_number, tag_name) context.path_on_end_event() | 5f226059fe7b1a35a6e4d7f26d8aa02d2f5767f5 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12681/5f226059fe7b1a35a6e4d7f26d8aa02d2f5767f5/rml2.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1652,
67,
1098,
67,
3256,
12,
3256,
16,
389,
2692,
67,
529,
16,
1677,
16,
819,
4672,
3536,
1407,
294,
2082,
1407,
3536,
225,
30009,
67,
4051,
273,
404,
380,
5003,
17285,
273,
5378,
711... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1652,
67,
1098,
67,
3256,
12,
3256,
16,
389,
2692,
67,
529,
16,
1677,
16,
819,
4672,
3536,
1407,
294,
2082,
1407,
3536,
225,
30009,
67,
4051,
273,
404,
380,
5003,
17285,
273,
5378,
711... | |
f.write(XMLHelper.writexml(self.asXML(), UnicodeStringIO(), | f.write(XMLHelper.writexml(xmlbody, UnicodeStringIO(), | def saveFile(self): if self.filename is None: self.saveAsFile() else: try: f = open(self.filename, 'w') f.write(XMLHelper.writexml(self.asXML(), UnicodeStringIO(), '', '\t', '\n')) f.close() self.modified = False except IOError: QMessageBox.critical(None, 'Error!', 'Error writing to file: ' + self.filename, 'OK') | d280253f621b6d1e91ce22b55f8a34f03dec1dd7 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/6671/d280253f621b6d1e91ce22b55f8a34f03dec1dd7/ScWindow.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1923,
812,
12,
2890,
4672,
309,
365,
18,
3459,
353,
599,
30,
365,
18,
5688,
1463,
812,
1435,
469,
30,
775,
30,
284,
273,
1696,
12,
2890,
18,
3459,
16,
296,
91,
6134,
284,
18,
2626,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1923,
812,
12,
2890,
4672,
309,
365,
18,
3459,
353,
599,
30,
365,
18,
5688,
1463,
812,
1435,
469,
30,
775,
30,
284,
273,
1696,
12,
2890,
18,
3459,
16,
296,
91,
6134,
284,
18,
2626,
... |
if alpha <= 0.0: raise ValueError, 'stdgamma: alpha must be > 0.0' | def stdgamma(self, alpha, *args): # *args for Py2.2 compatiblity random = self.random if alpha <= 0.0: raise ValueError, 'stdgamma: alpha must be > 0.0' | 1ef4463cc4c8aa15656fd0bdff420ad6c0ca29da /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/1ef4463cc4c8aa15656fd0bdff420ad6c0ca29da/random.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2044,
20589,
12,
2890,
16,
4190,
16,
380,
1968,
4672,
225,
468,
380,
1968,
364,
4707,
22,
18,
22,
4796,
495,
80,
560,
2744,
273,
365,
18,
9188,
309,
4190,
1648,
374,
18,
20,
30,
1002... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2044,
20589,
12,
2890,
16,
4190,
16,
380,
1968,
4672,
225,
468,
380,
1968,
364,
4707,
22,
18,
22,
4796,
495,
80,
560,
2744,
273,
365,
18,
9188,
309,
4190,
1648,
374,
18,
20,
30,
1002... | |
w = self._w | """ helper function to output general date value""" w = self._w | def timeProp(self, sts, pn, subj): w = self._w when = sts.any(subj, ICAL.sym(pn)) | 04829bd0c6467d1a5ceb640d5981b3facc7553d5 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3308/04829bd0c6467d1a5ceb640d5981b3facc7553d5/toIcal.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
813,
4658,
12,
2890,
16,
27099,
16,
11059,
16,
15333,
4672,
3536,
4222,
445,
358,
876,
7470,
1509,
460,
8395,
341,
273,
365,
6315,
91,
225,
1347,
273,
27099,
18,
2273,
12,
1717,
78,
16... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
813,
4658,
12,
2890,
16,
27099,
16,
11059,
16,
15333,
4672,
3536,
4222,
445,
358,
876,
7470,
1509,
460,
8395,
341,
273,
365,
6315,
91,
225,
1347,
273,
27099,
18,
2273,
12,
1717,
78,
16... |
if string in (b'', ''): | if not string: string.split | def unquote_to_bytes(string): """unquote_to_bytes('abc%20def') -> b'abc def'.""" # Note: strings are encoded as UTF-8. This is only an issue if it contains # unescaped non-ASCII characters, which URIs should not. if string in (b'', ''): return b'' if isinstance(string, str): string = string.encode('utf-8') res = string.split(b'%') if len(res) == 1: return string string = res[0] for item in res[1:]: try: string += bytes([int(item[:2], 16)]) + item[2:] except ValueError: string += b'%' + item return string | 82a3f8afcb0716cf524b4f731e25b4d6dca554a4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8546/82a3f8afcb0716cf524b4f731e25b4d6dca554a4/parse.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
25611,
67,
869,
67,
3890,
12,
1080,
4672,
3536,
318,
6889,
67,
869,
67,
3890,
2668,
18947,
9,
3462,
536,
6134,
317,
324,
11,
18947,
1652,
11,
12123,
468,
3609,
30,
2064,
854,
3749,
487... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
25611,
67,
869,
67,
3890,
12,
1080,
4672,
3536,
318,
6889,
67,
869,
67,
3890,
2668,
18947,
9,
3462,
536,
6134,
317,
324,
11,
18947,
1652,
11,
12123,
468,
3609,
30,
2064,
854,
3749,
487... |
return u'[[%s:%s:%s]]' % (self.site().family.name, self.site().lang, self.title(savetitle=True,decode=True)) | return u'[[:%s:%s:%s]]' % (self.site().family.name, self.site().lang, self.title(savetitle=True, decode=True)) | def aslink(self, forceInterwiki=False, textlink=False, noInterwiki=False): """Return a string representation in the form of a wikilink. | 05c6471df5f3cfa508eecc71702ba48548b8c454 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/4404/05c6471df5f3cfa508eecc71702ba48548b8c454/wikipedia.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
487,
1232,
12,
2890,
16,
2944,
2465,
13044,
33,
8381,
16,
977,
1232,
33,
8381,
16,
1158,
2465,
13044,
33,
8381,
4672,
3536,
990,
279,
533,
4335,
316,
326,
646,
434,
279,
341,
1766,
330... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
487,
1232,
12,
2890,
16,
2944,
2465,
13044,
33,
8381,
16,
977,
1232,
33,
8381,
16,
1158,
2465,
13044,
33,
8381,
4672,
3536,
990,
279,
533,
4335,
316,
326,
646,
434,
279,
341,
1766,
330... |
fiscal_date_start = fiscal_obj.browse(cr, uid, [data['form']['fiscalyear_id']], context=context)[0].date_start result_initial_bal['empty_fy_allow'] = True result_initial_bal['fiscalyear'] = fiscal_obj.search(cr, uid, [('date_stop', '<', fiscal_date_start), ('state', '=', 'draft')], context=context) result_initial_bal['date_from'] = '0001-01-01' result_initial_bal['date_to'] = (datetime.datetime.strptime(fiscal_date_start, "%Y-%m-%d") + timedelta(days=-1)).strftime('%Y-%m-%d') | if data['form']['fiscalyear_id']: fiscal_date_start = fiscal_obj.browse(cr, uid, [data['form']['fiscalyear_id']], context=context)[0].date_start result_initial_bal['empty_fy_allow'] = True result_initial_bal['fiscalyear'] = fiscal_obj.search(cr, uid, [('date_stop', '<', fiscal_date_start), ('state', '=', 'draft')], context=context) result_initial_bal['date_from'] = '0001-01-01' result_initial_bal['date_to'] = (datetime.datetime.strptime(fiscal_date_start, "%Y-%m-%d") + timedelta(days=-1)).strftime('%Y-%m-%d') | def _build_contexts(self, cr, uid, ids, data, context=None): if context is None: context = {} result = {} period_obj = self.pool.get('account.period') fiscal_obj = self.pool.get('account.fiscalyear') result['fiscalyear'] = 'fiscalyear_id' in data['form'] and data['form']['fiscalyear_id'] or False result['journal_ids'] = 'journal_ids' in data['form'] and data['form']['journal_ids'] or False result['chart_account_id'] = 'chart_account_id' in data['form'] and data['form']['chart_account_id'] or False result_initial_bal = result.copy() if data['form']['filter'] == 'filter_date': result['date_from'] = data['form']['date_from'] result['date_to'] = data['form']['date_to'] result_initial_bal['date_from'] = '0001-01-01' result_initial_bal['date_to'] = (datetime.datetime.strptime(data['form']['date_from'], "%Y-%m-%d") + timedelta(days=-1)).strftime('%Y-%m-%d') elif data['form']['filter'] == 'filter_period': if not data['form']['period_from'] or not data['form']['period_to']: raise osv.except_osv(_('Error'),_('Select a starting and an ending period')) result['periods'] = self._build_periods(cr, uid, data['form']['period_from'], data['form']['period_to']) first_period = self.pool.get('account.period').search(cr, uid, [], order='date_start', limit=1)[0] result_initial_bal['periods'] = self._build_periods(cr, uid, first_period, data['form']['period_from']) else: fiscal_date_start = fiscal_obj.browse(cr, uid, [data['form']['fiscalyear_id']], context=context)[0].date_start result_initial_bal['empty_fy_allow'] = True #Improve me => there should be something generic in account.move.line -> query get result_initial_bal['fiscalyear'] = fiscal_obj.search(cr, uid, [('date_stop', '<', fiscal_date_start), ('state', '=', 'draft')], context=context) result_initial_bal['date_from'] = '0001-01-01' result_initial_bal['date_to'] = (datetime.datetime.strptime(fiscal_date_start, "%Y-%m-%d") + timedelta(days=-1)).strftime('%Y-%m-%d') return result, result_initial_bal | d2cbb8c382dda4184e8b927a00ab5c29208a17cf /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/d2cbb8c382dda4184e8b927a00ab5c29208a17cf/account_report_common.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
3510,
67,
18332,
12,
2890,
16,
4422,
16,
4555,
16,
3258,
16,
501,
16,
819,
33,
7036,
4672,
309,
819,
353,
599,
30,
819,
273,
2618,
563,
273,
2618,
3879,
67,
2603,
273,
365,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
3510,
67,
18332,
12,
2890,
16,
4422,
16,
4555,
16,
3258,
16,
501,
16,
819,
33,
7036,
4672,
309,
819,
353,
599,
30,
819,
273,
2618,
563,
273,
2618,
3879,
67,
2603,
273,
365,
18,
... |
ret = ctxt.eval(""" ({ 'test': function(){ return [ [ 1, 'abla' ], [ 2, 'ajkss' ], ] } }) | ret = ctxt.eval(""" ({ 'test': function(){ return [ [ 1, 'abla' ], [ 2, 'ajkss' ], ] } }) | def testMultiDimArray(self): with JSContext() as ctxt: ret = ctxt.eval(""" ({ 'test': function(){ return [ [ 1, 'abla' ], [ 2, 'ajkss' ], ] } }) """).test() self.assertEquals([[1, 'abla'], [2, 'ajkss']], convert(ret)) | 550f86f8e1997ee990877076421feaf1936da8eb /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5451/550f86f8e1997ee990877076421feaf1936da8eb/PyV8.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
5002,
5225,
1076,
12,
2890,
4672,
598,
6756,
1042,
1435,
487,
14286,
30,
325,
273,
14286,
18,
8622,
2932,
3660,
10797,
296,
3813,
4278,
445,
1435,
95,
327,
225,
306,
306,
404,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
5002,
5225,
1076,
12,
2890,
4672,
598,
6756,
1042,
1435,
487,
14286,
30,
325,
273,
14286,
18,
8622,
2932,
3660,
10797,
296,
3813,
4278,
445,
1435,
95,
327,
225,
306,
306,
404,
16,
... |
DebugPrint (0, "ERROR: " + e + "exception caught while processing record " + record) | DebugPrint (0, "ERROR: " + str(e) + "exception caught while processing record ") | def Send(record): global failedSendCount global suppressedCount global successfulSendCount try: DebugPrint(0, "***********************************************************") DebugPrint(4, "DEBUG: In Send(record)") DebugPrint(4, "DEBUG: Printing record to send") record.Print() DebugPrint(4, "DEBUG: Printing record to send: OK") if (failedSendCount + len(OutstandingRecord)) >= Config.get_MaxPendingFiles(): responseString = "Fatal Error: too many pending files" DebugPrint(0, responseString) DebugPrint(0, "***********************************************************") return responseString # Assemble the record into xml DebugPrint(4, "DEBUG: Creating XML") record.XmlCreate() DebugPrint(4, "DEBUG: Creating XML: OK") # Parse it into nodes, etc (transitional: this will eventually be native format) DebugPrint(4, "DEBUG: parsing XML") xmlDoc = safeParseXML(string.join(record.XmlData,"")) DebugPrint(4, "DEBUG: parsing XML: OK") if not xmlDoc: responseString = "Internal Error: cannot parse internally generated XML record" DebugPrint(0, responseString) DebugPrint(0, "***********************************************************") return responseString DebugPrint(4, "DEBUG: Checking XML content") if not CheckXmlDoc(xmlDoc,False): DebugPrint(4, "DEBUG: Checking XML content: BAD") xmlDoc.unlink() responseString = "No unsuppressed usage records in this packet: not sending" suppressedCount += 1 DebugPrint(0, responseString) DebugPrint(0, "***********************************************************") return responseString DebugPrint(4, "DEBUG: Checking XML content: OK") DebugPrint(4, "DEBUG: Normalizing XML document") xmlDoc.normalize() DebugPrint(4, "DEBUG: Normalizing XML document: OK") # Generate the XML DebugPrint(4, "DEBUG: Generating data to send") record.XmlData = safeEncodeXML(xmlDoc).splitlines(True) DebugPrint(4, "DEBUG: Generating data to send: OK") # Close and clean up the document2 xmlDoc.unlink() dirIndex = 0 success = False ind = 0 f = 0 DebugPrint(4, "DEBUG: Back up record to send") while not success: (f,dirIndex) = OpenNewRecordFile(dirIndex) DebugPrint(1,"Will save in the record in:",f.name) DebugPrint(3,"DirIndex=",dirIndex) if f.name != "<stdout>": try: for line in record.XmlData: f.write(line) f.flush() if f.tell() > 0: success = True DebugPrint(0, 'Saved record to ' + f.name) else: DebugPrint(0,"failed to fill: ",f.name) if f.name != "<stdout>": os.remove(f.name) f.close() except: DebugPrint(0,"failed to fill with exception: ",f.name,"--", sys.exc_info(),"--",sys.exc_info()[0],"++",sys.exc_info()[1]) DebugPrint(4, "DEBUG: Backing up record to send: OK") # Currently, the recordXml is in a list format, with each item being a line of xml. # the collector web service requires the xml to be sent as a string. # This logic here turns the xml list into a single xml string. usageXmlString = "" for line in record.XmlData: usageXmlString = usageXmlString + line DebugPrint(3, 'UsageXml: ' + usageXmlString) connectionProblem = (__connectionRetries > 0) or (__connectionError) # Attempt to send the record to the collector response = __sendUsageXML(Config.get_MeterName(), usageXmlString) responseString = response.get_message() DebugPrint(0, 'Response code: ' + str(response.get_code())) DebugPrint(0, 'Response message: ' + response.get_message()) # Determine if the call was successful based on the response # code. Currently, 0 = success if response.get_code() == 0: DebugPrint(1, 'Response indicates success, ' + f.name + ' will be deleted') successfulSendCount += 1 os.remove(f.name) else: failedSendCount += 1 if (f.name == "<stdout>"): DebugPrint(0, 'Record send failed and no backup made: record lost!') responseString += "\nFatal: failed record lost!" match = re.search(r'^<(?:[^:]*:)?RecordIdentity.*/>$', usageXmlString, re.MULTILINE) if match: DebugPrint(0, match.group(0)) responseString += "\n", match.group(0) match = re.search(r'^<(?:[^:]*:)?GlobalJobId.*/>$', usageXmlString, re.MULTILINE) if match: DebugPrint(0, match.group(0)) responseString += "\n", match.group(0) responseString += "\n" + usageXmlString else: DebugPrint(1, 'Response indicates failure, ' + f.name + ' will not be deleted') if (connectionProblem) and (response.get_code() == 0): # Reprocess failed records before attempting more new ones SearchOutstandingRecord() Reprocess() DebugPrint(0, responseString) DebugPrint(0, "***********************************************************") return responseString except Exception, e: DebugPrint (0, "ERROR: " + e + "exception caught while processing record " + record) DebugPrint (0, " This record has been LOST") return "ERROR: record lost due to internal error!" | be7d341306bb627bcfdb5bc98f49959b337acc43 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/637/be7d341306bb627bcfdb5bc98f49959b337acc43/Gratia.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2479,
12,
3366,
4672,
2552,
2535,
3826,
1380,
2552,
24448,
1380,
2552,
6873,
3826,
1380,
225,
775,
30,
4015,
5108,
12,
20,
16,
315,
5021,
19482,
636,
7388,
13,
4015,
5108,
12,
24,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2479,
12,
3366,
4672,
2552,
2535,
3826,
1380,
2552,
24448,
1380,
2552,
6873,
3826,
1380,
225,
775,
30,
4015,
5108,
12,
20,
16,
315,
5021,
19482,
636,
7388,
13,
4015,
5108,
12,
24,
16,
... |
if pagename == request.page.page_name: | if request.page is not None and pagename == request.page.page_name: | def _check(request, pagename, user, right): if pagename == request.page.page_name: p = request.page # reuse is good else: p = Page(request, pagename) acl = p.getACL(request) # this will be fast in a reused page obj return acl.may(request, user, right) | 4ac66734a618cae15451937df4daab3758de88c2 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/888/4ac66734a618cae15451937df4daab3758de88c2/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1893,
12,
2293,
16,
4262,
1069,
16,
729,
16,
2145,
4672,
309,
590,
18,
2433,
353,
486,
599,
471,
4262,
1069,
422,
590,
18,
2433,
18,
2433,
67,
529,
30,
293,
273,
590,
18,
2433,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1893,
12,
2293,
16,
4262,
1069,
16,
729,
16,
2145,
4672,
309,
590,
18,
2433,
353,
486,
599,
471,
4262,
1069,
422,
590,
18,
2433,
18,
2433,
67,
529,
30,
293,
273,
590,
18,
2433,
... |
super(Piece, self).mousePressEvent(event) | def mousePressEvent(self, event): blank = self.findBlank() if blank: blank_len = blank.sum() curr_len = self.sum() self.swap(blank, True) if abs(curr_len-blank_len) % 2 == 0: global check if check == 0: check = 1 elif check == 1: check = 0 | 83348beebf88c995890ffc0cc077649358bab80d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1230/83348beebf88c995890ffc0cc077649358bab80d/maze_rev.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7644,
11840,
1133,
12,
2890,
16,
871,
4672,
7052,
273,
365,
18,
4720,
7796,
1435,
309,
7052,
30,
7052,
67,
1897,
273,
7052,
18,
1364,
1435,
4306,
67,
1897,
225,
273,
365,
18,
1364,
143... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7644,
11840,
1133,
12,
2890,
16,
871,
4672,
7052,
273,
365,
18,
4720,
7796,
1435,
309,
7052,
30,
7052,
67,
1897,
273,
7052,
18,
1364,
1435,
4306,
67,
1897,
225,
273,
365,
18,
1364,
143... | |
child.splitText(len(string.rstrip(data))) | child.splitText(len(data.rstrip())) | def build_para(doc, parent, start, i): children = parent.childNodes after = start + 1 have_last = 0 BREAK_ELEMENTS = PARA_LEVEL_ELEMENTS + RECURSE_INTO_PARA_CONTAINERS # Collect all children until \n\n+ is found in a text node or a # member of BREAK_ELEMENTS is found. for j in range(start, i): after = j + 1 child = children[j] nodeType = child.nodeType if nodeType == ELEMENT: if child.tagName in BREAK_ELEMENTS: after = j break elif nodeType == TEXT: pos = string.find(child.data, "\n\n") if pos == 0: after = j break if pos >= 1: child.splitText(pos) break else: have_last = 1 if (start + 1) > after: raise ConversionError( "build_para() could not identify content to turn into a paragraph") if children[after - 1].nodeType == TEXT: # we may need to split off trailing white space: child = children[after - 1] data = child.data if string.rstrip(data) != data: have_last = 0 child.splitText(len(string.rstrip(data))) para = doc.createElement(PARA_ELEMENT) prev = None indexes = range(start, after) indexes.reverse() for j in indexes: node = parent.childNodes[j] parent.removeChild(node) para.insertBefore(node, prev) prev = node if have_last: parent.appendChild(para) parent.appendChild(doc.createTextNode("\n\n")) return len(parent.childNodes) else: nextnode = parent.childNodes[start] if nextnode.nodeType == TEXT: if nextnode.data and nextnode.data[0] != "\n": nextnode.data = "\n" + nextnode.data else: newnode = doc.createTextNode("\n") parent.insertBefore(newnode, nextnode) nextnode = newnode start = start + 1 parent.insertBefore(para, nextnode) return start + 1 | 0572f502bec39f741f2ea3d4e8871d02a15520b9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/0572f502bec39f741f2ea3d4e8871d02a15520b9/docfixer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1361,
67,
25072,
12,
2434,
16,
982,
16,
787,
16,
277,
4672,
2325,
273,
982,
18,
3624,
3205,
1839,
273,
787,
397,
404,
1240,
67,
2722,
273,
374,
605,
22176,
67,
10976,
55,
273,
3939,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1361,
67,
25072,
12,
2434,
16,
982,
16,
787,
16,
277,
4672,
2325,
273,
982,
18,
3624,
3205,
1839,
273,
787,
397,
404,
1240,
67,
2722,
273,
374,
605,
22176,
67,
10976,
55,
273,
3939,
... |
self.assertRaises(_ped.CreateException, self.set_end, -1) self.assertRaises(_ped.CreateException, self.set_end, 1000000000) | self.assertRaises(_ped.CreateException, self.g.set_end, -1) self.assertRaises(_ped.CreateException, self.g.set_end, 1000000000) | def runTest(self): self.g.set_end(50) self.assert_(self.g.start == 0) self.assert_(self.g.length == 50) self.assert_(self.g.end == 50) | 7ea094a5a9369af6702e865d752b1bac938f2e5d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5978/7ea094a5a9369af6702e865d752b1bac938f2e5d/test_geometry.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
4709,
12,
2890,
4672,
365,
18,
75,
18,
542,
67,
409,
12,
3361,
13,
365,
18,
11231,
67,
12,
2890,
18,
75,
18,
1937,
422,
374,
13,
365,
18,
11231,
67,
12,
2890,
18,
75,
18,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
4709,
12,
2890,
4672,
365,
18,
75,
18,
542,
67,
409,
12,
3361,
13,
365,
18,
11231,
67,
12,
2890,
18,
75,
18,
1937,
422,
374,
13,
365,
18,
11231,
67,
12,
2890,
18,
75,
18,
2... |
charset = mimeview.get_charset(old_node.content_type) or \ default_charset old_content = util.to_utf8(old_content, charset) | charset = mimeview.get_charset(old_node.content_type) if not charset: charset = mimeview.detect_unicode(old_content) old_content = util.to_utf8(old_content, charset or default_charset) | def _render_html(self, req, repos, chgset, diff_options): """HTML version""" req.hdf['title'] = '[%s]' % chgset.rev req.hdf['changeset'] = { 'revision': chgset.rev, 'time': util.format_datetime(chgset.date), 'author': util.escape(chgset.author or 'anonymous'), 'message': wiki_to_html(chgset.message or '--', self.env, req, escape_newlines=True) } | 4bc2b42b3bd4c3eb01aa8a6468603f1efe53ed1c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2831/4bc2b42b3bd4c3eb01aa8a6468603f1efe53ed1c/changeset.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
5902,
67,
2620,
12,
2890,
16,
1111,
16,
13686,
16,
462,
75,
542,
16,
3122,
67,
2116,
4672,
3536,
4870,
1177,
8395,
1111,
18,
26428,
3292,
2649,
3546,
273,
21848,
87,
3864,
738,
46... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
5902,
67,
2620,
12,
2890,
16,
1111,
16,
13686,
16,
462,
75,
542,
16,
3122,
67,
2116,
4672,
3536,
4870,
1177,
8395,
1111,
18,
26428,
3292,
2649,
3546,
273,
21848,
87,
3864,
738,
46... |
runSystemEx('mv VMware_conversion/shared/converted_img.vmdk %s' % os.path.join(vmWareDir, 'clovr.9-04.x86-64.beta-%s.vmdk' % options('general.version'))) | runSystemEx('mv VMware_conversion/shared/converted_img.vmdk %s' % os.path.join(vmWareDir, 'clovr.9-04.x86-64.%s.vmdk' % options('general.version'))) | def main(options, _args): runSystemEx('svn copy https://clovr.svn.sourceforge.net/svnroot/clovr/trunk https://clovr.svn.sourceforge.net/svnroot/clovr/tags/%s -m "Cutting release %s"' % (options('general.version'), options('general.version')), log=True) runSystemEx('svn copy https://vappio.svn.sourceforge.net/svnroot/vappio/trunk https://vappio.svn.sourceforge.net/svnroot/vappio/tags/%s -m "Cutting release %s"' % (options('general.version'), options('general.version')), log=True) runSystemEx('scp %s:/export/%s .' % (options('general.remote_name'), options('general.image')), log=True) runSystemEx('cp %s /usr/local/projects/clovr/images' % options('general.image'), log=True) runSystemEx('cp %s VMware_conversion/shared/convert_img.img' % options('general.image'), log=True) waitForPasswordChange() bundleChannel = threads.runThreadWithChannel(bundleAMI)[1].sendWithChannel(options) convertChannel = threads.runThreadWithChannel(convertImage)[1].sendWithChannel(options) try: amiId = bundleChannel.receive() logPrint('AMI: ' + amiId) except Exception, err: amiId = None errorPrint('Bundling AMI failed for some reason. Error message:') errorPrint(str(err)) try: convertChannel.receive() vmWareDir = 'clovr-vmware.beta-%s' % options('general.version') runSystemEx('mkdir -p ' + vmWareDir) runSystemEx('mv VMware_conversion/shared/converted_img.vmdk %s' % os.path.join(vmWareDir, 'clovr.9-04.x86-64.beta-%s.vmdk' % options('general.version'))) runSystemEx('mkdir -p %s %s' % (os.path.join(vmWareDir, 'keys'), os.path.join(vmWareDir, 'user_data'))) runSystemEx('cp -rv /usr/local/projects/clovr/shared ' + vmWareDir) fout = open(os.path.join(vmWareDir, 'start_clovr.vmx'), 'w') clovrConf = config.configFromMap(dict(version=options('general.version'))) for line in open('/usr/local/projects/clovr/start_clovr.vmx'): fout.write(config.replaceStr(line, clovrConf)) except Exception, err: errorPrint('Converting image failed. Error message:') errorPrint(str(err)) | ddca11fdef05bad1b5f0af6e0eeed6a0c3fd5136 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5741/ddca11fdef05bad1b5f0af6e0eeed6a0c3fd5136/releaseCutS2.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
2116,
16,
389,
1968,
4672,
1086,
3163,
424,
2668,
31505,
1610,
2333,
2207,
7550,
18090,
18,
31505,
18,
3168,
1884,
908,
18,
2758,
19,
31505,
3085,
19,
7550,
18090,
19,
26133,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
2116,
16,
389,
1968,
4672,
1086,
3163,
424,
2668,
31505,
1610,
2333,
2207,
7550,
18090,
18,
31505,
18,
3168,
1884,
908,
18,
2758,
19,
31505,
3085,
19,
7550,
18090,
19,
26133,
2... |
extra_info = fft_opt_info | extra_info = [fft_opt_info, djbfft_info], | def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration from numpy.distutils.system_info import get_info config = Configuration('fftpack',parent_package, top_path) fft_opt_info = get_info('fft_opt') config.add_data_dir('tests') config.add_library('dfftpack', sources=[join('dfftpack','*.f')]) sources = ['fftpack.pyf','src/zfft.c','src/drfft.c','src/zrfft.c', 'src/zfftnd.c'] config.add_extension('_fftpack', sources=sources, libraries=['dfftpack'], extra_info = fft_opt_info ) config.add_extension('convolve', sources = ['convolve.pyf','src/convolve.c'], libraries = ['dfftpack'], extra_info = fft_opt_info ) return config | e6df4f9da588ebde3b652de8d07b242580d5e77b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12971/e6df4f9da588ebde3b652de8d07b242580d5e77b/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1664,
12,
2938,
67,
5610,
2218,
2187,
3669,
67,
803,
33,
7036,
4672,
628,
3972,
18,
4413,
5471,
18,
23667,
67,
1367,
1930,
4659,
628,
3972,
18,
4413,
5471,
18,
4299,
67,
1376,
1930,
33... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1664,
12,
2938,
67,
5610,
2218,
2187,
3669,
67,
803,
33,
7036,
4672,
628,
3972,
18,
4413,
5471,
18,
23667,
67,
1367,
1930,
4659,
628,
3972,
18,
4413,
5471,
18,
4299,
67,
1376,
1930,
33... |
result['name'] += ':' + self.pool.get('account.period').read(cr, uid, [data['period_id']], context=context)[0]['code'] | period_code = period_obj.read(cr, uid, [data['period_id']], context=context)[0]['code'] result['name'] += period_code and (':' + period_code) or '' | def account_tax_chart_open_window(self, cr, uid, ids, context=None): """ Opens chart of Accounts @param cr: the current row, from the database cursor, @param uid: the current user’s ID for security checks, @param ids: List of account chart’s IDs @return: dictionary of Open account chart window on given fiscalyear and all Entries or posted entries """ mod_obj = self.pool.get('ir.model.data') act_obj = self.pool.get('ir.actions.act_window') period_obj = self.pool.get('account.period') if context is None: context = {} data = self.read(cr, uid, ids, [], context=context)[0] result = mod_obj._get_id(cr, uid, 'account', 'action_tax_code_tree') id = mod_obj.read(cr, uid, [result], ['res_id'], context=context)[0]['res_id'] result = act_obj.read(cr, uid, [id], context=context)[0] if data['period_id']: fiscalyear_id = period_obj.read(cr, uid, [data['period_id']], context=context)[0]['fiscalyear_id'][0] result['context'] = str({'period_id': data['period_id'], \ 'fiscalyear_id': fiscalyear_id, \ 'state': data['target_move']}) else: result['context'] = str({'state': data['target_move']}) | d49ef5795e0c3e6fc925c0887750b1d8be5150d9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/d49ef5795e0c3e6fc925c0887750b1d8be5150d9/account_tax_chart.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2236,
67,
8066,
67,
11563,
67,
3190,
67,
5668,
12,
2890,
16,
4422,
16,
4555,
16,
3258,
16,
819,
33,
7036,
4672,
3536,
6066,
773,
4980,
434,
30849,
632,
891,
4422,
30,
326,
783,
1027,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2236,
67,
8066,
67,
11563,
67,
3190,
67,
5668,
12,
2890,
16,
4422,
16,
4555,
16,
3258,
16,
819,
33,
7036,
4672,
3536,
6066,
773,
4980,
434,
30849,
632,
891,
4422,
30,
326,
783,
1027,
... |
def keyfunc((k, v), ltrs=uppercase+'_'): | def keyfunc((k, v), letters=string.ascii_uppercase + '_'): | def keyfunc((k, v), ltrs=uppercase+'_'): # hack: mutate the subitems dicts to a list in the keyfunc v[1] = sorted((si, se) for (si, (se, void)) in v[1].iteritems()) # now calculate the key letter = k[0].upper() if letter in ltrs: return letter else: # get all other symbols under one heading return 'Symbols' | ee79eebe5d0978484fe53c99134a639351f09248 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7032/ee79eebe5d0978484fe53c99134a639351f09248/environment.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
498,
644,
12443,
79,
16,
331,
3631,
328,
28007,
33,
5797,
3593,
6797,
4623,
4672,
468,
11769,
30,
15887,
326,
720,
3319,
15838,
358,
279,
666,
316,
326,
498,
644,
331,
63,
21,
65,
273,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
498,
644,
12443,
79,
16,
331,
3631,
328,
28007,
33,
5797,
3593,
6797,
4623,
4672,
468,
11769,
30,
15887,
326,
720,
3319,
15838,
358,
279,
666,
316,
326,
498,
644,
331,
63,
21,
65,
273,... |
if not line: self.cmd_missing_arg() return | def ftp_STRU(self, line): # obsolete (backward compatibility with older ftp clients) if not line: self.cmd_missing_arg() return if line in ('f','F'): self.respond('200 File transfer structure set to: F.') else: self.respond('504 Unimplemented STRU type.') | 42cfb92815cfe91b8e9573da1b07cf8dce19a9ff /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7236/42cfb92815cfe91b8e9573da1b07cf8dce19a9ff/FTPServer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13487,
67,
3902,
57,
12,
2890,
16,
980,
4672,
468,
27191,
261,
29275,
8926,
598,
12156,
13487,
7712,
13,
309,
980,
316,
7707,
74,
17023,
42,
11,
4672,
365,
18,
17863,
2668,
6976,
1387,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13487,
67,
3902,
57,
12,
2890,
16,
980,
4672,
468,
27191,
261,
29275,
8926,
598,
12156,
13487,
7712,
13,
309,
980,
316,
7707,
74,
17023,
42,
11,
4672,
365,
18,
17863,
2668,
6976,
1387,
... | |
log = open(CFG_LOGDIR + '/alertengine.log', 'a') log.write(strftime('%Y%m%d%H%M%S log.write(msg + '\n') log.close() | logfile = open(CFG_LOGDIR + '/alertengine.log', 'a') logfile.write(strftime('%Y%m%d%H%M%S logfile.write(msg + '\n') logfile.close() | def log(msg): try: log = open(CFG_LOGDIR + '/alertengine.log', 'a') log.write(strftime('%Y%m%d%H%M%S#')) log.write(msg + '\n') log.close() except Exception: register_exception() | 77e85824dd4755257c9421e9ecbb76badd7c7e7a /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12027/77e85824dd4755257c9421e9ecbb76badd7c7e7a/alert_engine.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
613,
12,
3576,
4672,
775,
30,
613,
273,
1696,
12,
19727,
67,
4842,
4537,
397,
1173,
11798,
8944,
18,
1330,
2187,
296,
69,
6134,
613,
18,
2626,
12,
701,
9982,
29909,
61,
9,
81,
9,
72,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
613,
12,
3576,
4672,
775,
30,
613,
273,
1696,
12,
19727,
67,
4842,
4537,
397,
1173,
11798,
8944,
18,
1330,
2187,
296,
69,
6134,
613,
18,
2626,
12,
701,
9982,
29909,
61,
9,
81,
9,
72,... |
v -- vector | a -- vector | def lcm(a, b=None, integer=False): """ The least common multiple of a and b, or if a is a list and b is omitted the least common multiple of all elements of v. NOTE: Use integer=True to make this vastly faster if you are working with lists of integers. INPUT: a -- number b -- number (optional) integer -- (default: False); if True, do an integer LCM or v -- vector integer -- (default: False); if True, do an integer LCM NOTE -- this is *vastly* faster than doing the generic LCM EXAMPLES: sage: LCM(97,100) 9700 sage: LCM(0,2) 0 sage: LCM(-3,-5) 15 sage: LCM([1,2,3,4,5/3]) 60 sage: v = LCM(range(1,10000),integer=True) # *very* fast! sage: len(str(v)) 4349 """ if integer: if isinstance(a,(list,tuple)): return sage.rings.integer.LCM_list(a) else: return integer_ring.ZZ(a).lcm(\ integer_ring.ZZ(b)) if isinstance(a, (list,tuple)): return __LCM_list(a) if not isinstance(a, RingElement): a = integer_ring.ZZ(a) return a.lcm(b) | bdc03f71082d359479de239409797484223e36e0 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9417/bdc03f71082d359479de239409797484223e36e0/arith.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
328,
7670,
12,
69,
16,
324,
33,
7036,
16,
3571,
33,
8381,
4672,
3536,
1021,
4520,
2975,
3229,
434,
279,
471,
324,
16,
578,
309,
279,
353,
279,
666,
471,
324,
353,
16195,
326,
4520,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
328,
7670,
12,
69,
16,
324,
33,
7036,
16,
3571,
33,
8381,
4672,
3536,
1021,
4520,
2975,
3229,
434,
279,
471,
324,
16,
578,
309,
279,
353,
279,
666,
471,
324,
353,
16195,
326,
4520,
2... |
self.currentArticle = None | def __init__(self, env=None, strict=False, debug=False, mathcache=None, lang=None, test_mode=False): localedir = os.path.join(os.path.abspath(os.path.dirname(__file__)), 'locale') translation = gettext.NullTranslations() if lang: try: translation = gettext.translation('mwlib.rl', localedir, [lang]) except IOError, exc: log.warn(str(exc)) translation.install(unicode=True) | 6c73cbec4a25a8a3842a4b065d0a1cb07ec07963 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12390/6c73cbec4a25a8a3842a4b065d0a1cb07ec07963/rlwriter.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1550,
33,
7036,
16,
5490,
33,
8381,
16,
1198,
33,
8381,
16,
4233,
2493,
33,
7036,
16,
3303,
33,
7036,
16,
1842,
67,
3188,
33,
8381,
4672,
1191,
23210,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1550,
33,
7036,
16,
5490,
33,
8381,
16,
1198,
33,
8381,
16,
4233,
2493,
33,
7036,
16,
3303,
33,
7036,
16,
1842,
67,
3188,
33,
8381,
4672,
1191,
23210,
... | |
if type(value) != type(''): | if not isinstance(value, StringType): | def config_all(self, option, value): """Set configuration options for all subwidgets (and self).""" if option == '': return elif type(option) != type(''): option = `option` if type(value) != type(''): value = `value` names = self._subwidget_names() for name in names: self.tk.call(name, 'configure', '-' + option, value) | f6251f98f783a2e87344774cc7be0958561e5a16 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/f6251f98f783a2e87344774cc7be0958561e5a16/Tix.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
642,
67,
454,
12,
2890,
16,
1456,
16,
460,
4672,
3536,
694,
1664,
702,
364,
777,
720,
18148,
261,
464,
365,
13,
12123,
309,
1456,
422,
875,
30,
327,
1327,
618,
12,
3482,
13,
480,
618... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
642,
67,
454,
12,
2890,
16,
1456,
16,
460,
4672,
3536,
694,
1664,
702,
364,
777,
720,
18148,
261,
464,
365,
13,
12123,
309,
1456,
422,
875,
30,
327,
1327,
618,
12,
3482,
13,
480,
618... |
self.__apply_limit_offset_order_get_by(params, query) | def _get_by(self, params): # Decide which loading strategy to use if "full" in params: if self.use_subqueries: return self.__get_by_subquery(params) else: return self.__get_by_n_queries(params) | a98a29e6ee7d9727d7755c479eb91f66965e37df /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/13812/a98a29e6ee7d9727d7755c479eb91f66965e37df/query_reader.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
1637,
12,
2890,
16,
859,
4672,
225,
468,
3416,
831,
1492,
7153,
6252,
358,
999,
309,
315,
2854,
6,
316,
859,
30,
309,
365,
18,
1202,
67,
1717,
13748,
30,
327,
365,
1618... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
1637,
12,
2890,
16,
859,
4672,
225,
468,
3416,
831,
1492,
7153,
6252,
358,
999,
309,
315,
2854,
6,
316,
859,
30,
309,
365,
18,
1202,
67,
1717,
13748,
30,
327,
365,
1618... | |
image2 = self.image_class(id, self.getName(), imgdata, 'image/jpeg') | image2 = Image(id, self.getName(), imgdata, 'image/jpeg') | def set(self, instance, value, **kwargs): # Do we have to delete the image? if value=="DELETE_IMAGE": ObjectField.set(self, instance, None, **kwargs) return | 589c8a00edeed82af11646c142742440f2908f59 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12165/589c8a00edeed82af11646c142742440f2908f59/Field.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
12,
2890,
16,
791,
16,
460,
16,
2826,
4333,
4672,
468,
2256,
732,
1240,
358,
1430,
326,
1316,
35,
309,
460,
31713,
6460,
67,
13603,
6877,
1033,
974,
18,
542,
12,
2890,
16,
791,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
12,
2890,
16,
791,
16,
460,
16,
2826,
4333,
4672,
468,
2256,
732,
1240,
358,
1430,
326,
1316,
35,
309,
460,
31713,
6460,
67,
13603,
6877,
1033,
974,
18,
542,
12,
2890,
16,
791,
... |
'user_type' : 'existing' | 'user_type' : 'existing', 'domain': '[]', 'access_mode': 'readonly' | def generate_random_pass(): pass_chars = RANDOM_PASS_CHARACTERS[:] random.shuffle(pass_chars) return ''.join(pass_chars[0:10]) | ed4f2f3f1ea49ce5193367036970bb0e0c3381f5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/ed4f2f3f1ea49ce5193367036970bb0e0c3381f5/wizard_share.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2103,
67,
9188,
67,
5466,
13332,
1342,
67,
7549,
273,
534,
28159,
67,
10884,
67,
7305,
6526,
11367,
10531,
65,
2744,
18,
31980,
12,
5466,
67,
7549,
13,
327,
875,
18,
5701,
12,
5466,
67... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2103,
67,
9188,
67,
5466,
13332,
1342,
67,
7549,
273,
534,
28159,
67,
10884,
67,
7305,
6526,
11367,
10531,
65,
2744,
18,
31980,
12,
5466,
67,
7549,
13,
327,
875,
18,
5701,
12,
5466,
67... |
sage: for E in cremona_optimal_curves(range(15)): ....: print E.label() ....: E.prove_BSD(verbosity=2) ....: 11a1 | sage: EllipticCurve('11a').prove_BSD(verbosity=2) | def prove_BSD(self, verbosity=0, simon=False, proof=None): """ Attempts to prove the Birch and Swinnerton-Dyer conjectural formula for `E`, returning a list of primes `p` for which this function fails to prove BSD(E,p). Here, BSD(E,p) is the statement: "the Birch and Swinnerton-Dyer formula holds up to a rational number coprime to `p`." INPUT: - ``verbosity`` - int, how much information about the proof to print. - 0 - print nothing - 1 - print sketch of proof - 2 - print information about remaining primes - ``simon`` - bool (default False), whether to use two_descent or simon_two_descent at p=2. - ``proof`` - bool or None (default: None, see proof.elliptic_curve or sage.structure.proof). If False, this function just immediately returns the empty list. EXAMPLE:: sage: for E in cremona_optimal_curves(range(15)): ....: print E.label() ....: E.prove_BSD(verbosity=2) ....: 11a1 p = 2: true by 2-descent True for p not in {2, 5} by Kolyvagin. True for p=5 by Mazur [] 14a1 p = 2: true by 2-descent True for p not in {2, 3} by Kolyvagin. Remaining primes: p = 3: reducible, surjective, good ordinary [3] | 7e12d1cab8922a4b992851b4d1773b845c6c9b38 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/7e12d1cab8922a4b992851b4d1773b845c6c9b38/ell_rational_field.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
450,
537,
67,
30780,
12,
2890,
16,
11561,
33,
20,
16,
3142,
265,
33,
8381,
16,
14601,
33,
7036,
4672,
3536,
6020,
9585,
358,
450,
537,
326,
605,
481,
343,
471,
5434,
267,
82,
1051,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
450,
537,
67,
30780,
12,
2890,
16,
11561,
33,
20,
16,
3142,
265,
33,
8381,
16,
14601,
33,
7036,
4672,
3536,
6020,
9585,
358,
450,
537,
326,
605,
481,
343,
471,
5434,
267,
82,
1051,
2... |
debug("writing %s" % str(tags)) | def write(self, filename = None): """Saves any changes you've made to the file, or to a different file if you specify one. Any existing tag will be removed.""" filename = filename or self.filename f = file(filename, "ab+") offset = self.__tag_start(f) | 6666dcf7f5e5406d88f110933303ca37afecc2ac /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4764/6666dcf7f5e5406d88f110933303ca37afecc2ac/apev2.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
12,
2890,
16,
1544,
273,
599,
4672,
3536,
13370,
1281,
3478,
1846,
8081,
7165,
358,
326,
585,
16,
578,
358,
279,
3775,
585,
309,
1846,
4800,
1245,
18,
5502,
2062,
1047,
903,
506,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
12,
2890,
16,
1544,
273,
599,
4672,
3536,
13370,
1281,
3478,
1846,
8081,
7165,
358,
326,
585,
16,
578,
358,
279,
3775,
585,
309,
1846,
4800,
1245,
18,
5502,
2062,
1047,
903,
506,
... | |
if len(result) and not self.in_list_item: | if self.in_table and line[0:2] != '||': self.close_table() if len(result) and not self.in_list_item and not self.in_table: | def format(self, text, out): self.out = out self._open_tags = [] self._list_stack = [] self.in_pre = 0 self.indent_level = 0 self.paragraph_open = 0 | af46c235a463f240f94203ff5eaa106197438805 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9317/af46c235a463f240f94203ff5eaa106197438805/Wiki.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
740,
12,
2890,
16,
977,
16,
596,
4672,
365,
18,
659,
273,
596,
365,
6315,
3190,
67,
4156,
273,
5378,
365,
6315,
1098,
67,
3772,
273,
5378,
225,
365,
18,
267,
67,
1484,
273,
374,
365,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
740,
12,
2890,
16,
977,
16,
596,
4672,
365,
18,
659,
273,
596,
365,
6315,
3190,
67,
4156,
273,
5378,
365,
6315,
1098,
67,
3772,
273,
5378,
225,
365,
18,
267,
67,
1484,
273,
374,
365,... |
if self.IsModal(): self.EndModal(True) else: self.Destroy() | self.Destroy() | def OnCloseWindow(self, event): """ Stops the timer from running and closes the window. """ if self.timer != None: self.timer.Stop() if self.IsModal(): self.EndModal(True) else: self.Destroy() | 7e88fc1a4c30ed82fefbe46871504fb5e95daa0e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/7e88fc1a4c30ed82fefbe46871504fb5e95daa0e/SplashScreen.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2755,
4605,
3829,
12,
2890,
16,
871,
4672,
3536,
934,
4473,
326,
5441,
628,
3549,
471,
13445,
326,
2742,
18,
3536,
309,
365,
18,
12542,
480,
599,
30,
365,
18,
12542,
18,
4947,
1435,
36... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2755,
4605,
3829,
12,
2890,
16,
871,
4672,
3536,
934,
4473,
326,
5441,
628,
3549,
471,
13445,
326,
2742,
18,
3536,
309,
365,
18,
12542,
480,
599,
30,
365,
18,
12542,
18,
4947,
1435,
36... |
if string.find(line, "=") != -1: cols = string.split(line) | if line.find( "=") != -1: cols = line.split() | def reference_cross_reference(self, line): assert self.data.references, "RX: missing RN" # The basic (older?) RX line is of the form: # RX MEDLINE; 85132727. # but there are variants of this that need to be dealt with (see below) # CLD1_HUMAN in Release 39 and DADR_DIDMA in Release 33 # have extraneous information in the RX line. Check for # this and chop it out of the line. # (noticed by katel@worldpath.net) ind = string.find(line, '[NCBI, ExPASy, Israel, Japan]') if ind >= 0: line = line[:ind] | ade8039411341728bf4bd4dc20b4f5d9433c8f82 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7167/ade8039411341728bf4bd4dc20b4f5d9433c8f82/SProt.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2114,
67,
14653,
67,
6180,
12,
2890,
16,
980,
4672,
1815,
365,
18,
892,
18,
14353,
16,
315,
54,
60,
30,
3315,
534,
50,
6,
468,
1021,
5337,
261,
1498,
8535,
534,
60,
980,
353,
434,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2114,
67,
14653,
67,
6180,
12,
2890,
16,
980,
4672,
1815,
365,
18,
892,
18,
14353,
16,
315,
54,
60,
30,
3315,
534,
50,
6,
468,
1021,
5337,
261,
1498,
8535,
534,
60,
980,
353,
434,
... |
Calculate skeleton features for the object OBJIMG. | Calculate skeleton features for the object OBJIMG. | def objskelfeats(objimg): """ | 15dcb93d41cb9c0de9deddcdb3914d7e9a78295c /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12199/15dcb93d41cb9c0de9deddcdb3914d7e9a78295c/imgskelfeats.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1081,
7771,
292,
29428,
12,
2603,
6081,
4672,
3536,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1081,
7771,
292,
29428,
12,
2603,
6081,
4672,
3536,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
class TestPluginManagerConstructor(TestPlugin): | class TestPluginManagerBase(TestPlugin): def gen_config(self, path_list=[], black_list=[], activate_list=[]): return { "path_list": path_list, "black_list": black_list, "activate_list": activate_list, } def gen_pluginmanager(self, path_list=[], black_list=[], activate_list=[]): return PluginManager( config=self.gen_config(path_list, black_list, activate_list) ) def gen_pre_pluginlist(self, path_list=[]): pm = self.gen_pluginmanager(path_list) return pm._walk_for_plugin() class TestPluginManagerConstructor(TestPluginManagerBase): | def testDefault(self): self.assertEquals(self.p.type, u"research") self.assertEquals(self.p.priority, 0) self.assertNone(self.p.investigate_class) | dcf180560aa938d6de9c27482c89df4f17ac6384 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5115/dcf180560aa938d6de9c27482c89df4f17ac6384/plugin.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
1868,
12,
2890,
4672,
365,
18,
11231,
8867,
12,
2890,
18,
84,
18,
723,
16,
582,
6,
266,
3072,
7923,
365,
18,
11231,
8867,
12,
2890,
18,
84,
18,
8457,
16,
374,
13,
365,
18,
11... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
1868,
12,
2890,
4672,
365,
18,
11231,
8867,
12,
2890,
18,
84,
18,
723,
16,
582,
6,
266,
3072,
7923,
365,
18,
11231,
8867,
12,
2890,
18,
84,
18,
8457,
16,
374,
13,
365,
18,
11... |
data['comments'] = row[commentsCol-1] | try: data['comments'] = row[commentsCol--1] except: pass | def on_buttonCSVImport_clicked(self, widget): logging.debug('>>') #Determine values dateCol = self.cbCSVDate.get_active() distanceCol = self.cbCSVDistance.get_active() durationCol = self.cbCSVDuration.get_active() titleCol = self.cbCSVTitle.get_active() sportCol = self.cbCSVSport.get_active() avgspeedCol = self.cbCSVAvgSpeed.get_active() maxspeedCol = self.cbCSVMaxSpeed.get_active() calCol = self.cbCSVCal.get_active() accCol = self.cbCSVAccent.get_active() desCol = self.cbCSVDescent.get_active() hrCol = self.cbCSVHR.get_active() maxHRCol = self.cbCSVMaxHR.get_active() paceCol = self.cbCSVPace.get_active() maxPaceCol = self.cbCSVMaxPace.get_active() commentsCol = self.cbCSVComments.get_active() | dd4ff95048f583245124e63653737a807c710257 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11865/dd4ff95048f583245124e63653737a807c710257/windowimportdata.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
67,
5391,
16570,
5010,
67,
7475,
329,
12,
2890,
16,
3604,
4672,
2907,
18,
4148,
2668,
34,
1870,
13,
468,
8519,
924,
1509,
914,
273,
365,
18,
7358,
16570,
1626,
18,
588,
67,
3535,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
67,
5391,
16570,
5010,
67,
7475,
329,
12,
2890,
16,
3604,
4672,
2907,
18,
4148,
2668,
34,
1870,
13,
468,
8519,
924,
1509,
914,
273,
365,
18,
7358,
16570,
1626,
18,
588,
67,
3535,
... |
print 'Just only single cpu online, quiting...' sys.exit() | e_msg = 'Single CPU online detected, test not supported.' raise error.TestNAError(e_msg) | def initialize(self): # Check if the kernel supports cpu hotplug if utils.running_config(): utils.check_for_kernel_feature('HOTPLUG_CPU') | b5ff17c8595b21f7c392429328125e91aa0f6973 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12268/b5ff17c8595b21f7c392429328125e91aa0f6973/cpu_hotplug.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4046,
12,
2890,
4672,
468,
2073,
309,
326,
5536,
6146,
8326,
12811,
22045,
309,
2990,
18,
8704,
67,
1425,
13332,
2990,
18,
1893,
67,
1884,
67,
8111,
67,
7238,
2668,
44,
1974,
18567,
43,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4046,
12,
2890,
4672,
468,
2073,
309,
326,
5536,
6146,
8326,
12811,
22045,
309,
2990,
18,
8704,
67,
1425,
13332,
2990,
18,
1893,
67,
1884,
67,
8111,
67,
7238,
2668,
44,
1974,
18567,
43,
... |
sys.stderr.write ('%s\n' % msg) | sys.stderr.write('%s\n' % msg) | def warning(msg): sys.stderr.write ('%s\n' % msg) | e75848841ac96982524448e932e12fd6ff9cbe0c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12637/e75848841ac96982524448e932e12fd6ff9cbe0c/fdb.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3436,
12,
3576,
4672,
2589,
18,
11241,
18,
2626,
29909,
87,
64,
82,
11,
738,
1234,
13,
225,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3436,
12,
3576,
4672,
2589,
18,
11241,
18,
2626,
29909,
87,
64,
82,
11,
738,
1234,
13,
225,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
"section":"condor" | "section":"condor", | def __init__(self,job,time,cp,ifo,name,opts,dag,qscanCommand): """ job = A CondorDAGJob that can run an instance of remote qscan. """ self.friendlyName = name self.id = ifo + '-' + name + '-' + repr(time) pipeline.CondorDAGNode.__init__(self,job) self.add_macro("macroid", self.id) self.jobName = job.name | 9e152961c57c114ca32f6d71b76f936ddd7c7b79 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5758/9e152961c57c114ca32f6d71b76f936ddd7c7b79/fu_Condor.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
4688,
16,
957,
16,
4057,
16,
20299,
16,
529,
16,
4952,
16,
30204,
16,
85,
9871,
2189,
4672,
3536,
1719,
273,
432,
735,
72,
280,
40,
1781,
2278,
716,
84... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
4688,
16,
957,
16,
4057,
16,
20299,
16,
529,
16,
4952,
16,
30204,
16,
85,
9871,
2189,
4672,
3536,
1719,
273,
432,
735,
72,
280,
40,
1781,
2278,
716,
84... |
if (node.klass in [vtk.vtkInteractorStyleTrackball, vtk.vtkStructuredPointsGeometryFilter, vtk.vtkConstrainedPointHandleRepresentation]): | if node.klass in obsolete_list: | def is_abstract(): """is_abstract tries to instantiate the class. If it's abstract, this will raise.""" # Consider obsolete classes abstract if (node.klass in [vtk.vtkInteractorStyleTrackball, vtk.vtkStructuredPointsGeometryFilter, vtk.vtkConstrainedPointHandleRepresentation]): return True try: getattr(vtk, node.name)() except TypeError: # VTK raises type error on abstract classes return True return False | cc822ab826b338e5e055e2c1494dcd695f84ac55 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/6341/cc822ab826b338e5e055e2c1494dcd695f84ac55/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
353,
67,
17801,
13332,
3536,
291,
67,
17801,
9327,
358,
10275,
326,
667,
18,
971,
518,
1807,
8770,
16,
333,
903,
1002,
12123,
468,
23047,
27191,
3318,
8770,
309,
756,
18,
22626,
316,
271... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
353,
67,
17801,
13332,
3536,
291,
67,
17801,
9327,
358,
10275,
326,
667,
18,
971,
518,
1807,
8770,
16,
333,
903,
1002,
12123,
468,
23047,
27191,
3318,
8770,
309,
756,
18,
22626,
316,
271... |
"type" : "JClassDepJob" | "type" : "JCompileJob", | def listJobTriggers(): return { "api" : { "type" : "JClassDepJob" }, "collect-environment-info" : { "type" : "JSimpleJob" }, | 68e5b9fa3d83b95f5606b31661a65c2f8fae1008 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5718/68e5b9fa3d83b95f5606b31661a65c2f8fae1008/Generator.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
666,
2278,
15114,
13332,
327,
288,
225,
315,
2425,
6,
294,
288,
225,
315,
723,
6,
294,
315,
46,
9937,
2278,
3113,
19879,
225,
315,
14676,
17,
10274,
17,
1376,
6,
294,
288,
315,
723,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
666,
2278,
15114,
13332,
327,
288,
225,
315,
2425,
6,
294,
288,
225,
315,
723,
6,
294,
315,
46,
9937,
2278,
3113,
19879,
225,
315,
14676,
17,
10274,
17,
1376,
6,
294,
288,
315,
723,
... |
), \ | ), | def is_newer(file1, file2): """ Return True if either file2 does not exist or is older than file1. If file1 does not exist, always return False. """ if not os.path.exists(file1): return False if not os.path.exists(file2): return True if os.path.getmtime(file2) < os.path.getmtime(file1): return True return False | 697d48ce28ff4d12606adf26a8fb095216c31678 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/697d48ce28ff4d12606adf26a8fb095216c31678/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
353,
67,
2704,
264,
12,
768,
21,
16,
585,
22,
4672,
3536,
2000,
1053,
309,
3344,
585,
22,
1552,
486,
1005,
578,
353,
12156,
2353,
585,
21,
18,
225,
971,
585,
21,
1552,
486,
1005,
16,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
353,
67,
2704,
264,
12,
768,
21,
16,
585,
22,
4672,
3536,
2000,
1053,
309,
3344,
585,
22,
1552,
486,
1005,
578,
353,
12156,
2353,
585,
21,
18,
225,
971,
585,
21,
1552,
486,
1005,
16,... |
atom.AtomBase._TakeAttributeFromElementTree(self, attribute, | atom.AtomBase._TakeAttributeFromElementTree(self, attribute, | def _TakeAttributeFromElementTree(self, attribute, element_tree): if attribute == 'value': self.value = element_tree.attrib[attribute] del element_tree.attrib[attribute] else: atom.AtomBase._TakeAttributeFromElementTree(self, attribute, element_tree) | 6a016768dbbd7d5ee9c62bcd2d2ce3bf7f6d1048 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6580/6a016768dbbd7d5ee9c62bcd2d2ce3bf7f6d1048/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
13391,
1499,
1265,
1046,
2471,
12,
2890,
16,
1566,
16,
930,
67,
3413,
4672,
309,
1566,
422,
296,
1132,
4278,
365,
18,
1132,
273,
930,
67,
3413,
18,
14588,
63,
4589,
65,
1464,
930,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
13391,
1499,
1265,
1046,
2471,
12,
2890,
16,
1566,
16,
930,
67,
3413,
4672,
309,
1566,
422,
296,
1132,
4278,
365,
18,
1132,
273,
930,
67,
3413,
18,
14588,
63,
4589,
65,
1464,
930,... |
mdpath = os.path.join(cmds['basedir'], cmds['olddir'], cmds['update-info-dir']) if os.path.isdir(mdpath): for file in os.listdir(mdpath): os.remove(os.path.join(mdpath, file)) os.rmdir(mdpath) | def main(args): cmds, directories = parseArgs(args) directory = directories[0] # start the sanity/stupidity checks if not os.path.exists(os.path.join(cmds['basedir'], directory)): errorprint(_('Directory must exist')) sys.exit(1) if not os.path.isdir(os.path.join(cmds['basedir'], directory)): errorprint(_('Directory of packages must be a directory.')) sys.exit(1) if not os.access(cmds['outputdir'], os.W_OK): errorprint(_('Directory must be writable.')) sys.exit(1) if cmds['split']: oldbase = cmds['basedir'] cmds['basedir'] = os.path.join(cmds['basedir'], directory) if not checkAndMakeDir(os.path.join(cmds['outputdir'], cmds['tempdir'])): sys.exit(1) if not checkAndMakeDir(os.path.join(cmds['outputdir'], cmds['finaldir'])): sys.exit(1) if os.path.exists(os.path.join(cmds['outputdir'], cmds['olddir'])): errorprint(_('Old data directory exists, please remove: %s') % cmds['olddir']) sys.exit(1) if cmds.has_key('update-info-location'): if not checkAndMakeDir(os.path.join(cmds['basedir'], cmds['tempdir'], cmds['update-info-dir'])): errorprint(_('Error: cannot open/write to update info dir %s' % a)) usage() # make sure we can write to where we want to write to: for direc in ['tempdir', 'finaldir']: for file in ['primaryfile', 'filelistsfile', 'otherfile', 'repomdfile']: filepath = os.path.join(cmds['outputdir'], cmds[direc], cmds[file]) if os.path.exists(filepath): if not os.access(filepath, os.W_OK): errorprint(_('error in must be able to write to metadata files:\n -> %s') % filepath) usage() if cmds['split']: cmds['basedir'] = oldbase mdgen = SplitMetaDataGenerator(cmds) mdgen.doPkgMetadata(directories) else: mdgen = MetaDataGenerator(cmds) mdgen.doPkgMetadata(directory) mdgen.doRepoMetadata() if os.path.exists(os.path.join(cmds['outputdir'], cmds['finaldir'])): try: os.rename(os.path.join(cmds['outputdir'], cmds['finaldir']), os.path.join(cmds['outputdir'], cmds['olddir'])) except: errorprint(_('Error moving final %s to old dir %s' % (os.path.join(cmds['outputdir'], cmds['finaldir']), os.path.join(cmds['outputdir'], cmds['olddir'])))) sys.exit(1) try: os.rename(os.path.join(cmds['outputdir'], cmds['tempdir']), os.path.join(cmds['outputdir'], cmds['finaldir'])) except: errorprint(_('Error moving final metadata into place')) # put the old stuff back os.rename(os.path.join(cmds['outputdir'], cmds['olddir']), os.path.join(cmds['outputdir'], cmds['finaldir'])) sys.exit(1) for file in ['primaryfile', 'filelistsfile', 'otherfile', 'repomdfile', 'groupfile']: if cmds[file]: fn = os.path.basename(cmds[file]) else: continue oldfile = os.path.join(cmds['outputdir'], cmds['olddir'], fn) if os.path.exists(oldfile): try: os.remove(oldfile) except OSError, e: errorprint(_('Could not remove old metadata file: %s') % oldfile) errorprint(_('Error was %s') % e) sys.exit(1) # Clean up any update metadata mdpath = os.path.join(cmds['basedir'], cmds['olddir'], cmds['update-info-dir']) if os.path.isdir(mdpath): for file in os.listdir(mdpath): os.remove(os.path.join(mdpath, file)) os.rmdir(mdpath) | 69aebff0d9b62beebdfeb44f7aecb7234afcdfd1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9557/69aebff0d9b62beebdfeb44f7aecb7234afcdfd1/genpkgmetadata.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
1968,
4672,
13446,
16,
6402,
273,
1109,
2615,
12,
1968,
13,
1867,
273,
6402,
63,
20,
65,
468,
787,
326,
16267,
19,
334,
416,
24237,
4271,
309,
486,
1140,
18,
803,
18,
1808,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
1968,
4672,
13446,
16,
6402,
273,
1109,
2615,
12,
1968,
13,
1867,
273,
6402,
63,
20,
65,
468,
787,
326,
16267,
19,
334,
416,
24237,
4271,
309,
486,
1140,
18,
803,
18,
1808,
... | |
shear_file=None): | shear_file=None, outdir=None, mohrify=False): | def MakeCommand(executable, conf_file, image_file, cat_file=None, stars_file=None, psf_file=None, fitpsf_file=None, shear_file=None): #root, input_prefix = MakeRoot(image_file) root = MakeRoot(image_file) command = \ [executable, conf_file, 'root='+root] #if input_prefix != '': # command.append('input_prefix='+input_prefix) if cat_file is not None: command.append('cat_file='+cat_file) if stars_file is not None: command.append('stars_file='+stars_file) if psf_file is not None: command.append('psf_file='+psf_file) if fitpsf_file is not None: command.append('fitpsf_file='+fitpsf_file) if shear_file is not None: command.append('shear_file='+shear_file) return command | f2cd5c779583c2f63b6d536dbfab856040b9817b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6617/f2cd5c779583c2f63b6d536dbfab856040b9817b/wl-run.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4344,
2189,
12,
17751,
16,
2195,
67,
768,
16,
1316,
67,
768,
16,
6573,
67,
768,
33,
7036,
16,
24084,
67,
768,
33,
7036,
16,
28687,
67,
768,
33,
7036,
16,
4845,
1121,
74,
67,
768,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4344,
2189,
12,
17751,
16,
2195,
67,
768,
16,
1316,
67,
768,
16,
6573,
67,
768,
33,
7036,
16,
24084,
67,
768,
33,
7036,
16,
28687,
67,
768,
33,
7036,
16,
4845,
1121,
74,
67,
768,
3... |
for tablita in parsed: if project == "wikinews": if lang == "it": logg = 'Wikinotizie:Benvenuto log' welcom = "{{subst:benvenuto|%s}}" sign = 'Tooby' elif project == 'wiktionaty': if lang == 'it': logg = 'Wikizionario:Benvenuto log' welcom = '{{Utente:Filnik/Benve|nome={{subst:PAGENAME}}}} %s' | for found_result in parsed: | def DefineSign(welcomesite, SignPage): Page = wikipedia.Page(welcomesite, SignPage) SignText = Page.get() reg = "\* (.*?)\n" list_sign = re.findall(reg, SignText) return list_sign | bf50a20b7e53b37ac1a5a7ffeb2895c56bcf1f38 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/4404/bf50a20b7e53b37ac1a5a7ffeb2895c56bcf1f38/welcome.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13184,
2766,
12,
91,
292,
10127,
1137,
16,
4383,
1964,
4672,
3460,
273,
21137,
18,
1964,
12,
91,
292,
10127,
1137,
16,
4383,
1964,
13,
4383,
1528,
273,
3460,
18,
588,
1435,
960,
273,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13184,
2766,
12,
91,
292,
10127,
1137,
16,
4383,
1964,
4672,
3460,
273,
21137,
18,
1964,
12,
91,
292,
10127,
1137,
16,
4383,
1964,
13,
4383,
1528,
273,
3460,
18,
588,
1435,
960,
273,
1... |
self._do('tag_raise', aboveThis) | self._do('raise', aboveThis) | def tkraise(self, aboveThis=None): self._do('tag_raise', aboveThis) | 6678cbe7be6a7b9da83ce546bb9042448533cbbe /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/6678cbe7be6a7b9da83ce546bb9042448533cbbe/Canvas.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13030,
11628,
12,
2890,
16,
5721,
2503,
33,
7036,
4672,
365,
6315,
2896,
2668,
2692,
67,
11628,
2187,
5721,
2503,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13030,
11628,
12,
2890,
16,
5721,
2503,
33,
7036,
4672,
365,
6315,
2896,
2668,
2692,
67,
11628,
2187,
5721,
2503,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
jQuery(" return(msg); }, GoodbyeWorld: function (name) { return("Goodbye " + name); } } ] """ | c.parse_consumes_field() chunks.append(c) if i >= length: print "ERROR: Unexpected EOF: extension file %s has mismatched braces (ignoring file)" % filename return None start = js.find('{', end) return chunks | def __init__(self, debug=False): self.__debug = debug # if debug is False condense javascript to optimize self.__extension_cache = {} #lets put a dummy in for now to test self.__extension_cache['build_message'] = """ [{ info: {consumes:['build_message'], author: 'John (J5) Palmieri <johnp@redhat.com>', version: '0.1', name: 'Hello World Message' }, run: function (data) { msg = this.HelloWorld(data.task_id); jQuery("#" + data.uid).css('background-color', 'blue'); return(msg); }, HelloWorld: function (name) { return("Hello " + name); } }, { info: {consumes:['build_message'], author: 'John (J5) Palmieri <johnp@redhat.com>', version: '0.1', name: 'Goodbye World Message' }, run: function (data) { msg = this.GoodbyeWorld(data.task_id); jQuery("#" + data.uid).css('background-color', 'red'); return(msg); }, GoodbyeWorld: function (name) { return("Goodbye " + name); } } ] """ | 27ec9ed03663e715b7c0422359e98616637fde6d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12732/27ec9ed03663e715b7c0422359e98616637fde6d/extensions.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1198,
33,
8381,
4672,
365,
16186,
4148,
273,
1198,
468,
309,
1198,
353,
1083,
6941,
3558,
11341,
358,
10979,
365,
16186,
6447,
67,
2493,
273,
2618,
225,
46... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1198,
33,
8381,
4672,
365,
16186,
4148,
273,
1198,
468,
309,
1198,
353,
1083,
6941,
3558,
11341,
358,
10979,
365,
16186,
6447,
67,
2493,
273,
2618,
225,
46... |
module = __import__(client_name + ".common_lib." + library, globals(), locals(), [""]) | module = __import__(client_name + ".common_lib", globals(), locals(), [library]) module = getattr(module, library) | def _setup_common_library(root_module_name): """ Setup aliases for all the common libraries, e.g. common -> autotest_lib.client.common_lib common.error -> autotest_lib.client.common_lib.error """ # convert the root_module_name into a client module name parts = root_module_name.split(".") if parts[-1] == "client": client_name = root_module_name else: client_name = root_module_name + ".client" # import the top-level common library common_lib = __import__(client_name + ".common_lib", globals(), locals(), [""]) sys.modules["common"] = common_lib # patch up all the root_module_name.*.common libs for module_name in sys.modules.iterkeys(): if (module_name.startswith(root_module_name + ".") and module_name.endswith(".common")): sys.modules[module_name] = common_lib # import the specific common libraries for library in common_lib.__all__: module = __import__(client_name + ".common_lib." + library, globals(), locals(), [""]) setattr(common_lib, library, module) sys.modules["common.%s" % library] = module | 554f19bb46af6e7cb74f4dd0e6421df900b0d2bc /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12268/554f19bb46af6e7cb74f4dd0e6421df900b0d2bc/setup_modules.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8401,
67,
6054,
67,
12083,
12,
3085,
67,
2978,
67,
529,
4672,
3536,
10939,
6900,
364,
777,
326,
2975,
14732,
16,
425,
18,
75,
18,
2975,
317,
2059,
352,
395,
67,
2941,
18,
2625,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8401,
67,
6054,
67,
12083,
12,
3085,
67,
2978,
67,
529,
4672,
3536,
10939,
6900,
364,
777,
326,
2975,
14732,
16,
425,
18,
75,
18,
2975,
317,
2059,
352,
395,
67,
2941,
18,
2625,
... |
subpanel.pack(side = TOP, fill = BOTH, expand = 1) | subpanel.grid(sticky=N+S+E+W) panel.rowconfigure(0, weight=1) panel.columnconfigure(0, weight=1) subpanel.rowconfigure(0, weight=1) subpanel.columnconfigure(0, weight=1) | def build_tab_authors(self): panel = TFrame(self.root, style='FlatFrame', borderwidth = 5) self.nb.add(panel, text='Authors') subpanel = TFrame(panel, style='RoundedFrame', borderwidth = 4) subpanel.pack(side = TOP, fill = BOTH, expand = 1) sb=TScrollbar(subpanel) sb.pack(side = RIGHT, fill=Y, expand = 1) text=Text(subpanel, bg='white',highlightthickness=0) text.pack(side = LEFT, fill = BOTH, expand = 1) text['yscrollcommand'] = sb.set sb['command'] = text.yview txt=_('\nTo report bugs please use project bugtracker: https://bugs.launchpad.net/sk1/.\n\n') txt+=_('Igor Novikov\n igor.e.novikov@gmail.com\n Project Leader\n\n') txt+=_('Acknowledgments:\n=======================================\n') txt+=_('Valek Fillipov\n CDR format reverse engineering\n\n') txt+=_('Alexandre Prokoudine\n alexandre.prokoudine@gmail.com\n Information support (http://linuxgraphics.ru)') text['state']=NORMAL text.insert(END, txt) text['state']=DISABLED | f97d4ae13dc3170337eb5ab568147410c3748b0b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/3123/f97d4ae13dc3170337eb5ab568147410c3748b0b/aboutdlg.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1361,
67,
7032,
67,
19368,
12,
2890,
4672,
6594,
273,
399,
3219,
12,
2890,
18,
3085,
16,
2154,
2218,
16384,
3219,
2187,
5795,
2819,
273,
1381,
13,
365,
18,
6423,
18,
1289,
12,
13916,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1361,
67,
7032,
67,
19368,
12,
2890,
4672,
6594,
273,
399,
3219,
12,
2890,
18,
3085,
16,
2154,
2218,
16384,
3219,
2187,
5795,
2819,
273,
1381,
13,
365,
18,
6423,
18,
1289,
12,
13916,
1... |
print "Breakpoint in", filename, "at", lineno | def set_break(self, filename, lineno, temporary=0, cond = None): filename = self.canonic(filename) import linecache # Import as late as possible line = linecache.getline(filename, lineno) if not line: return 'Line %s:%d does not exist' % (filename, lineno) if not self.breaks.has_key(filename): self.breaks[filename] = [] list = self.breaks[filename] if not lineno in list: list.append(lineno) bp = Breakpoint(filename, lineno, temporary, cond) print "Breakpoint in", filename, "at", lineno | 6c500e53391ae0febda8e0e2dd201372f6ba876f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/6c500e53391ae0febda8e0e2dd201372f6ba876f/bdb.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
67,
8820,
12,
2890,
16,
1544,
16,
7586,
16,
6269,
33,
20,
16,
6941,
273,
599,
4672,
1544,
273,
365,
18,
4169,
15506,
12,
3459,
13,
1930,
4739,
557,
807,
468,
6164,
487,
26374,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
67,
8820,
12,
2890,
16,
1544,
16,
7586,
16,
6269,
33,
20,
16,
6941,
273,
599,
4672,
1544,
273,
365,
18,
4169,
15506,
12,
3459,
13,
1930,
4739,
557,
807,
468,
6164,
487,
26374,
4... | |
raise TypeError, "Do not know how to handle given type: %s" %\ o_descr | raise TypeError("Do not know how to handle given type: %s" % \ o_descr) | def rebin_axis_1D(obj,axis_out): """ This function rebins the primary axis for a SOM or a SO based on the given NessiList axis. Parameters: ---------- -> obj is the SOM or SO to be rebinned -> axis_out is a NessiList containing the axis to rebin the SOM or SO to Returns: ------- <- A SOM or SO that has been rebinned according to the provided axis Exceptions: ---------- <- TypeError is raised if the rebinning axis given is not a NessiList <- TypeError is raised if object being rebinned is not a SOM or a SO """ # import the helper functions import hlr_utils # set up for working through data try: axis_out.__type__ except AttributeError: raise TypeError, "Rebinning axis must be a NessiList!" (o_descr,d_descr)=hlr_utils.get_descr(obj) if o_descr == "number" or o_descr == "list": raise TypeError, "Do not know how to handle given type: %s" %\ o_descr else: pass (result,res_descr)=hlr_utils.empty_result(obj) result=hlr_utils.copy_som_attr(result,res_descr,obj,o_descr) # iterate through the values import axis_manip for i in range(hlr_utils.get_length(obj)): axis_in = hlr_utils.get_value(obj,i,o_descr,"x",0) val = hlr_utils.get_value(obj,i,o_descr) err2 = hlr_utils.get_err2(obj,i,o_descr) value=axis_manip.rebin_axis_1D(axis_in, val, err2, axis_out) xvals=[] xvals.append(axis_out) map_so = hlr_utils.get_map_so(obj,None,i) hlr_utils.result_insert(result,res_descr,value,map_so,"all",0,xvals) return result | 766ac4ad90bc5477427ed44f10694c1d5e7af5bd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/763/766ac4ad90bc5477427ed44f10694c1d5e7af5bd/hlr_rebin_axis_1D.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
283,
4757,
67,
4890,
67,
21,
40,
12,
2603,
16,
4890,
67,
659,
4672,
3536,
1220,
445,
283,
11862,
326,
3354,
2654,
364,
279,
348,
1872,
578,
279,
7460,
2511,
603,
326,
864,
423,
403,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
283,
4757,
67,
4890,
67,
21,
40,
12,
2603,
16,
4890,
67,
659,
4672,
3536,
1220,
445,
283,
11862,
326,
3354,
2654,
364,
279,
348,
1872,
578,
279,
7460,
2511,
603,
326,
864,
423,
403,
... |
A dict containing the result of the retrieve operation. | A list containing the result of the retrieve operation. | def RetrieveAllOwners(self, group_id): """Retrieve all owners of the given group | a4313b2e95974b9e4449afa950498d08d2e34e61 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5609/a4313b2e95974b9e4449afa950498d08d2e34e61/service.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10708,
1595,
5460,
414,
12,
2890,
16,
1041,
67,
350,
4672,
3536,
5767,
777,
25937,
434,
326,
864,
1041,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10708,
1595,
5460,
414,
12,
2890,
16,
1041,
67,
350,
4672,
3536,
5767,
777,
25937,
434,
326,
864,
1041,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
settings_defaults = {} | settings_defaults = {'output_encoding': 'latin-1'} | def setMode(self, mode): self.mode = mode | ed688792a0f4c20b7802333819981a0ad813078d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5620/ed688792a0f4c20b7802333819981a0ad813078d/python_latex.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
2309,
12,
2890,
16,
1965,
4672,
365,
18,
3188,
273,
1965,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
2309,
12,
2890,
16,
1965,
4672,
365,
18,
3188,
273,
1965,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
elem(0, "X", "Singlet", 0.001, 1.1, 1, 0), elem(1, "H", "Hydrogen", 1.6737, 1.2, 1, 30), | elem(0, "X", "Singlet", 17.000, 1.1, 1, 0), elem(1, "H", "Hydrogen", 1.6737, 0.77, 1, 30), | def prec(self): # /* mass rvdw evdw nbonds symbol */ # {0.100, 0.5, 0.130, 1, "X"}, /* 0 Singlet */ evd = evdw.get(self.symbol, 0.3+self.eltnum**2/190.0) lis = self.mass, self.rvdw, evd, self.bonds, self.symbol rec = ' {%7.3f, %5.2f, %6.3f, %d, "%s"}, ' % lis print rec+' //', self.name | 963f3ffaed3af052cd185d79968e60143c11b8b3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/963f3ffaed3af052cd185d79968e60143c11b8b3/gentab.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13382,
12,
2890,
4672,
468,
1748,
8039,
225,
5633,
22512,
282,
2113,
22512,
4264,
4096,
3273,
1195,
468,
288,
20,
18,
6625,
16,
225,
374,
18,
25,
16,
225,
374,
18,
24417,
16,
404,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13382,
12,
2890,
4672,
468,
1748,
8039,
225,
5633,
22512,
282,
2113,
22512,
4264,
4096,
3273,
1195,
468,
288,
20,
18,
6625,
16,
225,
374,
18,
25,
16,
225,
374,
18,
24417,
16,
404,
16,
... |
query.exclude(status__in=tko_rpc_utils._INVALID_STATUSES) | query = query.exclude(status__in=tko_rpc_utils._INVALID_STATUSES) | def get_latest_tests(group_by, header_groups=[], fixed_headers={}, machine_label_headers={}, extra_info=[], **filter_data): """ Similar to get_status_counts, but return only the latest test result per group. It still returns the same information (i.e. with pass count etc.) for compatibility. @param extra_info a list containing the field names that should be returned with each cell. The fields are returned in the extra_info field of the return dictionary. """ # find latest test per group query = models.TestView.objects.get_query_set_with_joins( filter_data, include_host_labels=bool(machine_label_headers)) query = models.TestView.query_objects(filter_data, initial_query=query, apply_presentation=False) query.exclude(status__in=tko_rpc_utils._INVALID_STATUSES) query = query.extra( select={'latest_test_idx' : 'MAX(%s)' % models.TestView.objects.get_key_on_this_table('test_idx')}) query = tko_rpc_utils.add_machine_label_headers(machine_label_headers, query) query = models.TestView.apply_presentation(query, filter_data) group_processor = tko_rpc_utils.GroupDataProcessor(query, group_by, header_groups, fixed_headers) group_processor.process_group_dicts() info = group_processor.get_info_dict() # fetch full info for these tests so we can access their statuses all_test_ids = [group['latest_test_idx'] for group in info['groups']] test_views = models.TestView.objects.in_bulk(all_test_ids) for group_dict in info['groups']: test_idx = group_dict.pop('latest_test_idx') group_dict['test_idx'] = test_idx test_view = test_views[test_idx] tko_rpc_utils.add_status_counts(group_dict, test_view.status) group_dict['extra_info'] = [] for field in extra_info: group_dict['extra_info'].append(getattr(test_view, field)) return rpc_utils.prepare_for_serialization(info) | 883108d29f533d60828037392b03b2b0133b55b4 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12268/883108d29f533d60828037392b03b2b0133b55b4/rpc_interface.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
13550,
67,
16341,
12,
1655,
67,
1637,
16,
1446,
67,
4650,
22850,
6487,
5499,
67,
2485,
28793,
5228,
67,
1925,
67,
2485,
28793,
2870,
67,
1376,
22850,
6487,
2826,
2188,
67,
892,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
13550,
67,
16341,
12,
1655,
67,
1637,
16,
1446,
67,
4650,
22850,
6487,
5499,
67,
2485,
28793,
5228,
67,
1925,
67,
2485,
28793,
2870,
67,
1376,
22850,
6487,
2826,
2188,
67,
892,
... |
blobs = np.array(objlist, dtype=np.object_) | blobs = np.empty(len(objlist), dtype=object) blobs[:] = objlist | def _smart_load_blobs(self, barray, refs): """ Load an ndarray of BLOBs from a set of refs refs, taking into account not to instantiate duplicate objects for the same BLOBs. The input array of refs must be one-dimensional. The output is a 1D array of blobs, corresponding to the refs. """ ##return np.ones(len(refs), dtype=object); assert len(refs.shape) == 1 | aa7de6b45c8cb85cb9ae180387eb122922f7d58e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/13091/aa7de6b45c8cb85cb9ae180387eb122922f7d58e/catalog.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
26416,
67,
945,
67,
10721,
87,
12,
2890,
16,
324,
1126,
16,
9047,
4672,
3536,
4444,
392,
8304,
434,
605,
6038,
87,
628,
279,
444,
434,
9047,
9047,
16,
13763,
1368,
2236,
486,
358,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
26416,
67,
945,
67,
10721,
87,
12,
2890,
16,
324,
1126,
16,
9047,
4672,
3536,
4444,
392,
8304,
434,
605,
6038,
87,
628,
279,
444,
434,
9047,
9047,
16,
13763,
1368,
2236,
486,
358,... |
self.doc_as_attr = (self.objpath[-1] != self.object.__name__) | if hasattr(self.object, '__name__'): self.doc_as_attr = (self.objpath[-1] != self.object.__name__) else: self.doc_as_attr = True | def import_object(self): ret = ModuleLevelDocumenter.import_object(self) # if the class is documented under another name, document it # as data/attribute if ret: self.doc_as_attr = (self.objpath[-1] != self.object.__name__) return ret | 48c4bf0cb388b8960e49639d19fb45c5880b6952 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5532/48c4bf0cb388b8960e49639d19fb45c5880b6952/autodoc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1930,
67,
1612,
12,
2890,
4672,
325,
273,
5924,
2355,
2519,
264,
18,
5666,
67,
1612,
12,
2890,
13,
468,
309,
326,
667,
353,
23051,
3613,
4042,
508,
16,
1668,
518,
468,
487,
501,
19,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1930,
67,
1612,
12,
2890,
4672,
325,
273,
5924,
2355,
2519,
264,
18,
5666,
67,
1612,
12,
2890,
13,
468,
309,
326,
667,
353,
23051,
3613,
4042,
508,
16,
1668,
518,
468,
487,
501,
19,
... |
ui.hide(self.searchbutton) self.searchtext.set_text("") self.library_browse(root=self.wd) if move_focus: self.library.grab_focus() | if self.library_search_visible(): self.libsearchfilter_toggle(move_focus) | def on_library_search_end(self, button, move_focus=True): ui.hide(self.searchbutton) self.searchtext.set_text("") self.library_browse(root=self.wd) if move_focus: self.library.grab_focus() | dfbaf3cf117d4b81ab04713e00bb504fb7791def /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/2312/dfbaf3cf117d4b81ab04713e00bb504fb7791def/main.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
67,
12083,
67,
3072,
67,
409,
12,
2890,
16,
3568,
16,
3635,
67,
13923,
33,
5510,
4672,
5915,
18,
11248,
12,
2890,
18,
3072,
5391,
13,
365,
18,
3072,
955,
18,
542,
67,
955,
2932,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
67,
12083,
67,
3072,
67,
409,
12,
2890,
16,
3568,
16,
3635,
67,
13923,
33,
5510,
4672,
5915,
18,
11248,
12,
2890,
18,
3072,
5391,
13,
365,
18,
3072,
955,
18,
542,
67,
955,
2932,... |
keys = self.keys() | def matvec(self, other): other = asarray(other) if other.shape[0] != self.shape[1]: raise ValueError, "Dimensions do not match." keys = self.keys() res = [0]*self.shape[0] for key in keys: res[int(key[0])] += self[key] * other[int(key[1]),...] return array(res) | a2bf57fc75113a68b861da7355c48c019c1ddc40 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12971/a2bf57fc75113a68b861da7355c48c019c1ddc40/sparse.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4834,
8799,
12,
2890,
16,
1308,
4672,
1308,
273,
10455,
12,
3011,
13,
309,
1308,
18,
4867,
63,
20,
65,
480,
365,
18,
4867,
63,
21,
14542,
1002,
2068,
16,
315,
10796,
741,
486,
845,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4834,
8799,
12,
2890,
16,
1308,
4672,
1308,
273,
10455,
12,
3011,
13,
309,
1308,
18,
4867,
63,
20,
65,
480,
365,
18,
4867,
63,
21,
14542,
1002,
2068,
16,
315,
10796,
741,
486,
845,
1... | |
if resp[:3] <> '229': | if resp[:3] != '229': | def parse229(resp, peer): '''Parse the '229' response for a EPSV request. Raises error_proto if it does not contain '(|||port|)' Return ('host.addr.as.numbers', port#) tuple.''' if resp[:3] <> '229': raise error_reply, resp left = resp.find('(') if left < 0: raise error_proto, resp right = resp.find(')', left + 1) if right < 0: raise error_proto, resp # should contain '(|||port|)' if resp[left + 1] <> resp[right - 1]: raise error_proto, resp parts = resp[left + 1:right].split(resp[left+1]) if len(parts) <> 5: raise error_proto, resp host = peer[0] port = int(parts[3]) return host, port | 0c5eac883ea9d53f54e55ccf224c50b7b2951488 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/0c5eac883ea9d53f54e55ccf224c50b7b2951488/ftplib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
3787,
29,
12,
12243,
16,
4261,
4672,
9163,
3201,
326,
296,
3787,
29,
11,
766,
364,
279,
512,
5857,
58,
590,
18,
20497,
555,
67,
9393,
309,
518,
1552,
486,
912,
7747,
20081,
96,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
3787,
29,
12,
12243,
16,
4261,
4672,
9163,
3201,
326,
296,
3787,
29,
11,
766,
364,
279,
512,
5857,
58,
590,
18,
20497,
555,
67,
9393,
309,
518,
1552,
486,
912,
7747,
20081,
96,
... |
if not match.group('title') or page.site().isInterwikiLink(match.group('title')) or match.group('section'): | if not match.group('title') \ or page.site().isInterwikiLink(match.group('title')) \ or match.group('section'): | def handleNextLink(self, page, text, match, context = 100): """ Returns a tuple (text, jumpToBeginning). text is the unicode string after the current link has been processed. jumpToBeginning is a boolean which specifies if the cursor position should be reset to 0. This is required after the user has edited the article. """ # ignore interwiki links and links to sections of the same page as well as section links if not match.group('title') or page.site().isInterwikiLink(match.group('title')) or match.group('section'): return text, False | 18d0a408a44b825cb28c4d9a9562e9286becf5e2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4404/18d0a408a44b825cb28c4d9a9562e9286becf5e2/selflink.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
2134,
2098,
12,
2890,
16,
1363,
16,
977,
16,
845,
16,
819,
273,
2130,
4672,
3536,
2860,
279,
3193,
261,
955,
16,
11833,
774,
8149,
2093,
2934,
977,
353,
326,
5252,
533,
1839,
326... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
2134,
2098,
12,
2890,
16,
1363,
16,
977,
16,
845,
16,
819,
273,
2130,
4672,
3536,
2860,
279,
3193,
261,
955,
16,
11833,
774,
8149,
2093,
2934,
977,
353,
326,
5252,
533,
1839,
326... |
ui.widget.set_active(font.justify == "fill")) | update_toggle(ui, font.justify == "fill")) | def update(ui, font): widget = ui.widget.get_proxies()[0] widget.set_active(font.par_type == "bullet") | f3c33abf52325609f58edfa6e4d18ba3e6c6ed23 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4738/f3c33abf52325609f58edfa6e4d18ba3e6c6ed23/editor.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
12,
4881,
16,
3512,
4672,
3604,
273,
5915,
18,
6587,
18,
588,
67,
20314,
606,
1435,
63,
20,
65,
3604,
18,
542,
67,
3535,
12,
5776,
18,
1065,
67,
723,
422,
315,
70,
19994,
7923,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
12,
4881,
16,
3512,
4672,
3604,
273,
5915,
18,
6587,
18,
588,
67,
20314,
606,
1435,
63,
20,
65,
3604,
18,
542,
67,
3535,
12,
5776,
18,
1065,
67,
723,
422,
315,
70,
19994,
7923,... |
MIME_TYPE = 'application/vnd.oasis.opendocument.text' | def getparent(self): return self.parent | d758e0c8ec3cf1e09ca98cd6ef2658cd7802279e /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1278/d758e0c8ec3cf1e09ca98cd6ef2658cd7802279e/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
2938,
12,
2890,
4672,
327,
365,
18,
2938,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
2938,
12,
2890,
4672,
327,
365,
18,
2938,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... | |
if re.search('UNAVAILABLE',urlLocality): | if re.search( 'UNAVAILABLE', urlLocality ): | def __parse_file_metadata(self,urlDict): statDict = self.__parse_stat(urlDict['stat']) if statDict['File']: statDict['Checksum'] = '' if urlDict.has_key('checksum') and (urlDict['checksum'] != '0x'): statDict['Checksum'] = urlDict['checksum'] if urlDict.has_key('locality'): urlLocality = urlDict['locality'] if re.search('ONLINE',urlLocality): statDict['Cached'] = 1 else: statDict['Cached'] = 0 if re.search('NEARLINE',urlLocality): statDict['Migrated'] = 1 else: statDict['Migrated'] = 0 statDict['Lost'] = 0 if re.search('LOST',urlLocality): statDict['Lost'] = 1 statDict['Unavailable'] = 0 if re.search('UNAVAILABLE',urlLocality): statDict['Unavailable'] = 1 return statDict | 70e66af095cb6701e39b1e701e4a2ce4d012b4f7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/70e66af095cb6701e39b1e701e4a2ce4d012b4f7/SRM2Storage.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2670,
67,
768,
67,
4165,
12,
2890,
16,
718,
5014,
4672,
610,
5014,
273,
365,
16186,
2670,
67,
5642,
12,
718,
5014,
3292,
5642,
19486,
309,
610,
5014,
3292,
812,
3546,
30,
610,
50... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2670,
67,
768,
67,
4165,
12,
2890,
16,
718,
5014,
4672,
610,
5014,
273,
365,
16186,
2670,
67,
5642,
12,
718,
5014,
3292,
5642,
19486,
309,
610,
5014,
3292,
812,
3546,
30,
610,
50... |
print str | logging.debug(str) | def __debug_print(self, str1, str2=""): """ Nicely print two strings and an arrow. | 0ae821998f2c2f979b34b606091ddc8d892b33a9 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12268/0ae821998f2c2f979b34b606091ddc8d892b33a9/kvm_config.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
4148,
67,
1188,
12,
2890,
16,
609,
21,
16,
609,
22,
1546,
6,
4672,
3536,
423,
335,
2357,
1172,
2795,
2064,
471,
392,
12274,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
4148,
67,
1188,
12,
2890,
16,
609,
21,
16,
609,
22,
1546,
6,
4672,
3536,
423,
335,
2357,
1172,
2795,
2064,
471,
392,
12274,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
self._dual = Cone(rays, lattice=self.lattice().dual(), check=False) | self._dual = Cone(rays, lattice=self.dual_lattice(), check=False) | def dual(self): r""" Return the dual cone of ``self``. | a3e3306597435e3589001abd0b89714b99258af9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9417/a3e3306597435e3589001abd0b89714b99258af9/cone.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
24557,
12,
2890,
4672,
436,
8395,
2000,
326,
24557,
356,
73,
434,
12176,
2890,
68,
8338,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
24557,
12,
2890,
4672,
436,
8395,
2000,
326,
24557,
356,
73,
434,
12176,
2890,
68,
8338,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
assert u"%r, %r" % (u"abc", "abc") == u"u'abc', 'abc'" | value = u"%r, %r" % (u"abc", "abc") if value != u"u'abc', 'abc'": print '*** formatting failed for "%s"' % 'u"%r, %r" % (u"abc", "abc")' | test('capwords', u'abc\t def \nghi', u'Abc Def Ghi') | 8e118135be00ebd40ca7b87637cac566657008ec /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/8e118135be00ebd40ca7b87637cac566657008ec/test_unicode.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1842,
2668,
5909,
3753,
2187,
582,
11,
18947,
64,
88,
282,
1652,
225,
521,
3368,
12266,
2187,
582,
11,
5895,
71,
10922,
611,
12266,
6134,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1842,
2668,
5909,
3753,
2187,
582,
11,
18947,
64,
88,
282,
1652,
225,
521,
3368,
12266,
2187,
582,
11,
5895,
71,
10922,
611,
12266,
6134,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
self.keywords = ['HOSTID', 'RUNID', 'SLICEID', 'LEVEL', 'LOG', 'DATE', 'NODE', 'TIMESTAMP', 'COMMENT', 'STATUS', 'pipeline'] | self.keywords = ['HOSTID', 'RUNID', 'SLICEID', 'LEVEL', 'LOG', 'DATE', 'NODE', 'TIMESTAMP', 'COMMENT', 'STATUS', 'PIPELINE'] | def __init__(self, dbHostName, portNumber): MySQLBase.__init__(self, dbHostName, portNumber) self.keywords = ['HOSTID', 'RUNID', 'SLICEID', 'LEVEL', 'LOG', 'DATE', 'NODE', 'TIMESTAMP', 'COMMENT', 'STATUS', 'pipeline'] self.keywordSet = set(self.keywords) | bf9ea936ed2a244ec8736465ad45cb4fa0a434f5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6552/bf9ea936ed2a244ec8736465ad45cb4fa0a434f5/DatabaseLogger.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1319,
20946,
16,
1756,
1854,
4672,
13485,
2171,
16186,
2738,
972,
12,
2890,
16,
1319,
20946,
16,
1756,
1854,
13,
365,
18,
11771,
273,
10228,
8908,
734,
218... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1319,
20946,
16,
1756,
1854,
4672,
13485,
2171,
16186,
2738,
972,
12,
2890,
16,
1319,
20946,
16,
1756,
1854,
13,
365,
18,
11771,
273,
10228,
8908,
734,
218... |
A larger example, illustrating some precision issues. The default algorithm uses Pari and makes sure the result has the requested precision. Note that Pari may complain if the requested precision is too low: | A larger example. The default algorithm uses Pari and makes sure the result has the requested precision. | def elliptic_logarithm(self, embedding=None, precision=100, algorithm='pari'): """ Returns the elliptic logarithm of this elliptic curve point. | a9b66a2ea206d0cbe3c288b4e2cd2235a0a8c7e6 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/a9b66a2ea206d0cbe3c288b4e2cd2235a0a8c7e6/ell_point.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
415,
549,
21507,
67,
1330,
31249,
12,
2890,
16,
15853,
33,
7036,
16,
6039,
33,
6625,
16,
4886,
2218,
1065,
77,
11,
4672,
3536,
2860,
326,
415,
549,
21507,
613,
31249,
434,
333,
415,
54... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
415,
549,
21507,
67,
1330,
31249,
12,
2890,
16,
15853,
33,
7036,
16,
6039,
33,
6625,
16,
4886,
2218,
1065,
77,
11,
4672,
3536,
2860,
326,
415,
549,
21507,
613,
31249,
434,
333,
415,
54... |
blockhead = self.expr_compile(s1) | cond.append(self.expr_compile(s1)) | def compile(self,s = None): " main parsing and compilation of tdl statements" # print "compile <%s>" % s, s,key = self.get_next_statement(s=s) # print " :key <%s>" % key # print " :s <%s>" % s if s in ('','#'): return None if s.startswith('#'): return None if s == self.EOF: return [self.EOF] ret = [] # these keywords can never legally occur at a top-level compilation # and will always be found while the appropriate (if,while,for,def) # block is being processed if key in ('else','elif','endif','endfor','endwhile','enddef','endtry'): raise EvalException, 'syntax error: %s' % key | d0335fdee64e7efe121c90260271a4efee99ab7e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11796/d0335fdee64e7efe121c90260271a4efee99ab7e/Eval.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4074,
12,
2890,
16,
87,
273,
599,
4672,
315,
2774,
5811,
471,
8916,
434,
268,
5761,
6317,
6,
468,
1172,
315,
11100,
23930,
87,
2984,
738,
225,
272,
16,
272,
16,
856,
273,
365,
18,
58... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4074,
12,
2890,
16,
87,
273,
599,
4672,
315,
2774,
5811,
471,
8916,
434,
268,
5761,
6317,
6,
468,
1172,
315,
11100,
23930,
87,
2984,
738,
225,
272,
16,
272,
16,
856,
273,
365,
18,
58... |
bundleRoot = 'Democracy.app/Contents' | bundleRoot = os.path.join(self.dist_dir, 'Democracy.app/Contents') | def run(self): global root, bundleRoot, rsrcRoot template_compiler.compileAllTemplates(root) py2app.run(self) # Setup some variables we'll need | 89a6271596a9ca34438750e9734456de052cbcb0 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12354/89a6271596a9ca34438750e9734456de052cbcb0/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
2552,
1365,
16,
3440,
2375,
16,
3597,
1310,
2375,
1542,
67,
9576,
18,
11100,
1595,
8218,
12,
3085,
13,
2395,
22,
2910,
18,
2681,
12,
2890,
13,
468,
10939,
2690,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
2552,
1365,
16,
3440,
2375,
16,
3597,
1310,
2375,
1542,
67,
9576,
18,
11100,
1595,
8218,
12,
3085,
13,
2395,
22,
2910,
18,
2681,
12,
2890,
13,
468,
10939,
2690,
3... |
NUMBER '0200000000' (1, 14) (1, 24) | NUMBER '200000000' (1, 14) (1, 23) | ... " def y(m): # A whitespace after the colon\\n" | 0c1f7c0c8adacac5f2960357b1c35f23352b4e70 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8546/0c1f7c0c8adacac5f2960357b1c35f23352b4e70/test_tokenize.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1372,
6647,
315,
225,
1652,
677,
12,
81,
4672,
468,
432,
7983,
1839,
326,
13336,
1695,
82,
6,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1372,
6647,
315,
225,
1652,
677,
12,
81,
4672,
468,
432,
7983,
1839,
326,
13336,
1695,
82,
6,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
Makes the frame a CONN_TERM frame | Makes the frame a MULTIPLEXER_CONN_TERM frame | def initConnTermFrame(self,referenceID): """ <Purpose> Makes the frame a CONN_TERM frame | c0a4cb118694010ecadc522882f9681e1ce41a0b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7263/c0a4cb118694010ecadc522882f9681e1ce41a0b/Multiplexer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1208,
3543,
4065,
3219,
12,
2890,
16,
6180,
734,
4672,
3536,
411,
10262,
4150,
34,
490,
3223,
326,
2623,
279,
3492,
50,
67,
15176,
2623,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1208,
3543,
4065,
3219,
12,
2890,
16,
6180,
734,
4672,
3536,
411,
10262,
4150,
34,
490,
3223,
326,
2623,
279,
3492,
50,
67,
15176,
2623,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
call([sys.executable, path], cwd=dirname(path)) | call([sys.executable, path, '--prefix', self.prefix], cwd=dirname(path)) | def run(self, fn): path = join(self.meta_dir, fn) if not isfile(path): return from subprocess import call call([sys.executable, path], cwd=dirname(path)) | d175bd41ca1a79c29a85239e2e93877450651493 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7707/d175bd41ca1a79c29a85239e2e93877450651493/main.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
16,
2295,
4672,
589,
273,
1233,
12,
2890,
18,
3901,
67,
1214,
16,
2295,
13,
309,
486,
8814,
12,
803,
4672,
327,
628,
6652,
1930,
745,
745,
3816,
9499,
18,
17751,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
16,
2295,
4672,
589,
273,
1233,
12,
2890,
18,
3901,
67,
1214,
16,
2295,
13,
309,
486,
8814,
12,
803,
4672,
327,
628,
6652,
1930,
745,
745,
3816,
9499,
18,
17751,
16,
... |
layout1 = self.create_pango_layout("%s: %s" % (time1)) | layout1 = self.create_pango_layout(" %s: %s " % (time1)) | def draw(self, context): self.dark = self.get_style().dark[gtk.STATE_NORMAL] self.light = self.get_style().light[gtk.STATE_NORMAL] hpad, vpad = 7, 1 #FIXME: The clock will always redraw everything, #while the next line is outcommented #Problem is that saving the redrawingAll in self, #will make natural redraws, as when hidden under another window, #only draw the one side, and leave the other empty. #ra = self.redrawingAll ra = True if ra or self.player == 0: time0 = self.names[0],self.formatTime(self.getDeciSeconds(0)) layout0 = self.create_pango_layout("%s: %s" % (time0)) layout0.set_font_description(pango.FontDescription("Sans Serif 17")) if ra or self.player == 1: time1 = self.names[1],self.formatTime(self.getDeciSeconds(1)) layout1 = self.create_pango_layout("%s: %s" % (time1)) layout1.set_font_description(pango.FontDescription("Sans Serif 17")) if ra: w = layout1.get_pixel_size()[0] + layout0.get_pixel_size()[0] self.set_size_request(w+hpad*4, self.get_size_request()[1]) rect = self.get_allocation() context.rectangle( float(rect.width)/2 * self.player, 0, float(rect.width)/2, rect.height) context.set_source_color(self.dark) context.fill_preserve() context.new_path() if ra or self.player == 0: if self.player == 0: context.set_source_color(self.light) context.move_to(hpad,vpad) context.show_layout(layout0) if ra or self.player == 1: if self.player == 1: context.set_source_color(self.light) else: context.set_source_color(self.dark) context.move_to(float(rect.width)/2+hpad,vpad) context.show_layout(layout1) | 2acae0ecedcbae24d727a712ea9375809b03ce61 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5339/2acae0ecedcbae24d727a712ea9375809b03ce61/ChessClock.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3724,
12,
2890,
16,
819,
4672,
365,
18,
25045,
273,
365,
18,
588,
67,
4060,
7675,
25045,
63,
4521,
79,
18,
7998,
67,
15480,
65,
365,
18,
5099,
273,
365,
18,
588,
67,
4060,
7675,
5099... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3724,
12,
2890,
16,
819,
4672,
365,
18,
25045,
273,
365,
18,
588,
67,
4060,
7675,
25045,
63,
4521,
79,
18,
7998,
67,
15480,
65,
365,
18,
5099,
273,
365,
18,
588,
67,
4060,
7675,
5099... |
self.output_write(" * %s: %s\n" % (arch, ", ".join(sorted([b for b in broken if b not in nuninst_comp[arch]])))) | self.output_write(" * %s: %s\n" % (arch, ", ".join(sorted([b for b in nuninst[arch] if b not in nuninst_comp[arch]])))) | def iter_packages(self, packages, selected, hint=False, nuninst=None): """Iter on the list of actions and apply them one-by-one | 333b76f04c83e940fd1667457dc7d9c0a58c6aa6 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2784/333b76f04c83e940fd1667457dc7d9c0a58c6aa6/britney.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1400,
67,
10308,
12,
2890,
16,
5907,
16,
3170,
16,
7380,
33,
8381,
16,
290,
318,
8591,
33,
7036,
4672,
3536,
2360,
603,
326,
666,
434,
4209,
471,
2230,
2182,
1245,
17,
1637,
17,
476,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1400,
67,
10308,
12,
2890,
16,
5907,
16,
3170,
16,
7380,
33,
8381,
16,
290,
318,
8591,
33,
7036,
4672,
3536,
2360,
603,
326,
666,
434,
4209,
471,
2230,
2182,
1245,
17,
1637,
17,
476,
... |
try: engine = page[2].engine except Exception as e: wx.MessageBox("There was an error with your database connection. \n\n" + e.__str__()) raise options = page[2].option engine.keep_raw_data = page[1].keepdata.Value engine.use_local = page[1].uselocal opts = dict() for key in options.keys(): opts[key] = options[key].GetValue() engine.opts = opts engine.get_cursor() | def CheckValues(self, evt): """Users can't continue from this page without checking at least one dataset.""" if len(self.scriptlist.GetCheckedStrings()) == 0 and evt.Direction: evt.Veto() elif evt.Direction: checked = self.scriptlist.GetCheckedStrings() warn = [script.name for script in dbtk_list if not script.public and script.name in checked] if warn: warning = "Warning: the following datasets are not " warning += "publicly available. You must have the raw " warning += "data files in the correct location first." warning += "\n\n" + ','.join(warn) wx.MessageBox(warning) | 2882624101c2c22357d412dc6fe912c5cddd0a0b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12606/2882624101c2c22357d412dc6fe912c5cddd0a0b/dbtk_ui.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2073,
1972,
12,
2890,
16,
6324,
4672,
3536,
6588,
848,
1404,
1324,
628,
333,
1363,
2887,
6728,
622,
4520,
1245,
3709,
12123,
309,
562,
12,
2890,
18,
4263,
1098,
18,
967,
11454,
7957,
107... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2073,
1972,
12,
2890,
16,
6324,
4672,
3536,
6588,
848,
1404,
1324,
628,
333,
1363,
2887,
6728,
622,
4520,
1245,
3709,
12123,
309,
562,
12,
2890,
18,
4263,
1098,
18,
967,
11454,
7957,
107... | |
cmd = base+"sites.rotate='1 2 3' sites.translate='4 5 6' selection='%s'"%( | cmd = base+'sites.rotate="1 2 3" sites.translate="4 5 6" selection="%s"'%( | def run(file_name = "phe_e.pdb"): file_name = libtbx.env.find_in_repositories( relative_path="phenix_regression/pdb/%s"%file_name, test=os.path.isfile) output = "modified.pdb" xrsp_init = xray_structure_plus(file_name = file_name) base = \ "mmtbx.pdbtools %s output.pdb.file_name=%s --quiet "%(file_name, output) for selection_str in [None, "chain A or chain C"]: selection = xrsp_init.selection(selection_strings = selection_str) if(selection_str is None): assert selection.size() == selection.count(True) else: assert selection.size() == 36 and selection.count(True) == 24 # cmd = base + "adp.randomize=true selection='%s'"%str(selection_str) check_adp_rand(cmd, xrsp_init, output, selection, selection_str) # cmd = base + "adp.set_b_iso=10.0 selection='%s'"%str(selection_str) check_adp_set_b_iso(cmd, xrsp_init, output, selection, selection_str) # cmd = base + "adp.shift_b_iso=20.0 selection='%s'"%str(selection_str) check_adp_rand(cmd, xrsp_init, output, selection, selection_str) # cmd = base + "adp.scale_adp=2.0 selection='%s'"%str(selection_str) check_adp_rand(cmd, xrsp_init, output, selection, selection_str) # cmd = base + "adp.convert_to_iso=true selection='%s'"%str(selection_str) check_adp_to_iso(cmd, xrsp_init, output, selection, selection_str) # cmd = base + "adp.convert_to_aniso=true selection='%s'"%str(selection_str) check_adp_to_aniso(cmd, xrsp_init, output, selection, selection_str) # shake = 1.5 cmd = base+"sites.shake=%s selection='%s'"%(str(shake), str(selection_str)) check_sites_shake(cmd, xrsp_init, output, selection, selection_str, shake) # cmd = base+"sites.rotate='1 2 3' sites.translate='4 5 6' selection='%s'"%( str(selection_str)) check_sites_rt(cmd, xrsp_init, output, selection, selection_str) # cmd = base+"occupancies.randomize=true selection='%s'"%(str(selection_str)) check_occ_randomize(cmd, xrsp_init, output, selection, selection_str) # remove_selection_str = "element C" cmd = base+"remove.selection='%s' selection='%s'"%( str(remove_selection_str),str(selection_str)) check_remove_selection(cmd, xrsp_init, output, selection, selection_str, remove_selection_str) # test_quiet(file_name) # cmd = base check_all_none(cmd, xrsp_init, output) | d3b962327f9bfcab4633f0d80f058ec29ecde30b /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/696/d3b962327f9bfcab4633f0d80f058ec29ecde30b/tst_pdbtools.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
768,
67,
529,
273,
315,
84,
580,
67,
73,
18,
17414,
6,
4672,
585,
67,
529,
273,
2561,
18587,
92,
18,
3074,
18,
4720,
67,
267,
67,
26626,
12,
3632,
67,
803,
1546,
13819,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
768,
67,
529,
273,
315,
84,
580,
67,
73,
18,
17414,
6,
4672,
585,
67,
529,
273,
2561,
18587,
92,
18,
3074,
18,
4720,
67,
267,
67,
26626,
12,
3632,
67,
803,
1546,
13819,
6... |
mask = N.ones(self._jmim._n_z.value, dtype=int) n_cols, n_nz = self._jmim.opt_dJ_dim(JMI_DER_CPPAD, JMI_DER_DENSE_ROW_MAJOR, independent_variables, mask) | mask = N.ones(self._m._n_z.value, dtype=int) n_cols, n_nz = self._m.jmimodel.opt_dJ_dim(pyjmi.JMI_DER_CPPAD, pyjmi.JMI_DER_DENSE_ROW_MAJOR, independent_variables, mask) | def getCostJacobian(self, independent_variables, mask=None): """ Returns the jacobian for the cost function with respect to the independent variable independent_variable. @param independent_variable: A mask consisting of the independent variables (JMI_DER_X etc.) requested. @todo: independent_variable should not be a mask. It is not pythonesque. @return: The cost gradient/jacobian. """ assert self._jmim._n_z.value != 0 if mask is None: mask = N.ones(self._jmim._n_z.value, dtype=int) n_cols, n_nz = self._jmim.opt_dJ_dim(JMI_DER_CPPAD, JMI_DER_DENSE_ROW_MAJOR, independent_variables, mask) jac = N.zeros(n_nz, dtype=c_jmi_real_t) self._jmim.opt_dJ(JMI_DER_CPPAD, JMI_DER_DENSE_ROW_MAJOR, independent_variables, jac) return jac.reshape( (1, len(jac)) ) | 733964346bd705d3e3fa78795693f8a99289c936 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7711/733964346bd705d3e3fa78795693f8a99289c936/shooting.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1927,
669,
46,
1077,
26968,
12,
2890,
16,
14807,
67,
7528,
16,
3066,
33,
7036,
4672,
3536,
2860,
326,
20138,
26968,
364,
326,
6991,
445,
598,
8762,
358,
326,
14807,
2190,
14807,
67,
6105... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1927,
669,
46,
1077,
26968,
12,
2890,
16,
14807,
67,
7528,
16,
3066,
33,
7036,
4672,
3536,
2860,
326,
20138,
26968,
364,
326,
6991,
445,
598,
8762,
358,
326,
14807,
2190,
14807,
67,
6105... |
import traceback | def handleError(self, record): """ Handle errors which occur during an emit() call. | b89e7c9bc97d3f9332ecc94c88a055c0fe947625 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/b89e7c9bc97d3f9332ecc94c88a055c0fe947625/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
15676,
12,
2890,
16,
1409,
4672,
3536,
5004,
1334,
1492,
3334,
4982,
392,
3626,
1435,
745,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
15676,
12,
2890,
16,
1409,
4672,
3536,
5004,
1334,
1492,
3334,
4982,
392,
3626,
1435,
745,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... | |
def __init__(data = None) | def __init__(data = None): | def __init__(data = None) if data == None: quickfix.IntField.__init__(self, 814) else quickfix.IntField.__init__(self, 814, data) | 484890147d4b23aac4b9d0e85e84fceab7e137c3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8819/484890147d4b23aac4b9d0e85e84fceab7e137c3/quickfix_fields.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
892,
273,
599,
4672,
309,
501,
422,
599,
30,
9549,
904,
18,
1702,
974,
16186,
2738,
972,
12,
2890,
16,
1725,
3461,
13,
469,
9549,
904,
18,
1702,
974,
16186,
2738,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
892,
273,
599,
4672,
309,
501,
422,
599,
30,
9549,
904,
18,
1702,
974,
16186,
2738,
972,
12,
2890,
16,
1725,
3461,
13,
469,
9549,
904,
18,
1702,
974,
16186,
2738,
... |
node.setAttribute("index", "index") | node.setAttribute("index", "yes") | def fixup_refmodindexes_chunk(container): # node is probably a <para>; let's see how often it isn't: if container.tagName != PARA_ELEMENT: sys.stderr.write("--- fixup_refmodindexes_chunk(%s)\n" % container) module_entries = find_all_elements(container, "module") if not module_entries: return index_entries = find_all_elements_from_set(container, REFMODINDEX_ELEMENTS) removes = [] for entry in index_entries: children = entry.childNodes if len(children) != 0: sys.stderr.write( "--- unexpected number of children for %s node:\n" % entry.tagName) sys.stderr.write(entry.toxml() + "\n") continue found = 0 module_name = entry.getAttribute("name") for node in module_entries: if len(node.childNodes) != 1: continue this_name = node.childNodes[0].data if this_name == module_name: found = 1 node.setAttribute("index", "index") if found: removes.append(entry) for node in removes: container.removeChild(node) | 080c1b5af693afd471728e7fa66f744a61e75ed4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/080c1b5af693afd471728e7fa66f744a61e75ed4/docfixer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2917,
416,
67,
1734,
1711,
11265,
67,
6551,
12,
3782,
4672,
468,
756,
353,
8656,
279,
411,
25072,
34,
31,
2231,
1807,
2621,
3661,
16337,
518,
5177,
1404,
30,
309,
1478,
18,
2692,
461,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2917,
416,
67,
1734,
1711,
11265,
67,
6551,
12,
3782,
4672,
468,
756,
353,
8656,
279,
411,
25072,
34,
31,
2231,
1807,
2621,
3661,
16337,
518,
5177,
1404,
30,
309,
1478,
18,
2692,
461,
... |
l.append(r) | if r not in seen: l.append(r) | def revfix(repo, val, defval): if not val and val != 0 and defval is not None: return defval return repo.changelog.rev(repo.lookup(val)) | cd2486859449c3d576b952797b611c767a5b266d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11312/cd2486859449c3d576b952797b611c767a5b266d/cmdutil.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5588,
904,
12,
7422,
16,
1244,
16,
1652,
1125,
4672,
309,
486,
1244,
471,
1244,
480,
374,
471,
1652,
1125,
353,
486,
599,
30,
327,
1652,
1125,
327,
3538,
18,
24083,
12970,
18,
9083,
12... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5588,
904,
12,
7422,
16,
1244,
16,
1652,
1125,
4672,
309,
486,
1244,
471,
1244,
480,
374,
471,
1652,
1125,
353,
486,
599,
30,
327,
1652,
1125,
327,
3538,
18,
24083,
12970,
18,
9083,
12... |
wikipedia.output(u'Graph could not be saved as %s' % filename) | pywikibot.output(u'Graph could not be saved as %s' % filename) | def run(self): for format in config.interwiki_graph_formats: filename = 'interwiki-graphs/' + getFilename(self.originPage, format) if self.graph.write(filename, prog = 'dot', format = format): wikipedia.output(u'Graph saved as %s' % filename) else: wikipedia.output(u'Graph could not be saved as %s' % filename) | 9fb477f51590bae04c7375cb3b5f4e60a3cffe02 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9005/9fb477f51590bae04c7375cb3b5f4e60a3cffe02/interwiki_graph.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
364,
740,
316,
642,
18,
2761,
13044,
67,
4660,
67,
11962,
30,
1544,
273,
296,
2761,
13044,
17,
30016,
2473,
397,
14632,
12,
2890,
18,
10012,
1964,
16,
740,
13,
30... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
364,
740,
316,
642,
18,
2761,
13044,
67,
4660,
67,
11962,
30,
1544,
273,
296,
2761,
13044,
17,
30016,
2473,
397,
14632,
12,
2890,
18,
10012,
1964,
16,
740,
13,
30... |
>>> seguid_dict = SeqIO.to_dict(SeqIO.parse(handle, format), \ key_function = lambda rec : seguid(rec.seq)) | >>> seguid_dict = SeqIO.to_dict(SeqIO.parse(handle, format), ... key_function = lambda rec : seguid(rec.seq)) | def to_dict(sequences, key_function=None) : """Turns a sequence iterator or list into a dictionary. - sequences - An iterator that returns SeqRecord objects, or simply a list of SeqRecord objects. - key_function - Optional function which when given a SeqRecord returns a unique string for the dictionary key. e.g. key_function = lambda rec : rec.name or, key_function = lambda rec : rec.description.split()[0] If key_function is ommitted then record.id is used, on the assumption that the records objects returned are SeqRecords with a unique id field. If there are duplicate keys, an error is raised. Example usage, defaulting to using the record.id as key: >>> from Bio import SeqIO >>> handle = open("GenBank/cor6_6.gb", "rU") >>> format = "genbank" >>> id_dict = SeqIO.to_dict(SeqIO.parse(handle, format)) >>> print id_dict.keys() ['L31939.1', 'AJ237582.1', 'X62281.1', 'AF297471.1', 'X55053.1', 'M81224.1'] >>> print id_dict["L31939.1"].description Brassica rapa (clone bif72) kin mRNA, complete cds. A more complex example, using the key_function argument in order to use a sequence checksum as the dictionary key: >>> from Bio import SeqIO >>> from Bio.SeqUtils.CheckSum import seguid >>> handle = open("GenBank/cor6_6.gb", "rU") >>> format = "genbank" >>> seguid_dict = SeqIO.to_dict(SeqIO.parse(handle, format), \ key_function = lambda rec : seguid(rec.seq)) >>> for key, record in seguid_dict.iteritems() : ... print key, record.id SabZaA4V2eLE9/2Fm5FnyYy07J4 X55053.1 l7gjJFE6W/S1jJn5+1ASrUKW/FA X62281.1 /wQvmrl87QWcm9llO4/efg23Vgg AJ237582.1 TtWsXo45S3ZclIBy4X/WJc39+CY M81224.1 uVEYeAQSV5EDQOnFoeMmVea+Oow AF297471.1 BUg6YxXSKWEcFFH0L08JzaLGhQs L31939.1 """ if key_function is None : key_function = lambda rec : rec.id d = dict() for record in sequences : key = key_function(record) if key in d : raise ValueError("Duplicate key '%s'" % key) d[key] = record return d | 1402c9724064176b78d630b217fa4a42a1bfa012 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7167/1402c9724064176b78d630b217fa4a42a1bfa012/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
358,
67,
1576,
12,
17190,
16,
498,
67,
915,
33,
7036,
13,
294,
3536,
31873,
279,
3102,
2775,
578,
666,
1368,
279,
3880,
18,
225,
300,
8463,
225,
300,
1922,
2775,
716,
1135,
14367,
2115... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
358,
67,
1576,
12,
17190,
16,
498,
67,
915,
33,
7036,
13,
294,
3536,
31873,
279,
3102,
2775,
578,
666,
1368,
279,
3880,
18,
225,
300,
8463,
225,
300,
1922,
2775,
716,
1135,
14367,
2115... |
""" {% get_posts_by_tag <tag> [<limit>] as <varname> %} """ tag, limit, varname = parse_token(token) return NewsItemNode(varname, limit, filters={'tags__contains':tag}) | """ {% get_posts_by_tag <tag> [<limit>] as <varname> %} """ tag, limit, varname = parse_token(token) return NewsItemNode(varname, limit, filters={'tags__contains':tag}) | def get_posts_by_tag(parser,token): """ {% get_posts_by_tag <tag> [<limit>] as <varname> %} """ tag, limit, varname = parse_token(token) return NewsItemNode(varname, limit, filters={'tags__contains':tag}) | 2d3c8ed0ed7c2f6a9f8202b37fc595f1f8bfb7d6 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/2484/2d3c8ed0ed7c2f6a9f8202b37fc595f1f8bfb7d6/news_tags.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
11660,
67,
1637,
67,
2692,
12,
4288,
16,
2316,
4672,
3536,
288,
9,
336,
67,
11660,
67,
1637,
67,
2692,
411,
2692,
34,
306,
32,
3595,
34,
65,
487,
411,
1401,
529,
34,
738,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
11660,
67,
1637,
67,
2692,
12,
4288,
16,
2316,
4672,
3536,
288,
9,
336,
67,
11660,
67,
1637,
67,
2692,
411,
2692,
34,
306,
32,
3595,
34,
65,
487,
411,
1401,
529,
34,
738,
... |
command = config["transfers"]["afterfinish"].replace("$", utils.escapeCommand(newname)) os.system("%s &" % command) | command = [x.replace("$", newname) for x in config["transfers"]["afterfinish"].split(' ')] proc = Popen(command) | def FileDownload(self, msg): """ A file download is in progress""" needupdate = 1 config = self.eventprocessor.config.sections for i in self.downloads: if i.conn != msg.conn: continue try: if i.transfertimer is not None: i.transfertimer.cancel() curtime = time.time() i.currentbytes = msg.file.tell() if i.lastbytes is None: i.lastbytes = i.currentbytes if i.starttime is None: i.starttime = curtime if i.lasttime is None: i.lasttime = curtime - 1 #i.status = "%s" %(str(i.currentbytes)) i.status = "Transferring" oldelapsed = i.timeelapsed i.timeelapsed = curtime - i.starttime if curtime > i.starttime and i.currentbytes > i.offset: i.speed = (i.currentbytes - i.lastbytes)/(curtime - i.lasttime)/1024 if i.speed <= 0.0: i.timeleft = "∞" else: i.timeleft = self.getTime((i.size - i.currentbytes)/i.speed/1024) i.lastbytes = i.currentbytes i.lasttime = curtime if i.size > i.currentbytes: if oldelapsed == i.timeelapsed: needupdate = 0 #i.status = str(i.currentbytes) i.status = "Transferring" else: msg.file.close() basename = self.CleanPath(self.encode(string.split(i.filename,'\\')[-1], i.user)) downloaddir = config["transfers"]["downloaddir"] if i.path and i.path[0] == '/': folder = self.CleanPath(i.path) else: folder = os.path.join(downloaddir, self.encode(i.path)) if not os.access(folder, os.F_OK): os.makedirs(folder) (newname, identicalfile) = self.getRenamedEnhanced(os.path.join(folder, basename), msg.file.name) if newname: try: if win32: os.rename(msg.file.name, u"%s" % newname) else: os.rename(msg.file.name, newname) except OSError: try: f1 = open(msg.file.name, "rb") d = f1.read() if win32: f1 = open(u"%s" % newname, "wb") else: f1 = open(newname, "wb") f1.write(d) f1.close() os.remove(msg.file.name) except OSError: self.eventprocessor.logMessage(_("Couldn't move '%(tempfile)s' to '%(file)s'") % {'tempfile':self.decode(msg.file.name), 'file':self.decode(newname)}) i.status = "Finished" if newname: self.eventprocessor.logMessage(_("Download finished: %(file)s") % {'file':self.decode(newname)}) self.eventprocessor.logTransfer(_("Download finished: user %(user)s, file %(file)s") % {'user':i.user, 'file':self.decode(i.filename)}) else: self.eventprocessor.logMessage(_("File %(file)s is identical to %(identical)s, not saving.") % {'file':self.decode(msg.file.name), 'identical':identicalfile}) self.eventprocessor.logTransfer(_("Download finished but not saved since it's a duplicate: user %(user)s, file %(file)s") % {'user':i.user, 'file':self.decode(i.filename)}) self.queue.put(slskmessages.ConnClose(msg.conn)) #if i.speed is not None: #self.queue.put(slskmessages.SendSpeed(i.user, int(i.speed*1024))) #Removed due to misuse. Replaced by SendUploadSpeed i.conn = None if newname: if win32: self.addToShared(u"%s" % newname) else: self.addToShared(newname) self.eventprocessor.sendNumSharedFoldersFiles() self.SaveDownloads() self.downloadspanel.update(i) if config["transfers"]["shownotification"]: self.eventprocessor.frame.NewNotification(_("%(file)s downloaded from %(user)s") % {'user':i.user, "file":newname.rsplit(os.sep, 1)[1]}, title=_("Nicotine+ :: file downloaded")) | 0d0924480854d90f7d8e3057496ba8dab38cb4a7 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8738/0d0924480854d90f7d8e3057496ba8dab38cb4a7/transfers.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1387,
7109,
12,
2890,
16,
1234,
4672,
3536,
432,
585,
4224,
353,
316,
4007,
8395,
1608,
2725,
273,
404,
642,
273,
365,
18,
2575,
8700,
18,
1425,
18,
11657,
364,
277,
316,
365,
18,
7813... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1387,
7109,
12,
2890,
16,
1234,
4672,
3536,
432,
585,
4224,
353,
316,
4007,
8395,
1608,
2725,
273,
404,
642,
273,
365,
18,
2575,
8700,
18,
1425,
18,
11657,
364,
277,
316,
365,
18,
7813... |
res = RecapDecompose(m) | res = RecapDecompose(m,onlyUseReactions=[1]) | def testAmideRxn(self): m = Chem.MolFromSmiles('C1CC1C(=O)NC1OC1') res = RecapDecompose(m) self.failUnless(res) self.failUnless(len(res.GetLeaves())==2) ks = res.GetLeaves().keys() self.failUnless('[*]C(=O)C1CC1' in ks) self.failUnless('[*]NC1CO1' in ks) m = Chem.MolFromSmiles('C1CC1C(=O)N(C)C1OC1') res = RecapDecompose(m) self.failUnless(res) self.failUnless(len(res.GetLeaves())==2) ks = res.GetLeaves().keys() self.failUnless('[*]C(=O)C1CC1' in ks) self.failUnless('[*]N(C)C1CO1' in ks) | e793cf1d5a1abc952de6d7bd7feb70328e8a57c4 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9195/e793cf1d5a1abc952de6d7bd7feb70328e8a57c4/Recap.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
9864,
831,
54,
22695,
12,
2890,
4672,
312,
273,
26542,
18,
49,
355,
1265,
9552,
1449,
2668,
39,
21,
6743,
21,
39,
12,
33,
51,
13,
10346,
21,
10809,
21,
6134,
400,
273,
868,
590... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
9864,
831,
54,
22695,
12,
2890,
4672,
312,
273,
26542,
18,
49,
355,
1265,
9552,
1449,
2668,
39,
21,
6743,
21,
39,
12,
33,
51,
13,
10346,
21,
10809,
21,
6134,
400,
273,
868,
590... |
for x in testmap: | for x in self.fspath.listdir(): | def collect(self): l = [] for x in testmap: name = x.basename regrtest = self.get(name) if regrtest is not None: #if option.extracttests: # l.append(InterceptedRunModule(name, self, regrtest)) #else: l.append(RunFileExternal(name, parent=self, regrtest=regrtest)) return l | 6b808a12b6d15071694d05215329617d4736baf9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6934/6b808a12b6d15071694d05215329617d4736baf9/conftest.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3274,
12,
2890,
4672,
328,
273,
5378,
364,
619,
316,
365,
18,
2556,
803,
18,
1098,
1214,
13332,
508,
273,
619,
18,
13909,
960,
3797,
395,
273,
365,
18,
588,
12,
529,
13,
309,
960,
37... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3274,
12,
2890,
4672,
328,
273,
5378,
364,
619,
316,
365,
18,
2556,
803,
18,
1098,
1214,
13332,
508,
273,
619,
18,
13909,
960,
3797,
395,
273,
365,
18,
588,
12,
529,
13,
309,
960,
37... |
result = rdflib.sparql.Query.query(graph, select, where) | result = Query.query(sparqlGr, select, where) | def getGeoPosition(self, foaf, sha1mail): """ Obtain geography information from foaf @param foaf: a foaf uri @param sha1mail: mail addess enconded @return: coordinates """ graph = self.__getGraph(foaf) if (graph != None): select = ('?lat', '?long') where = GraphPattern([ ('?x', RDF['type'], FOAF['Person']), ('?x', FOAF['mbox_sha1sum'], sha1mail), ('?x', FOAF['based_near'], '?y'), ('?y', GEO['lat'], '?lat'), ('?y', GEO['long'], '?long') ]) result = rdflib.sparql.Query.query(graph, select, where) for one in result: return [one[0], one[1]] return [None, None] | 8ef5a273f0122121e99747775ac7d5c8b3c6b2ca /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2232/8ef5a273f0122121e99747775ac7d5c8b3c6b2ca/foaf.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7162,
4361,
2555,
12,
2890,
16,
18261,
1727,
16,
6056,
21,
4408,
4672,
3536,
24850,
8244,
15669,
1779,
628,
18261,
1727,
225,
632,
891,
18261,
1727,
30,
279,
18261,
1727,
2003,
632,
891,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7162,
4361,
2555,
12,
2890,
16,
18261,
1727,
16,
6056,
21,
4408,
4672,
3536,
24850,
8244,
15669,
1779,
628,
18261,
1727,
225,
632,
891,
18261,
1727,
30,
279,
18261,
1727,
2003,
632,
891,
... |
parentId = simbase.air.districtId self.parentId = parentId | if not hasattr(self, 'parentId'): parentId = simbase.air.districtId self.parentId = parentId | def generateWithRequiredAndId(self, doId, zoneId, optionalFields=[]): assert self.notify.debugStateCall(self) # have we already allocated a doId? if self.__preallocDoId: assert doId == self.__preallocDoId self.__preallocDoId = 0 | f1d59344483f34f952f612134837e6f2ce54900a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7242/f1d59344483f34f952f612134837e6f2ce54900a/DistributedObjectAI.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2103,
1190,
3705,
1876,
548,
12,
2890,
16,
741,
548,
16,
29939,
16,
3129,
2314,
33,
8526,
4672,
1815,
365,
18,
12336,
18,
4148,
1119,
1477,
12,
2890,
13,
468,
1240,
732,
1818,
11977,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2103,
1190,
3705,
1876,
548,
12,
2890,
16,
741,
548,
16,
29939,
16,
3129,
2314,
33,
8526,
4672,
1815,
365,
18,
12336,
18,
4148,
1119,
1477,
12,
2890,
13,
468,
1240,
732,
1818,
11977,
2... |
raise Unauthorised, self._( | raise exceptions.Unauthorised, self._( | def _changenode(self, cn, nodeid, props): """Change the node based on the contents of the form.""" # check for permission if not self.editItemPermission(props): raise Unauthorised, self._( 'You do not have permission to edit %(class)s' ) % {'class': cn} | 7b35d0c71d26a7290bec5255325716330d7a66bf /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1906/7b35d0c71d26a7290bec5255325716330d7a66bf/actions.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
24083,
20680,
12,
2890,
16,
6227,
16,
756,
350,
16,
3458,
4672,
3536,
3043,
326,
756,
2511,
603,
326,
2939,
434,
326,
646,
12123,
468,
866,
364,
4132,
309,
486,
365,
18,
4619,
118... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
24083,
20680,
12,
2890,
16,
6227,
16,
756,
350,
16,
3458,
4672,
3536,
3043,
326,
756,
2511,
603,
326,
2939,
434,
326,
646,
12123,
468,
866,
364,
4132,
309,
486,
365,
18,
4619,
118... |
if enter_chroot: cmd = ['./enter_chroot.sh', '--'] + cmd | if isinstance(cmd, basestring): if enter_chroot: cmd = './enter_chroot.sh -- ' + cmd cmd_str = cmd else: if enter_chroot: cmd = ['./enter_chroot.sh', '--'] + cmd cmd_str = ' '.join(cmd) | def RunCommand(cmd, print_cmd=True, error_ok=False, error_message=None, exit_code=False, redirect_stdout=False, redirect_stderr=False, cwd=None, input=None, enter_chroot=False): """Runs a shell command. Keyword arguments: cmd - cmd to run. Should be input to subprocess.POpen. If a string, converted to an array using split(). print_cmd -- prints the command before running it. error_ok -- does not raise an exception on error. error_message -- prints out this message when an error occurrs. exit_code -- returns the return code of the shell command. redirect_stdout -- returns the stdout. redirect_stderr -- holds stderr output until input is communicated. cwd -- the working directory to run this cmd. input -- input to pipe into this command through stdin. enter_chroot -- this command should be run from within the chroot. If set, cwd must point to the scripts directory. Raises: Exception: Raises generic exception on error with optional error_message. """ # Set default for variables. stdout = None stderr = None stdin = None output = '' # Modify defaults based on parameters. if redirect_stdout: stdout = subprocess.PIPE if redirect_stderr: stderr = subprocess.PIPE if input: stdin = subprocess.PIPE if enter_chroot: cmd = ['./enter_chroot.sh', '--'] + cmd # Print out the command before running. if print_cmd: Info('RunCommand: %s' % ' '.join(cmd)) try: proc = subprocess.Popen(cmd, cwd=cwd, stdin=stdin, stdout=stdout, stderr=stderr) (output, error) = proc.communicate(input) if exit_code: return proc.returncode if not error_ok and proc.returncode: raise Exception('Command "%s" failed.\n' % (' '.join(cmd)) + (error_message or error or output or '')) except Exception,e: if not error_ok: raise else: Warning(str(e)) return output | 6db8f522a1af07dbf7c6aafb310eb4925fadd71b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3785/6db8f522a1af07dbf7c6aafb310eb4925fadd71b/cros_build_lib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1939,
2189,
12,
4172,
16,
1172,
67,
4172,
33,
5510,
16,
555,
67,
601,
33,
8381,
16,
555,
67,
2150,
33,
7036,
16,
2427,
67,
710,
33,
8381,
16,
3136,
67,
10283,
33,
8381,
16,
3136,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1939,
2189,
12,
4172,
16,
1172,
67,
4172,
33,
5510,
16,
555,
67,
601,
33,
8381,
16,
555,
67,
2150,
33,
7036,
16,
2427,
67,
710,
33,
8381,
16,
3136,
67,
10283,
33,
8381,
16,
3136,
6... |
RouterStats.table.c.circ_to_rate : RouterStats.table.c.circ_fail_to/RouterStats.table.c.circ_try_to, RouterStats.table.c.circ_bi_rate : | RouterStats.table.c.circ_to_rate: RouterStats.table.c.circ_fail_to/RouterStats.table.c.circ_try_to, RouterStats.table.c.circ_bi_rate: | def _compute_stats_query(stats_clause): # http://www.sqlalchemy.org/docs/04/sqlexpression.html#sql_update to_s = select([func.count(Extension.id)], and_(stats_clause, Extension.table.c.to_node_idhex == RouterStats.table.c.router_idhex)).as_scalar() from_s = select([func.count(Extension.id)], and_(stats_clause, Extension.table.c.from_node_idhex == RouterStats.table.c.router_idhex)).as_scalar() f_to_s = select([func.count(FailedExtension.id)], and_(stats_clause, FailedExtension.table.c.to_node_idhex == RouterStats.table.c.router_idhex)).as_scalar() f_from_s = select([func.count(FailedExtension.id)], and_(stats_clause, FailedExtension.table.c.from_node_idhex == RouterStats.table.c.router_idhex)).as_scalar() avg_ext = select([func.avg(Extension.delta)], and_(stats_clause, Extension.table.c.to_node_idhex==RouterStats.table.c.router_idhex, Extension.table.c.hop==0, Extension.table.c.row_type=='extension')).as_scalar() | c8ad6cda0735bca71a66e2ae8dc82c8034ffc7ce /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3224/c8ad6cda0735bca71a66e2ae8dc82c8034ffc7ce/SQLSupport.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
9200,
67,
5296,
67,
2271,
12,
5296,
67,
18128,
4672,
468,
1062,
2207,
5591,
18,
4669,
24182,
18,
3341,
19,
8532,
19,
3028,
19,
11410,
4149,
1337,
18,
2620,
7,
4669,
67,
2725,
358,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
9200,
67,
5296,
67,
2271,
12,
5296,
67,
18128,
4672,
468,
1062,
2207,
5591,
18,
4669,
24182,
18,
3341,
19,
8532,
19,
3028,
19,
11410,
4149,
1337,
18,
2620,
7,
4669,
67,
2725,
358,... |
cal = LocaleTextCalendar(locale=options.locale) | cal = LocaleTextCalendar(locale=locale) | def main(args): import optparse parser = optparse.OptionParser(usage="usage: %prog [options] [year [month]]") parser.add_option( "-w", "--width", dest="width", type="int", default=2, help="width of date column (default 2, text only)" ) parser.add_option( "-l", "--lines", dest="lines", type="int", default=1, help="number of lines for each week (default 1, text only)" ) parser.add_option( "-s", "--spacing", dest="spacing", type="int", default=6, help="spacing between months (default 6, text only)" ) parser.add_option( "-m", "--months", dest="months", type="int", default=3, help="months per row (default 3, text only)" ) parser.add_option( "-c", "--css", dest="css", default="calendar.css", help="CSS to use for page (html only)" ) parser.add_option( "-L", "--locale", dest="locale", default=None, help="locale to be used from month and weekday names" ) parser.add_option( "-e", "--encoding", dest="encoding", default=None, help="Encoding to use for output" ) parser.add_option( "-t", "--type", dest="type", default="text", choices=("text", "html"), help="output type (text or html)" ) (options, args) = parser.parse_args(args) if options.locale and not options.encoding: parser.error("if --locale is specified --encoding is required") sys.exit(1) if options.type == "html": if options.locale: cal = LocaleHTMLCalendar(locale=options.locale) else: cal = HTMLCalendar() encoding = options.encoding if encoding is None: encoding = sys.getdefaultencoding() optdict = dict(encoding=encoding, css=options.css) if len(args) == 1: print cal.formatyearpage(datetime.date.today().year, **optdict) elif len(args) == 2: print cal.formatyearpage(int(args[1]), **optdict) else: parser.error("incorrect number of arguments") sys.exit(1) else: if options.locale: cal = LocaleTextCalendar(locale=options.locale) else: cal = TextCalendar() optdict = dict(w=options.width, l=options.lines) if len(args) != 3: optdict["c"] = options.spacing optdict["m"] = options.months if len(args) == 1: result = cal.formatyear(datetime.date.today().year, **optdict) elif len(args) == 2: result = cal.formatyear(int(args[1]), **optdict) elif len(args) == 3: result = cal.formatmonth(int(args[1]), int(args[2]), **optdict) else: parser.error("incorrect number of arguments") sys.exit(1) if options.encoding: result = result.encode(options.encoding) print result | cd469a0757f6a55b468e50bf09177c79ce0782a5 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12029/cd469a0757f6a55b468e50bf09177c79ce0782a5/calendar.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
1968,
4672,
1930,
2153,
2670,
2082,
273,
2153,
2670,
18,
1895,
2678,
12,
9167,
1546,
9167,
30,
738,
14654,
306,
2116,
65,
306,
6874,
306,
7496,
65,
4279,
13,
2082,
18,
1289,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
1968,
4672,
1930,
2153,
2670,
2082,
273,
2153,
2670,
18,
1895,
2678,
12,
9167,
1546,
9167,
30,
738,
14654,
306,
2116,
65,
306,
6874,
306,
7496,
65,
4279,
13,
2082,
18,
1289,
... |
def point(self, multiplier=1): point = 0 result = self.reference() if self.homeScore >=0 and self.homeScore == result.homeScore: point = point + 1 elif self.homeScore > 4 and result.homeScore > 4: point = point + 1 if self.awayScore >=0 and self.awayScore == result.awayScore: point = point + 1 elif self.awayScore > 4 and result.awayScore > 4: point = point + 1 if self.home_w() and result.home_w(): point = point + 2 elif self.home_d() and result.home_d(): point = point + 2 elif self.home_l() and result.home_l(): point = point + 2 return point * multiplier | def point(self, multiplier=1): point = 0 result = self.reference() if self.homeScore >=0 and self.homeScore == result.homeScore: point = point + 1 elif self.homeScore > 4 and result.homeScore > 4: point = point + 1 if self.awayScore >=0 and self.awayScore == result.awayScore: point = point + 1 elif self.awayScore > 4 and result.awayScore > 4: point = point + 1 if self.home_w() and result.home_w(): point = point + 2 elif self.home_d() and result.home_d(): point = point + 2 elif self.home_l() and result.home_l(): point = point + 2 return point * multiplier | 1ab0b278e22f8fdada61cf86113af2b2721da06f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11033/1ab0b278e22f8fdada61cf86113af2b2721da06f/model.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1634,
12,
2890,
16,
15027,
33,
21,
4672,
1634,
273,
374,
563,
273,
365,
18,
6180,
1435,
309,
365,
18,
8712,
7295,
1545,
20,
471,
365,
18,
8712,
7295,
422,
563,
18,
8712,
7295,
30,
16... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1634,
12,
2890,
16,
15027,
33,
21,
4672,
1634,
273,
374,
563,
273,
365,
18,
6180,
1435,
309,
365,
18,
8712,
7295,
1545,
20,
471,
365,
18,
8712,
7295,
422,
563,
18,
8712,
7295,
30,
16... | |
from pandac.PandaModules import NotifySeverity | from pandac.PandaModules import NSDebug, NSInfo, NSWarning, NSError | def getSeverity(self): from pandac.PandaModules import NotifySeverity if self.getDebug(): return NotifySeverity.NSDebug elif self.getInfo(): return NotifySeverity.NSInfo elif self.getWarning(): return NotifySeverity.NSWarning else: return NotifySeverity.NSError | 439beadb7b0354026959fb92d947dce0178d4bfc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7242/439beadb7b0354026959fb92d947dce0178d4bfc/Notifier.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1322,
6084,
560,
12,
2890,
4672,
628,
293,
464,
1077,
18,
52,
464,
69,
7782,
1930,
11472,
2829,
16,
11472,
966,
16,
11472,
6210,
16,
11472,
668,
309,
365,
18,
588,
2829,
13332,
327,
10... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1322,
6084,
560,
12,
2890,
4672,
628,
293,
464,
1077,
18,
52,
464,
69,
7782,
1930,
11472,
2829,
16,
11472,
966,
16,
11472,
6210,
16,
11472,
668,
309,
365,
18,
588,
2829,
13332,
327,
10... |
return value raise TextFieldException(('Value is not File, String or ' 'BaseUnit on %s: %r' % (self.getName(), type(value)))) | if encoding is not None and type(value) == type(u''): value = value.encode(encoding) else: value = str(value) if mimetype is None: mimetype, enc = guess_content_type('', value, mimetype) if not value: return default, mimetype return value, mimetype else: if ((isinstance(value, FileUpload) and value.filename != '') or (isinstance(value, FileType) and value.name != '')): f_name = '' if isinstance(value, FileUpload): f_name = value.filename if isinstance(value, FileType): f_name = value.name value = value.read() if mimetype is None: mimetype, enc = guess_content_type(f_name, value, mimetype) size = len(value) if size == 0: return default, mimetype return value, mimetype elif IBaseUnit.isImplementedBy(value): if mimetype is None: mimetype, enc = guess_content_type('', str(value), mimetype) return value, getattr(aq_base(value), 'mimetype', mimetype) raise TextFieldException('Value is not File, String or BaseUnit on %s: %r' % (self.getName(), type(value))) | def _process_input(self, value, default=None, **kwargs): # We also need to handle the case where there is a baseUnit # for this field containing a valid set of data that would # not be reuploaded in a subsequent edit, this is basically # migrated from the old BaseObject.set method if ((isinstance(value, FileUpload) and value.filename != '') or (isinstance(value, FileType) and value.name != '')): #OK, its a file, is it empty? if not value.read(1): # This new file has no length, so we keep # the orig return default value.seek(0) return value | 6c49241026527d58ec33dad69b98fdf197a65ae4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12165/6c49241026527d58ec33dad69b98fdf197a65ae4/Field.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2567,
67,
2630,
12,
2890,
16,
460,
16,
805,
33,
7036,
16,
2826,
4333,
4672,
468,
1660,
2546,
1608,
358,
1640,
326,
648,
1625,
1915,
353,
279,
1026,
2802,
468,
364,
333,
652,
4191,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2567,
67,
2630,
12,
2890,
16,
460,
16,
805,
33,
7036,
16,
2826,
4333,
4672,
468,
1660,
2546,
1608,
358,
1640,
326,
648,
1625,
1915,
353,
279,
1026,
2802,
468,
364,
333,
652,
4191,... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.