rem
stringlengths
1
226k
add
stringlengths
0
227k
context
stringlengths
6
326k
meta
stringlengths
143
403
input_ids
listlengths
256
256
attention_mask
listlengths
256
256
labels
listlengths
128
128
if dir: if not os.path.exists(dir): os.mkdir(dir, 0o700) else: print("Cannot determine location of $HOME. Not storing default user") return False
if not os.path.exists(dir): os.mkdir(dir, 0o700)
def _save_user_info(self, username, passwd, dir, file_name): ''' Save username and password in file "dir + file_name" If it's saved properly, return True, or else return False. ''' try: if dir: if not os.path.exists(dir): os.mkdir(dir, 0o700) else: print("Cannot determine location of $HOME. Not storing default user") return False csvfilepath = dir + file_name csvfile = open(csvfilepath, 'w') os.chmod(csvfilepath, 0o600) writer = csv.writer(csvfile) writer.writerow([username, passwd]) csvfile.close() except Exception as e: # just not store it print(e, "\nCannot write ~/.bind10/default_user.csv; default user is not stored") return False
6000bb171b66735393843be9a3cb2ab34f01c634 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6792/6000bb171b66735393843be9a3cb2ab34f01c634/bindcmd.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 5688, 67, 1355, 67, 1376, 12, 2890, 16, 2718, 16, 19093, 16, 1577, 16, 585, 67, 529, 4672, 9163, 7074, 2718, 471, 2201, 316, 585, 315, 1214, 397, 585, 67, 529, 6, 971, 518, 1807...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 5688, 67, 1355, 67, 1376, 12, 2890, 16, 2718, 16, 19093, 16, 1577, 16, 585, 67, 529, 4672, 9163, 7074, 2718, 471, 2201, 316, 585, 315, 1214, 397, 585, 67, 529, 6, 971, 518, 1807...
if _debug: print "\t%s: Called Plane.__init__()" % rendererName
Item.__init__(self) debugMsg("Called Plane.__init__()")
def __init__(self,scene): """ Initialisation of the Plane object """ if _debug: print "\t%s: Called Plane.__init__()" % rendererName
431189e7e535c0ce614e7ffbef2068439f08912a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8476/431189e7e535c0ce614e7ffbef2068439f08912a/plane.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 23694, 4672, 3536, 10188, 10742, 434, 326, 3008, 8806, 733, 3536, 309, 389, 4148, 30, 1172, 1548, 88, 9, 87, 30, 11782, 3008, 8806, 16186, 2738, 972, 10031...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 23694, 4672, 3536, 10188, 10742, 434, 326, 3008, 8806, 733, 3536, 309, 389, 4148, 30, 1172, 1548, 88, 9, 87, 30, 11782, 3008, 8806, 16186, 2738, 972, 10031...
content = content.encode('iso-8859-15', 'replace')
content = content.encode('Windows-1252', 'replace')
def post(self, content): if isinstance(content, unicode): content = content.encode('iso-8859-15', 'replace')
c4d8f7de2e451f8490c489c44dbd9103cebc38bb /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7471/c4d8f7de2e451f8490c489c44dbd9103cebc38bb/contact_thread.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1603, 12, 2890, 16, 913, 4672, 309, 1549, 12, 1745, 16, 5252, 4672, 913, 273, 913, 18, 3015, 2668, 9699, 17, 17258, 17, 3600, 2187, 296, 2079, 6134, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1603, 12, 2890, 16, 913, 4672, 309, 1549, 12, 1745, 16, 5252, 4672, 913, 273, 913, 18, 3015, 2668, 9699, 17, 17258, 17, 3600, 2187, 296, 2079, 6134, 2, -100, -100, -100, -100, -100, -1...
for group in sorted(available, key=lambda x: x.nameByLang(mylang)):
for group in available:
def returnGroupLists(self, userlist):
6be024406dc4d862c1e43975fa7754f098cc04ff /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5445/6be024406dc4d862c1e43975fa7754f098cc04ff/cli.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 327, 1114, 7432, 12, 2890, 16, 26860, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 327, 1114, 7432, 12, 2890, 16, 26860, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10...
' in ' + input
' in ' + build_file
def ExpandVariables(input, is_late, variables, build_file): # Look for the pattern that gets expanded into variables if not is_late: variable_re = early_variable_re else: variable_re = late_variable_re matches = map(None, variable_re.finditer(input)) output = input if matches != None: # Reverse the list of matches so that replacements are done right-to-left. # That ensures that earlier replacements won't mess up the string in a # way that causes later calls to find the earlier substituted text instead # of what's intended for replacement. matches.reverse() for match_group in matches: match = match_group.groupdict() # match['replace'] is the substring to look for, match['type'] # is the character code for the replacement type (< > <! >! <@ # >@ <!@ >!@), match['is_array'] contains a '[' for command # arrays, and match['content'] is the name of the variable (< >) # or command to run (<! >!). # run_command is true if a ! variant is used. run_command = '!' in match['type'] # expand_to_list is true if an @ variant is used. In that case, # the expansion should result in a list. Note that the caller # is to be expecting a list in return, and not all callers do # because not all are working in list context. Also, for list # expansions, there can be no other text besides the variable # expansion in the input. expand_to_list = '@' in match['type'] and input == match['replace'] if run_command: # Run the command in the build file's directory. build_file_dir = os.path.dirname(build_file) if build_file_dir == '': # If build_file is just a leaf filename indicating a file in the # current directory, build_file_dir might be an empty string. Set # it to None to signal to subprocess.Popen that it should run the # command in the current directory. build_file_dir = None if match['is_array']: command = eval('[%s]' % match['content']) else: command = match['content'] p = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, cwd=build_file_dir) (p_stdout, p_stderr) = p.communicate() if p.wait() != 0 or p_stderr: sys.stderr.write(p_stderr) # Simulate check_call behavior by reusing its exception. raise subprocess.CalledProcessError(p.returncode, command) replacement = p_stdout.rstrip() else: if not match['content'] in variables: raise KeyError, 'Undefined variable ' + match['content'] + \ ' in ' + input replacement = variables[match['content']] if isinstance(replacement, list): for item in replacement: if not isinstance(item, str) and not isinstance(item, int): raise TypeError, 'Variable ' + match['content'] + \ ' must expand to a string or list of strings; ' + \ 'list contains a ' + \ item.__class__.__name__ # Run through the list and handle variable expansions in it. Since # the list is guaranteed not to contain dicts, this won't do anything # with conditions sections. # # TODO(mark): I think this should be made more general: any time an # expansion is done, if there are more expandable tokens left in the # output, additional expansion phases should be done. It should not # be effective solely for lists. ProcessVariablesAndConditionsInList(replacement, is_late, variables, build_file) elif not isinstance(replacement, str) and \ not isinstance(replacement, int): raise TypeError, 'Variable ' + match['content'] + \ ' must expand to a string or list of strings; ' + \ 'found a ' + replacement.__class__.__name__ if expand_to_list: # Expanding in list context. It's guaranteed that there's only one # replacement to do in |input| and that it's this replacement. See # above. if isinstance(replacement, list): # If it's already a list, make a copy. output = replacement[:] else: # Split it the same way sh would split arguments. output = shlex.split(str(replacement)) else: # Expanding in string context. if isinstance(replacement, list): # When expanding a list into string context, turn the list items # into a string in a way that will work with a subprocess call. # # TODO(mark): This isn't completely correct. This should # call a generator-provided function that observes the # proper list-to-argument quoting rules on a specific # platform instead of just calling the POSIX encoding # routine. output = "".join([output[:match_group.start('replace')], gyp.common.EncodePOSIXShellList(replacement), output[match_group.end('replace'):]]) else: # Expanding into string context is easy, just replace it. output = "".join([output[:match_group.start('replace')], str(replacement), output[match_group.end('replace'):]]) return output
ac5e1106245db7117ff502314437cd6ee4ffad65 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6220/ac5e1106245db7117ff502314437cd6ee4ffad65/input.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 16429, 6158, 12, 2630, 16, 353, 67, 20293, 16, 3152, 16, 1361, 67, 768, 4672, 468, 10176, 364, 326, 1936, 716, 5571, 8406, 1368, 3152, 309, 486, 353, 67, 20293, 30, 2190, 67, 266, 273,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 16429, 6158, 12, 2630, 16, 353, 67, 20293, 16, 3152, 16, 1361, 67, 768, 4672, 468, 10176, 364, 326, 1936, 716, 5571, 8406, 1368, 3152, 309, 486, 353, 67, 20293, 30, 2190, 67, 266, 273,...
if self.has_handler('template.xhtml'):
try:
def get_template(self): if self.has_handler('template.xhtml'): return self.get_handler('template.xhtml') # Default, aruni return self.get_handler('/ui/aruni/template.xhtml')
561b04b86500f7f67e69de0f03a500447abb6080 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12681/561b04b86500f7f67e69de0f03a500447abb6080/skins.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 3202, 12, 2890, 4672, 775, 30, 327, 365, 18, 588, 67, 4176, 2668, 3202, 18, 26341, 6134, 468, 2989, 16, 419, 318, 77, 327, 365, 18, 588, 67, 4176, 2668, 19, 4881, 19, 297, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 3202, 12, 2890, 4672, 775, 30, 327, 365, 18, 588, 67, 4176, 2668, 3202, 18, 26341, 6134, 468, 2989, 16, 419, 318, 77, 327, 365, 18, 588, 67, 4176, 2668, 19, 4881, 19, 297, ...
print "%d %d %d" % (len (data), len (cdata), len (bdata)) EXEDATA = """ eNrtfQt4U9eV7pEl28IIJIJNlMRJlCBaN3JcE5uMiYDK2HJMsI1s2TIkDsbYMrLwK9IRj4akENkT DhvlMU06aW/bSZp2pnMnM5OZ5jY0bYONSQzkwcMp0JJpIO1NZQyEJATMU/df+2z5BalnvszMN/f7 fPi29jl7r7X2WmuvvdbaW0em9P6nJK0kSTqUWEyStknq5ZDGv46iTL31tanSK5PeuW2bpuSd2yp9 TUFLe6BtVaCuxVJf19raJltWei2BUKulqdVSuMRtaWlr8GZNmZJiFTT66x9r/ceEE/Pj5eAtR+e/ hDrrlg/n/zPqitujvN57c2z+TzjMyfmfCdgXeP9HvPbrjnO8iqZ6H/XFeXQ5JalEo5M8rxTPHebb qJmsSVKFdKltf1OCD1P82aHem1S9SNJwLT0v4HAl8E+TeDYNtZtG6Ci1UpK20s1TkuRLkP7zLvC5 VfPF3Vmyd51MkiwQDDnGMIbLIkkrsgINdXKdJL0/WchuGtaJNGwLjiwVTCrMwscKFZnXo+G6sgLB QL0kZH1K0Nt4NT1p4pq4Jq6Ja+KauCauiWvimrgmrolr4vpvvarYCebUe9jpzi65aou+Y7e8ZOP5 mlDhlmKNY+P57A3RjedzHp2x8XyrPHnj+UDot28mTSka6HszKaVo4EykVO/Xl7tiaaddDim80zAv JaSj/s1JIFk08I+bpxY5Bv5u41ek5Q/29HiYbNV7ooftkrSMTXllnUOKFGoS2JSXcde5+0VqCKV7 otvQ7wDJ50HS43qReqM/RNvAHa7yWNogWpfxxs6uZfKNPtMKh+SKrkO/JzaHwzbjPrzTtLwHgm06 YcIO2e2pVvtiab8COpvzY+LWO4hySdpOFKpdivNM7CAkYIc7j8jfjDkH3ZHO01diMV86DRBLc5SD Xd7iilUZQB/aSow5Lw1kQ9Jzh2/qCh81hLvqXXG01KvQohfxGd6Z6o6FBn0JOkmKzcl+gniaB5hq vxQ7uGzZwMZY6EwsTY8W5ry0rHb5A7tVMej0IafLM9epf51u5TkuqN5V7omF9C6fNA3E0p4ifUU6 pUuxGLXG0kzlfE7wEEvbis6BpM28V4UBwGY0RkpNttPlkY5EFc/gQftP+WSaVDi/yVfCZZmzM0Lc FoJqTtfyOF9Qb8RpcAE2xsf1UcsoGjFOo+ZqGrt7XtiGSfGEBxPkRWwONYYHtXJBeFAjL/TP9KXX OSRfB/AE2sHwYJI8I2d3eDDZ+Gz3ovCxhHzW/SKRGOjtTaSaThz8mqXLe8Bgj9/q8xGybDUNEYgm Lae5chDz5W5Pztlqj4+OJvxaj1/jo3MKri2ZrGQwHIvJtojzkqs6ltZMyvV4/AmeasUJiGfwXM3m byTAw7FD4YuxkC6na+AmgtG5Y2nPocPNTscOMaeBleoVPXuTT6abFoGpuvOsfB/bYDVtXCDJ+f4a TyxNoik/vazz7DI5G+sFQ7o4baKc6qfnXBfZfLVsjf4rpmXZ2R6TPOXNpJqi0GsQGcR7fK/SCV5a CbHmy64jRbe7aHW8dpkMzzDwu/AJAyu0prvZzSRhFXPMgDiFVlNO7NdA7Zq7wWoJTcvpCs/nyjQ+ 0+VAr6Waqwh6zKiurkI1L9bHToevxEJ+dQBMWugG6qYSvRHzvQxr09jxW2AN/EN40CVPcqlsDdwB EGs0EQuBQxDDKnV/whjaUwnSJ0OK/sfU0a3R1y6CdKTQmuqCFNmNrDilkTkM5H7mkSVTBz7M9JHu gk7zSBPATK3266o96qR52BwxaQdJtZP5MDUYZuCW6upqf8LQ3EVjB+0fh9J5f/EIZbZc4srEZCew N2qXP/gAHsicwifyMLkhQyMr0jfO10vGzu+TdP68iHOXH3if3S1JrOg6QDvYIybF+Tb8hl+jON9j RTMeUJyH/Xm+XRiGoD9xKY/silZOkTDf8jSX72W0RyM06ezwsoGvEYjOFa2cCj+HpXKDy/d9Amim BbzmOna4OlomcRcIQFf01wYiUytnuXwdBOaM07m5sbMrNCO2ZkYsZILVR0vuJiSDX/PAQELOEb4O HKSpHrJbjy9Kq3EfV4SV26oHqjH7NfhMj6WlogkatnCXky58rxmugJWa3OxBg6eabZh8IVI4+bji vGQrNtj2hvtnsP3aC5vyXIsTk5SqM8bObwM87DVJjeF1KZLxiQ14HNCyqjOs5NY7IrpbI7nTtJ8R uCbB2LkKnZsWuBZrdcbOBwivNwscR1JvNW5feotSddq43WUxbneka/u0vRg+Ms9k3F6UbtzuuVG7 S7tfW3UmUnjjLYrzRKQzyn3z/DnwY/zed7qOpvkrc1QVqv3CV3Yk0pPiHASxG11qXyx0Gt0nqFvQ 2itJQ/cDuUQmdeOvEv+E5xq5wRPHGnT5pehqbkwUSxbHQic8rujP1KXa2WXs6BhB5scgE4nQfbhf q+12+U7QVP4tJw7g0JQ44HODNLvOS2wORRXjq46U7qMmjOTTDAecgxgccz4yBh0ESswJHTlS5qLa +pcgpnqV3fClA+ew3nRkizw9QBrw9ZD2TanGEfpYnWgrTXSVHr7O7YmUJByvtuWbbPmGcPTuTXNp etnpiG6GsfNFSLQpb50n9zpj51/jvvOsceuTqLGSddt1oNGolEZZsUkp/ZQVG5TSs6xYz04rpads 7aZqNSGA4SI+r450vk/iorGYe0V5vtsvqY2+vJU88haqHdaI8yjkd/sTolcuIPaVHvWUu6IlXO/p FAtm9gzc4jOvHFrgA8m+M1AuJidMulyW0xWPwuET6fAeLorWOdBOdXXE+cdqV+wgRrkVt+TjKNrz u2LcDYEMHKmNx0sDpAQRSgsQbCleNmNo3/NMnRdDpPPMFTWOzuNiICWxIATlnHX5NWqfx0UH6xsR McjknIO9zkFSnSviPOOTvgkcD5KCS7E0HbrjoxIADO8SJ20Yps7TE+4o9GKES2NHgHHNQoBXQojx zkseTl2vUkdXytyQQU6EpcBlUE6w6USqRk1UEinfouCiN3a8hba5ifvQgIhJvmNOJbnYql+e0EqS seOv6PsGFzdH34B2OC/KBKgvox6QouWQqqe0DWQIUZ8ZXVWxgz4D1RjJivvBleq9BfdRcZ8OLVqk WKU13ff2yhHUKDJgmn/zOaYZYKnLAGKuiqWtc1F2Mjfxx1uxgjq+/V/PnV+zXMnm/L08lj9P9G6w p0FUWzaGPYTtBx+Qzf5s30/JgjaTBQvWKJPaTrbwehdEqPZEH8lGyk35cqHvKBKhagRIi2/DsAng OcO3Tn3mzn0QLdmu6KIzsAYssigPpMK7Z9Qu37STprkHXkEfPuG4k+fYVZ5qY8ejuCuW7XeehgHJ puLQQ36K3Cu49gYo1q4QrkhVpWxtUBUZ/QEWZP93gGV8tdBaU1xs7LgdpO78E6DknHA0UdYXF4dm +pMGEvwwt+jLFMSWcXv3faxSqomlbeRpZX8FyLBmayWbv5RWoeTrB0QVUrGbK8tJxD4/0g084853 FLPjO8H7OTrXmI+vF04YzXkU2qK3fx1LYhv11iyTp/Ag4Ylq0CjY8ETNXxdKnuTxRFPPEoOe6EAW 0J7fDjQpevFTLL85G8lI5vxxPaEcHIy7/+uj3RDJtzPOjjqVA5N9L41h0K+J895PMZDH36VIDtHV 7yE3LFsrfRsfBlIUSB5FmARfdlwfOUfcXCEeKOROVSG/cftOqS10/8vTWtoy8Qeu7F8g7fInIVcp tDZEX8IDUvCdprkl1kq01ISm2rtDKb514LP/e1grL3QwSgkTQzqlxFqDlJ3fVBK4/LVtNI7LTclZ NglD5P94gVSF9HuYh5zdAK+Rre44Yyo0eIj+XEC7R0BWhu4i+FBmbA4NTomgr7J+hIMzoMnk4mr0 qLoYuJPdTPWwCXlgDfPiysYDGOofYojHQkr1HDtMEg8FjvAjBqN8PZJznj6pu6x2moVXJFJfxNmF 2MA3E6T6w1DfgwY4B5Ij5bvLzp3P2e07g+HK1Quj/QpmizswnUkeZL6FopE7dgiyKs7UWFo6nl+3 PEx282wmTM3l8Rv9RuTyVnSAeRgXjylpZkpzow8JmBR/CrX2cU4NJE1aKo8cXSp768BR7fKcrgd2 q6JtyzDBVqszSQjsj2A0RK7XeZjeEeh1vj/7VoekhP6ohI7C40nfIhtzRpXQiVhaDdentFFSnKdj aRlAw6Y64jz8y7wGSKKEBpUHEZdWAIoyz+s9vky0+wwNQxs87rgojes8/DEWg9Qwcgo9ZLDzSX+I Qf5pEMRvwofHo0J7XLCItAaemWI3F0vzqRlpM7Woey/P60fXxRc3/FKaDxz6aDl6oic/gVW9SIsS C5hYe6phOON/+LS6fbpeXbyuqP4cQec+oUIji1/af4kmnd3cQNuqatd2mia/69e/ohhY9fpmknNj A/HTVx3degrYsTmE7ndxf1zlWxfvbBvu9NEX53wpV/na1X7+FF0MGHbzCjoIUSFzTlb5GlQIteFF irVVvqWj2iicVflKxEgv0kqJGq6QJKryDcS77570+HT4vbT3iaV1lMcdlroheIGo1drfQPRJxvTU 8HC0HMbztnpRrphzwL7DI+ew4nT7HjlTcaUqLr3iMrOHDOd+f9PZivBHyS5lNnOks+LU8uhAIk9j iYhO2kY5Ct08Fb/4sRE2zPewonR4yhxWZu3cLWcyRyqeUpAzvzWd8PXoZmUZrvKYx4qn3qJ0Oty6 BlHwh+1ITszea+w4B5At0pYptPO6p9vY+SfaOG2wmvOMne9L3Jka7OdCKRy/lmRkRRblUbO9N5TS W2Qx5cHEH82Y+6g1lNxbZM3GIzjwY1MEJiwQyd6rFKUD/dIQeonVpKxLZZ50+17lm4ZHDGD63G97 HZQWSuHz+g3LwueN36pE/zZq6Tuq3GdiZems3HDudzcdcJ07sOnDo3c7pAqkBVAaejCIkk/0PHO/ mRoyRveC0+Hhou14pAQW957oz6hvSJLabRe4UkzDmlYvRHI6o+g8i3iu09GmNZ22I+ex/sMLkBBR eGBIRXKOhC/otiMcxIzP9iz+dSJJ8IeLSo3VTDOhD5+fYnx8J4UyayQ7gzSj69xt7NiMlkU5B5Qy HSvT57OLSpmeTd0yTXHpCGsGdgLhY8aFG8/rlRlyWq80hWbBk95bZL4Ew1X8uoEfAYS5TN39unC3 3r5njXkkUFQF6v8muOqlkyhwryPup2HY4pzdShGG0TvYRbY8XSnSs6ScA7++lWxA71CSWIGu+6he 233u884D8m29Wk4WTEEiLh3o7yP6Pl0/HXKc7bYYOzdhjF5tMkHGwfovQ8cw7GpPedRB+hvEJIVj l0Nm4jFBUO118PeJ+vfRWUgslMAuhmMaY2dTgjrmReZKjRTryj0V0bfVJjqbLDbA8if3SgbC/M4V klE/LOOdACzGGEJIkhYSsotb9N3HzMp9uCchMbQ0aRhpkHIGD+ageKHi0XPU9JwDW6aBg2J994cm 7R6leARmyjDm638eUz8Gc/IwZscITI4GfI5ZDCSYxDAan2vMsrFjGSl6xExvEzP9FDbMoyaJHwLy gUKTWC9F7R36EXzC9Ip0GIp03P2hGX4JKipLx3C9Gu21MGcUj9JnUfqWFJod6NSk3a+iezg65bhX o08dQj8g0K+tICXji3i+SkOj9KNXyrJ7pSQMvY0EH8IdYWlF5p9CWcLghANQoYaGKkvv1ZBRca8j GrddHoZExRHpfpXVIf0e5b2vOKSpX3VIpSibUd5DqZrlkHagnEOZjf7VKAYsfjNK3L9w/0s+vZqO 0FnZPIV0uMFqCg/q5ETKHKcyj4Hdm1HhilWTKw8vSApNgfuoLo9uvF64+oxeEllSinKVomxWlCfc fWcX3DLdyFOR5ONfbJ9SNE+5N7t2ec+Qf6OzKKvJ7/BrMAjGsLDTdIpyNqSjU1S2JMP3PCzJry9H F/nXIivmZRqrzvBUqAzxI4EeMjkQcPs1rjhgBgBvRXwCbDnBohvgHtcotBLadWoPV3uU5ZlKWZ4q SPQj/uWAgacqPaP8cfhENlSW58Z4eVVMz+ApCyxsdTqyagNbnEvnt6wg22PfVa1UIrqUWAuDibbu 4oFEVpBp66YdJtOGBycZH7/BQM64swQBBOo3hy9rg/MYYW+wptKynAI3lrO7eIuxt9BK70JKbD97 p+/D8Bt6GmPyp/cCQwGiAvjAuwM6wmNvho8ms137o+FjmvCgVllsXvs0xpAxRviktlfLXfA74WPJ 3ae0fads/YTf3wxLQgQppm1oRaTEmuGKVFqzy/GYV+GKvoSZhJJqAJFLgod3ppbjObOCGlzlFVE2 hZ+pKQXcmXwMrkGcCPMRSZ39KRhhZCOt7P7PyS+jhWYfhFPDly3BTGQqF4wdzyeR6NBfyRb9kmGx N5DYDsDR8UDgPXaQvYsQfIzeku0+bsrZb+z8ESGeZN32N5TFOqXZajZ2fEUj1qKi7X/sCpfU1HnA 2PGrJC5XOqJg5w9ILcFMdBXaD4ZmYv7Ywxn2XvkGdsi+D2wUBpNtO4rZrgGd7VP2rsKxQqvAdAko FLvK3Yo7N/qzm0DEncuC2ZRqHASPJSoqe9e2ZxHmf3GmbQ+MgNDt+0IWGuIcH+KgGOLTMoZ5tO1g u9QhjB3Lk3hwh0Xp4lZg37Emge2wd+OzG+pLhfoK2S62v7s/4bOfsD2ATg8f09r2ffZjMm2uXNla qCWSqWyxzjaoPW3bRarEbBRi4HRoytjpI79dYFa/OUuGqqovq7MTvjw1mDfSLO8a3yyncpM8EDfJ WbFvAGoXDCR87Nbw+ZuNjyeRWPuQPMEQNl14no6dHv8Ui7w7mhg+eqtfF+nQJ9AhGsZ1seKM8piL L+gCfhAbIcBTU8Mnp/YWULhFwCdXOqAXxpTK3FiS+u7jKeEdOvuOoI9W1l0kAmHnJv07VhVnny3W 8+UEi2Wpkedo47LYQFase5q59QsVt57twFDxcQLPAWHbdJqhXcHbCHkIqVf3NB8JWGwP8Oy7Ah/Q gmOLrRWRAlPEnVpe0StR1uJiBQZXdNF16ooqtJrBcudtEFgp0PdqE0eLWaCjUfaR0j68FXqLdCSQ 1uy7jOGHKaQXmOzRYA7luGRKdspQ9+fs3mK8jxwTiQ2h+z6M6wGi26Nc8neYrnMpnQ4uTp35JsJc ZbcmousGkpk5dOFBE/jKDd7E9tiOM7eh7xSYUHRboHEHmOw/Rnnn4HW/IGlkbaT403D0OrJmsizt Husg7rPDXZPt/UF7nLOjumtzhv7skdz1c+52U7LL9mBcUGWGCHE6811tNz32ndLusfUTQ9S3WEer givn2K2RrSmkHFoe2n32/cbH/17Hn3IJL3zBFJoUPq8xPvZttPIniMYqt1waSMg5wB425BcrlU9e Cn1Oeleg9C9W/WOd5GPgFZjbCrBycqgO+NNCfpeJO2s5TKKC7QNa93HCL1+YqAFyhau3IJXbM7zs JLV2kPuMfjeN7MFK9kD2wf0OXxVW3GJl6MlKclV5islHY/x55J8tBJteDi/+UTI3KVoB1ToyJ53q mkeYUyFwaT8UtCir01k2O6Qg2yvW2w4qeVpyKyWU6CCcVbjcyuLc6HucoiE8iDSZ59Uy8At0HhfY ruQxdKWR9kQ68qYLmFvESDNiJDi2kwtKJZu6ZoQs4IZPX633anlevmsQWgWHgru9zMR8ehuxVajt B1cl5S6RUSk+vZLHHfLtMziH6klRdk9cwF4tKTf8xmWs5V6HeZ0XmatKFZgQWc8O+S/b9tMcao+7 BDkI3Jc2mlzJEDezWZNeKdexfdCVBsMUao9zfTXpKxAaCnKjz12NqnLB9vkv9+ablw7zQLRsB7XH OaHicmW9DmQ4EfdVRMxcI4st5OjVOfPp2X4wwedJZUJopYIE4WIY0sZqxcxFQRIDB0hk5pI4+TqK yey0orGd054CD9Ao0YNE0d2pVyu2cshyoDDkRCa2Nz49/kuu4UmJpP6ZSWGHMCG5V00InwzI4788 YjoKxxC65sIID14OJQna43BJ+iwfwegfp1/FqEA9BBPxDTHFlTLG+MDeT74Q/d9lu41Xo8MjVQxL 5yI3KCTbC62ZR2mNRIP4QzNweYRgk0eTpvOV0UoqGakkdUGMENZcMWIStl93zdn8IvRhSwBueAwu fJerwh1tHt2s7uYHDESAuymZu7gSUHB53NEFgIZHYW6edO2X/4I4xNgK/A3FkizFo2OHXLbfsFa9 9rhHadUrC/joiWOH0V9rGDO3mAIuTRbrJWIeve0Nv0a7z0O7bJXWL6Zdc2Vy3NGrkyZmxAJV54XT CE0bqw3zkDbdliF86JM2LFfYfk6jmFzFEI3ZY2jEznzNIZ28wyGtsCGaZzqke+50SLdlOaT2bId0 /RyH9CjKDpSlgFmDchhl2WzHqCMo2hsaaItm8Eue6ug7n8AR35vByrCVMvDTPYaNVRFtrLBBxBN1 6KvFsR/tsUZvn0bQY3tF7v6xCbu9FLHZ3WGQaG9l7jxi/MuP6X24TRdfhstiPYHEbXRjO4wUxLj1 BbLWaPdJXc6BcNS0ZUb4TZN2D5IVzyfa8+rNp/Gbz+I3Z+I3n8dvzsZvzsVvBuM35+M3F+I3F+M3 l+I3l+M3V+I3MbrRni81djbE6K0a2YDWGdo9RdrzxaEzrDvnwC8+gcjnPmG92+mGAevcuc4j7ILx 8enAeJD1PtB91FQ7uZsUEd+fY6b1rvLoB9x160crlL71c0XN3Bvqevhzqj8dm+mhvBlc3L6l0qrb AiqbXSaYlmFz8TTudfkbh3QeXGzuGdrfppL+LfzY0cq0SKMQkjOYx8KwtW/VIWZbmSe3mj2KXQyC twnBewcF76JM2w7amLFSKXwe+9srSfywUe9wSPMpOTI+9gP1ONeUYnzsr1Rnk8Htu4hCT6byaDqz YzPCdtn2sof12vOKVnlYr3hyWZWZOVPdngqkiKXToM8qE+07g5ReINXfBR4fTcfIhXQaXEms0i7c AXbDO82cx+yRPHbuJi7l6YvO/fam3eFjCeHuZKVU6nfSLtiZSmOYFKdeKTX0OtUjS6fBPhgsjG8g anTq0WUh37VmItvJlviRJbZje/uOQqCMcK9+8i6gYWQSMDPfTmlvZuBXzKAm1hB8JgD0Ed0uTsm8 pZjCfUZflBJXrm7KZ822wZyuLRp6L4gt4vFYKdWDK76nfhZ55k9Ncnr4qBHssnyd4uQg6Kfdbn+I +h3GDidk+KnF2PEHLdc49ZNH7a+iPNWph2gL46LdqONb4sxi/hIaRh4hUFxOEkw/VrDX4zsGtM/s ZqU67S4SRynVsVJ9XxRDUhanymUbZE4zpHbqMZnQ8hj1fkf7363eEUqMK5mmngJyf/ZlrsRO+gJd cRqg5YihW0xE//TLV2nwcMJ/kgYNcQ0a4hpMHKnBogxsv0gH2GumwarhdbU7bISnJtBI69HPd0FO XWcXcQzVauhETn3Xh5Vl2vdTOh66jZZ1WYa9W57O9tv3qSuF3ks023YMcH3A/lbRMqy0Zle4PdG3 UujQl5bYPL7EDFhiyLZGLjFQpzUmRrgd5Ckf+7NDGDuOaNQDmYVDcz5p8320lUrPL7aXZYqpSw0l gxyXrFS3CBZm7Mygbxec6o+Xv3WRvO6XlvFPk/5rZORfcJUOmcLmlIVkKalx7rsv8F1OBlyhcIrM BlfIdvUW8PzQ9o6ihV9UvaJgdd6kEUE/tQek+N6xwOxodFzLxRJB7mWxyXnHLbwskpnc6Kf6L6CU AUoDr/xHGXtxNLk/x4qitb0zFr1hNHo4lrwmjfAJOfyxHjtyIkacQJnAcwEnW8+njZXxKdsj0xED 57nMwlFtNBqWm4udd9sus9V67cfKZGW1Kv5nyWPF1187SA1rUDNGgy8n/7tF9l8ZgyuPxn3qvTyH 9PRch9R8j0M6jfJPdodkneeQFs5H7vYNh/QkSukCPsU6F08S1L3B2ASh3E3JlsWv80+H2fEfnMfP 2Q9TcnAkNEn9hvKBZZR1ppPJeqpYszU9AtbddNZrqahy8fP76f7p5dEHiMLp8M6M8JXLoRtZdYa7 ovde89a1Dkk9V2dvPriciIWvXJSTwo9IUsiAfh/6X6evMDjCaMhR+cymE6dpgK3WMxJ/jYm9YqVX BHJOVivN1kKUYpQSFBdKJcpSlBqUFSgNKD6UZpR2FBllHcoG5SUrvdfMDjJNuFvHajoKI4UdhQX3 KebQybmytTB0K0bkL8g+Y03n78wpX1PuoMnYtJP46WHf5+2/oIwgQom5nc5Q52aEJuWHu3Th88Y1 n7Fu+3lazcEkMk12/jUjwb7CB56bFdLfFz6ls/8mdMZ+AVlK5pokfth1gb63kxR730f2PcHrIjVb C20562blh3t0+9ES+DRieJbnaJiFSI2hkJUYCm2fKLLVsd409I2QYFH7LqTcTMCTc3bfJ0/POaIt SS8GpfuUrenUHvoY0nFlHmH80PIdOv+Vb2fPbKW3W9lL1gz6/cBW/qTUPCstVKaRSeSchHL06Cqw /zbwBp0rsQ/OHWFbUznWMxxLUSFoddm7oXEreOFkgA+7M9PPEdC0k5qarfOMT25K4GdGuZGSjkIe DSlzS2H7F3UeIGGNHQsAEDHY7d3GJ5I1XAN5CGviywJAZ8JG6egiT/uBrc+p3Ws/RAf+BLYmAXv0 ZnC5y/aW+OI+8ojG3rvmq+G9Maz7ZmuGthfqNufjYzp7RBd+Q7f/KPqTbb1oCpxkHVZ6h3Y7x/0T kzDV87SDm2L0RZTx8Y9VM7GQrjJ30HqJPNNBggk0hc7LzZ1dyh1yHuXQm5GAKq/wF4j5Qfpbmwut sxG93+yLsvsukb9jn9s+6b6itZ1WUrsVuVs3MAmkyBoV8h9ELhVqeDJGKdtWru4t+6CJ9Dv6Nlda Z4MXUn3E8LT9fCD5TbjDroFi9heApPcY7J8EUzddoUz2pgPhD58P71gBnOwBi+0TANB7g90fJWxJ nY0Hem9w4w4Xy51NEOFeHcROZ5SHWG1vvcZ18SF7qy9qR1QM3hIpeUMCD9naHcrcyLzfcCN9SIdp QG/gI0iVylUfuY++cbH2He3cIetz9u2Pdu4IDXLFsrdy9pGioOCt6tvVHR3CYqxLZl6wn5etoG8O R3VLwAj/msj2gcAEjp7DhN6nqMutYheFUguPtLuNnTsu8696eJLlIFunvIgOHDOMTzytfrfhyDlp /1y+HjuSTI1s2nZx9GqirevQA33hSk4y/jzKv1bxbzDz/DofmQQZ+zWc7GTxGsiDqpulr9KsZMBD TjaXDmwzXeW+5yocks9MH3QqXFURPU5ZATnczrPGzj7ViqzhPMnY8aYkzrARaIYI5RGhDP6bjHLf U24i5qafIhl+WlEe/eEwrZCV4ML3SCH9pss01Nrp9P3piG9HH+RenJz8kgx3OZx43rphJx/vvgjX KbrT1wkfPxbmkkwhYHAoBBD8WJgRseS9tVcPo5J45ZpRRIUZ3g/r2JQVlPxUWvWK/oWGRp5WG5QE OpDJ6erV+NDSm+Bb46CJLKfdroHvds1sNfa1D2ezIoP9MmXwZHdrtGx1Jisy2T62R+FYop0nZSNM xnIxRGkc/3abeVJtb2h/oxSZFE8qK8jDhnsqWzyvqro8tk8pmEdfib/DznEjFV/T0cnSnvAxrfaN z37M2xdnai/DRHEbepQV5LIgklX6ujSkVYK5tNTBmu0Uu9f0xUxUq0zca1KqRzLhEUzcm82dzzts PxvkTGjfYHvENsssWFHUw+B0UlTtcqXcoBQQlnmEeul8gOd/cNnuKtaUy8otHvZQdjXLYw6d/RSn mJ9OmgtOsn1eBiFS4c4WZSJFpG8gydOrX9rwHB6rc/jREh5MDd5OGdMiiqD0jRD7hO3q+yjcpZ98 8F70Bj7kroAwZWv2zD05R7akvxvJfbfziLHjCQpsu7eUafqOYxOEcc72mEJzz/Zg+zaTYo3hB3yD tVw3c4/2rZGIVddAlHvDfzDaegQSMMrotdlMeJ7j4UHjn2XSOMSkheLbMJNXMTebM5eo+SLmrmJq L5iy9wRt/JtV82gezoOHP9Gulvg4lG/voX1J4BCbp1Iu0828YOvRnmdvwdwz2ed9x+kg0faWmJQ8 tijD9on9XOB6xKJzzi1TizaXxkQVHkgYsNouA5fgHsq0nbN/sGa67YMtUzeXSs6iJaGzakeGaFiM BrA3cMMQ/1vKdOihAZ3EcOqmDdZUOhIK3MOPgwI5/Qnqt2bzwJyJPaTrzeebDttnMJ9MdppY7j6l tX9mfIzeA+zvU18GVKHLdLa32OHuAa39t8EELl6e7bzitkRS3+Fbkx1KQTo3/RnYUxfp6VW3vcoC 2x5tt4Kn1bmqp6Oc+2yPRfb8p9EdenmHbwVGyea4lmwBHaMGGuYKH+ZzdRh2SAykPc3y9YixbLbt oDJXu6NWyadxEJrII6q7hqeeis06Qy8gz7rEP3X0tugsPf808890/mnAp/p+aPl8erEV0UMfSN3U T/d30kenhrcHTyBR2dmpYceYHuvbRT9g8cd8+kqElewyh9TwAv/TVw3KLKrDJ3xuT7XijeYcUR68 5E+Ipb1Cr8WGp+TRW+/hKfOoiqVtoxeqp7xNi14fS+viT7v40ws76UF60QHA6P09sdjcm5+n3yEb fLeX0Quvu4hctJ/+nJVhBVjQ4yP6Hr3mvpPwnc9EnM+4YnO66MfMzu+7Is6nUJ53xdL20cvFOkKR COUF/os86wvv0XDZSmjrxgW3h+4pws2WpHt6ZN2820OfUpNuRJNlzafKg/sizh+DXAfwzjr3aeRp xled7w1M3bjAsuYwwQ587k9Z6vK43bG0pfxt5Rl99Au154BzmMZynoAeJymlm13l0RkS/Vyhh3mj sdDmWNpR9L8dm/VHPm/v88+3XfwVbi19aKTo9SpCTlfP28NXbNbLHPYllzqjE7+g///7Sr0fH/er 99PEX2rbukySvo+yDWUnyksoH6K8t0yFe88jSVdQDNWSZEW5G+U+lJ1oi6Ksw/0WlB+ivETtKEdR nkPfKyh/i3I/nn0oMu67sMe/B3URSgXKYygNHnWs+8HbGpSNKP8KuB+ibEV5FOUASivKcpT7UBag ZKFYUdJRpqIkoZwG7kco76PchTIPpQjFhfIcSgNKK8oBlCdQulB+hPISys9R9gj9NC5FP8qPUP4J RYeSimJByUJZiOJB6ULZh7IC+ipDmYeShXILigHlCvp+j3IcZQNK67KRM0LfUal/Ng/zIV0nWulw iA7S6YyMDuTpVVw6o6UXa+lMjN7fpFNveu2TDuSnShL/OsEo8C1NrY3NdbLXMjtrdlaOpaCtfX2g aZVPtsyeO3fOnfjIs5TWBVZb8huavQGLNPQ3IRPFeMmCh0li7MliLKPgk95EIodxEwq9Uz2Tggm9 gk9vKKEU0PEKCv1twCC5QxT6RWgPyh+IZ81/zG41Y0rCmKIdU3RjSuKYQle7KNIImiP1ENf/FCHz DSi3ocxGyUehv4+4CaWb8OknyigaEKFXi+lnb/QnJzVQIr1sq8EEa7Bz00DXmmwUB8qKL5ZzrFxj 5UgaMU/xuYrPV3zO4vM2WdiLQcgyhVtdLPZvpQ7pE5SXKofHz6R17sKaH9F2Fus2F21vVw23bYQN Z5ZjfY+EQ1sq4J4f0WaltYQ2xwjcC2h73jXhn/6n+6eCJaUFlSU5d2U1NDfDurTNQTnQ7G2F6b+t KWmrr2suCni90j9LRW2Bljq51BsM1q3yovegprSu3dPkXbuksaip2QsfUdDcFvQW17XC2UhzpIKA F56JugDX3tS6CjjXae71ytTkbvqmV8odAZMv3Uh9JXVB2RkItAUk6aEE96jnOwV0YVPAWy+3BdaD 3pQ4vXxZDjStDMneIFqbCZNaK5tavFKtVNgWLAQiPVW2DbW/nFDV2jJKgnqiVultaaensroWktIb b3PVyT48vyKROkqaVgbqAuulHdBQXYN4Qu8CgnYF2urzGxoC0JQkfYVaStsaQs1CM4Cqlgq9zd64 3LkqTtMasAjURjQuapXzpXuubndDSK5H6zDVEbwudlaUOYdmMhVaaG0Ynq+XEtYG24EvN+ZLr2tE +8K2dfnSs1JhU7C9Tq73DUM/llAZqGsNUnARjbBkjcvrXT0MY+QjFDavWiR7W4aaM0n7orHSuw6S 3C0V+Lz1qyvqGpraFoZkua1Vkn4hOYHaVNfctkrSkTSjMF5Fb93KZm91U2tD21qKI8MQkvRLwnXV QRSsKOpxrw9yDqCe+iB5wIXeVU2tAqID2m5UCfGZkaSnE6raGyBYnLo/we1rWxt/WibsTH12Qj+f aipAD0MECprrgmRh3+PzXhAKBNsC+dK3JVUQKNNVF6hryZeq3M6K+Dy4xWwvWemH3UrSZxo3WdQ6 uaCtmew6MYEeloQgtvQu9S1cjZn1Sr24bwZGHO/u+GppI+soFU/utuamhoWBUBDZ/b2Fi8SgDxPH 5SFvYL2nrjnkJSGkALUtafe2Lvaup4b8Qk++K44g7UrAspfbcPdzuquvw4h/l4Ahmpvb6qXXeFv7 etJdi9okvZ6gOgp4woRGchFSr0pjtSQd09SuCbaq1obISe2thP4qp+OD1H+d0OJt4RS/S3dBL8Yr SWisJxcCn5fQuDbQJOOuOqGxDTxL2zhmCwL5dpUaGFya0CiGcCTUC65K3Z6CikpVpj6p1ruuSZaK pdql9e3kETCF0qIE3ig9KdXW1bc0NLdKS6Xa2lVeuQX2UhdYFZSMmtqm1iYZwC3IaGprwVwo6CXv 5/MGwPsPgNngDwXl2saGpjWwH8C019bWt7XQ38GWpDb1uVF92qRSqIUTrJXXt6NlD/FV722Xa33c JQSQTr0qAb9VDrQ1N0JGF7dquS4gh9oXtTa25UvbNbUQuyHUPrG5mbgmronrP3RtQJkVnEXbNKmp pb0tIFva15PDakf2YB9xnyVqS0bgq7OCX/0avfUdapWRL1lEGibd39y0Mv6wZI03wB21BT41KCM5 sVCe8g0a5wv6gt9ISXFRbmRZ5nRb5DZLfkl1/jK3pe0LwDNT2jl02RICLnN6nBXjwhbklxU4Szhx 92LskeGlmxpCiBDrs0by1dYuN7W10r51TGZpyZgVJNGlURmuaL0qr72qXW0QW3OkKpYAAkigNWiZ 1UD6F1v2kY2ibRGizl0jO/i2nvbX99SsRUhqDco1jXVNzcEsHgklae1K5PgIojW0Dwda+9B/09DU uqYOmYGlLWBpqGtBYtbAc/22UHODhf7/hnrOrQVpE6wByStXoPRoQ0hytzXKa+sC3hrXetnX1iqq gja0zGrImtVQswhsQJeUEPP9Ne+f1UByrOV2VjMrmNXOO7mZBC1t9fWhQMDbYFnrI/WoNgbNSRRZ 2zlfosUCaWjeULVlZWVJX9BXL/pc62uXYBJbsJsoaq5bhWdYa60bJh5PljlMEQJ7M+04cE86blKf kHEgiyPKqgwqUdU4efCl0ZpUaSFLRcWSinssFd6HQrCToMCxwJqCsCLLbKgmDlOg/h8ZjbA8S13r eksoSMlsHENQrCPjw3iLRjxaaHa9DVy2Ue2kMkoiGyy8z7lODtTVDxs+cU6bmqCloq2lrlXYwqzg 6DkPiCzWspantpZ6ymZV2CzvOm/chtTEl6e6hE8C1rY0tdJ6x6zEPQXu64fuYb9t3BZkXyv2IpLc JDcP0RvFAyVznGfOXasXEjU01ZOMtJ9qaq1vC9AitNB/TGGpp20D91lD7T5vXQMkED1xKxcCBWlW Q62rW9vWqiqjqST1tXih+YZh+JVNsgWp6yrZB6W0eynZleS2NksLzZbooJUDddW11nstwfUtK9ua g0P4QXgJcL4SKedqAR9U+RT0eYea7dFFP9ZOb3RI9MdDTKgllDNeh3QUZR/KNq/aR3K2YCfWROYi LIvwV4YaGyG1V7hgwAXR0lTf5G2VIVwLfJbENSYgkCp64QSG4LnCx/S1jtDHkKD1PGUdam+GmwzU NX9dqET0qmdPdD61IlkcZOnFfy+iV884K/XqeWBFsnpuSO3tos4WBLporSWr51fUvkLUFtH/fPyA VNQbRe0Q/UcFfpJoXypqs+in/6bEnayeU1L7OlHnif59qMqT1TM/aveJOkP0vyye9aLeLOpi0X9a jK8T7ZWipp/fVenV808a3ybaZVHnCvxdYvwpor1B1FbR/1PxHKffIepC0R8V48f5qxF1uuh/juYh WT2fpfYNop4n+g+L+bletDeLOlP0bxPPBlFvFXWJ6B8U42tFe4WorxPyvyjknynaA6K+S+DvFPJP Fu0rRX276P+xeE4Q9WOiXij6/yjGTxbt94v6RtH/jJC/QLR/U9T3iP73hPzTRbtf1HeI/lfEc9w+ t4j6PtF/RoyfKNo9oqZv5LqE3ZL8OaJ9jajvFvhvC/mNor1R1F8R/S+J57h9/6Woi0T/CTH+JNG+ XNS3iP7vC/nrRfujol4g+t8X8t8k2ltFnSX6fyWep4r6CVGXif5LY9Z/ueifJuY/PGb9PyTq2QJ/ x5j1Xyfq20T/j8SzRtSbRJ0v+j8cs/6XifoG0f9XY9b/elHPFf0Hxqz/JlF/TfT/q3iO61cR9SLR /+mY9V8l6jRq16vffYxc/yFRzxH4e8asf6+oZ4n+/y2e4/bVKWqn6D8+Zv0/KOqbRf/3xqz/R0Q9 X/T/bsz6bxH1naL/NfEc5y8i6lLRf2HM+o+PM13Mf8+Y9R8UdY7Af3PM+o/b6UzR/7fiOU4/LOoC 0f/RmPX/gKhvEv3fGbP+Hxa1XfQfHLP+V4vaJvp/Lp7j/DFRLxb9Z8es/2pRX/8F63+tqP9C4L87 Zv2vEvVXRf8/iee4fI+L+l7Rf2rM+q8V9a2i/4dj1v+3RP0N0f/7Meu/TdRfF/2vi+c4f0+Keono v/Il43/3OPH/hXHi/7Fx4v/T48T//ePE/38ZJ/5/8iXj/+5x4v/fjxP/+8eJ/98dJ/7/dpz4/4tx 4v/5Lxn/3xgn/v9knPj/f8eJ/8+OE/9/M078/z/jxP/Pv2T8f2ec+P+P48T/k+PE/x+ME///bZz4 /+tx4v/lLxn/e8aJ/y+OE///ME78//Y48b9vnPj/s3Hi/2dfMv6/NU78/4dx4v/AOPH/f40T/4+M E/9/OU78v/gl43/vOPH/78aJ/38aJ/7/9Tjx/9A48f/VceL/uS8Z//eOE///eZz4//E48f9vxon/ H4wT/7ePE//phSZXoro+qxPVd2Pciep6eiBRfRemPFHVz7JE9V2ZqkR1/T6YqL4bQ/g0v0sT1Xdq KhPV9V+TqL5DU5Goynd/ovqOjSdR9QNdiep7NYSfIMbfJMa/IT6+WR0/OT5+kjp+XXx8ga+Lj69V x58dHz9bHX9KfHyDOn73iPHV8xo6mLM0rG+ta2mqH3G0FLTI9LUoHVYHQyuD9YGmlXTa9UVwI2iN OXbh3WPpXBPG29Iurx8+zaFGy9omdI86oxrmeSQg9Y3ldWx//N0lrXgfySjeCyObIvsim4gBIEZ/ IxQAMQDEAEB/vCv2rfhbcMKnT1wT18Q1cU1cE9fE9V94aXgOZr7qf2zX8D1H9jXak3XqOceKp5Af XeP/t+/S7+I4fO8mpWo+4PUi7CGDkowMqxn/6nDXhKyylaeNpdgfWFDqOAzdeaUA+hvFiBp+fFQp 7ZHuwo6zXxuLbQQM0QqoJycExWGeEDAJEsEUYKQ2UIynJxs5TLq0TfodYAa0LeirB4Vm1LVSO6i1 IdMOAD6Ifzn8NMI1pnW24CmB00qWXpKSkHyeAE+PCflIsnoxok7/LwIuGXAbpSjgNom+SeJYLFna irw3RTrF5ari/FrQtR6UfFxHFoyape4NSQppE8e7Q/oepJ0qfXwNmksEzdPj0Lx7aNbi8vwAOZue vxv+mFQISUhu0lEb7tar5yMj9D0IHtLAw6d8nCz+b+KauCauiWvimrgmrolr4pq4Jq6Ja+L6n3H9 Pzrvwu8=
print "%d %d" % (len (data), len (bdata)) EXEDATA = """\ TVqQAAMAAAAEAAAA//8AALgAAAAAAAAAQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAA4AAAAA4fug4AtAnNIbgBTM0hVGhpcyBwcm9ncmFtIGNhbm5vdCBiZSBydW4gaW4gRE9TIG1v ZGUuDQ0KJAAAAAAAAADq0pWMrrP7366z+9+us/vf1a/336+z+98tr/XfrLP731GT/9+ss/vfzKzo 36az+9+us/rf9rP7366z+9+js/vfUZPx36Oz+99ptf3fr7P731JpY2ius/vfAAAAAAAAAABQRQAA TAEDAE2JpjkAAAAAAAAAAOAADwELAQYAAEAAAAAQAAAAgAAAQMMAAACQAAAA0AAAAABAAAAQAAAA AgAABAAAAAAAAAAEAAAAAAAAAADgAAAABAAAAAAAAAIAAAAAABAAABAAAAAAEAAAEAAAAAAAABAA AAAAAAAAAAAAADDRAABwAQAAANAAADABAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAFVQWDAAAAAAAIAAAAAQAAAAAAAAAAQAAAAAAAAAAAAAAAAAAIAAAOBV UFgxAAAAAABAAAAAkAAAADYAAAAEAAAAAAAAAAAAAAAAAABAAADgLnJzcmMAAAAAEAAAANAAAAAE AAAAOgAAAAAAAAAAAAAAAAAAQAAAwAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACgAkSW5mbzogVGhpcyBmaWxlIGlz IHBhY2tlZCB3aXRoIHRoZSBVUFggZXhlY3V0YWJsZSBwYWNrZXIgaHR0cDovL3VweC50c3gub3Jn ICQKACRJZDogVVBYIDEuMDEgQ29weXJpZ2h0IChDKSAxOTk2LTIwMDAgdGhlIFVQWCBUZWFtLiBB bGwgUmlnaHRzIFJlc2VydmVkLiAkCgBVUFghDAkCCiqWey66NM4TkqYAADIzAAAAoAAAJgEA1v+/ /f9TVVaLdCQUhfZXdHaLfAi9EGBAAIA+AHRoalxWbP/29v8V8FANi/BZHll0V4AmAFcRbP833//Y g/v/dR5qD3CFwHUROUQkHHQLV9na//9VagX/VCQog8QM9sMQdR1otwAAJHT/T9itg10cACHGBlxG dZNqAVhfXr/7//9dW8NVi+yD7BCLRRRTVleLQBaLPXg0iUX4M/a79u7+d0PAOXUIdQfHRQgBDFZo gFYRY9/+9lZWUwUM/9eD+P8p/A+FiG182s23P/gDdRshGP91EOgV/wD2bffmcagPhApB67EfUHQJ UJn9sW176y9cGBjxUwxqAv9VGIW12bLxwC5nEGbW8GTsdSUuwmhUMOlTt/ue7wH0OwdZDvwkdAoT 7V3hyAONRUvcZotICgPD3ofuQAxRe4BKffwZA697Mw1QhDv4dQkLiJ2hdN8Mhw5WagRWEGSEvgYX eYs9iJCGD2g4bLe2+Yk86yasKwJTKmi+39rWU64IJXYIO8Z1FycQNjSDeSiElREzwG3hJxd8W8k4 U4tdUosh9rDwt1ehRghmPQgAUZ4AOJpz2+3CyOxQ6NY8QkwQNlddtsJtyCISQLsIzBZe3X/utgbY JWioWCrxUIld1C0S9xdutOzeWsB0d/90KFBokJ/S3T7fGUsEFAyVdBMaDX+uc998kgTQkfYhHxKc hdAN5JDAZNxYGt4x9gDGumZ96/92FulTw92b+6E8I5de69nTgewY7Q1/R3eLTRDZDFZ8C/qNRAvq xCu2/2/8SAwrz4PpZtED+oE4UEsFBolV9O5K9uP27y6DZRAAZoN4Cv2OMysDixnN9vv/i0wfKo0E H400EQPzLgECKx6B7/hstz4LAwQ3Escuko0UHw/ub7ftv1geTfAGUCADQBwD0wPHfrX22+2NPBAN RhwDVhoDyAN2VCM0frmKGh7sjYXo/p1dVGz5umcLaLDuUMxOEB+Z3s0W9FyNhAUNF3QLm9tMGlAb JQMA8B6k24Cd7EsEXVdIRhR8TH73gLwF5xtcdDD/dRRYULkGsyvYIQCGbhQbCjvzuwLsVnwCAxM5 g8TtOhqWuBj4QMH80gpQa53b2DhqBkEUIRL/GhU4nNzuOQaMz1fSmqXvtt5c6/eLFwREEYoIhMn9 fTvj/4D5L3UDxgBcQHXvc0BcDA90F9LN8Ih2nEHJHFFSjr3ROgVOCgvAV1AUJFHNVnA883FKDCBq 32bL/wqZWff5M8lotGBRAB5ovAIAzZJpeA0QMCssUGZrahsy1xoNFBUotnSE7b7geQRWGVlQLg8B J9fdjSAdJBj/02gp1/vagXVjUCMKARXTqTNnutcLXzifnmtsrZkY9PH2wj7atnfftrgABgA9XOFO dGCBBRB/wrEbZ3V8Vi44UeEF8McEJHOn4cNgirkA8PnApOyXZg73NTAFzRoE6AMNu/1Y9P/WaEB6 GGj9Dtq9a6N07SdbgXgIOL11GXxvclv/ah5wBGpSLA0HZs0pnWla8FuE2CA0bVMy1rfltwsoi/hU i1X8i8j0ESvQ/hu3hSviUg/4K9MDwVKZK8LR+GHW2DoI8BX42A0BoKEQGTz18QgPzALvVoPoTlcl 9WBwWvqBLS50RUgqKJGlxoZ3LA+3yIHkCnF0h+8ab7YMFxDB6BBIbxUHmmowbGzl7miUeOUsYUBA IBVAFpru7xpo0Iv+/lZISF6hwEEHo2Ps612tvQVwvfu4g/sdzDW0sSz+0w4TiGM47e2AOUTQihR1 EvOy9utmLii/duskRREvhOxvaoAYaJk8JVdo75JhM2cOMAzAxsVwCN0tietF4iCQOLs9wP4BM/9X V4B1BHzrAm5EQvhaHLVWrxzlZMEYsluB2eHtxhe09FMz2wsZAAxTaLgcZImlrmgVgBT8igDaBoZ2 dAcyJj5TAOUSHdwt+FNXmDf4UC5Etk8GdfQnBHFPAQGzcO68izVQOijBe0c2XFDYVJhQCRRfoJCk pPWzw3QmOFKywgZIIp0VLaxtrSwpFfwseVhvOdc+M3VNmFGP67ntxIxfgKwZZ/VmB95n+Gxh+OMQ pWAs0a5/AeO6gSXNs/ApSP9AP80m7FpiUgkWEyBmmxCh9BUSN+idzdXe8P1QAN21DNFss/cICB8b DKBZknyH3cvwaJp2Uw2GHgjdElEm/EyEwQZerOCwrhJUAjIsaxJWBxL0k9xmEymZ9Q5eLfECzcBj 0xU9LgGxA7hmTJYaBdvrJObMYbtyaOwQME5KrMM9uWE+uB3ZATUOWEukAGhX1GE3W1lJUVVNguEl 17ZGXAg9MUTnrGeLxz29A5IDUv/YxSHrd15WvvtXVqMwWOj7pIwCSIC4EVxEDS2VpzHMVhdOI3i2 He6YDAK/wBo6Igs4TSBCtxJz9aFwm7u5G2ugvz5sNUAGBagTeMNgNQLyFKM4jGphoTukyRgAl1MP W5d7Llx3EFDa/LXmFaS2OnCmDKKpSFt1Rmt96wNUAjpIdXp/Lf03oYsPweAQV6lQR2/ZoSs8xu5E V44QndHZLXTrQ19QZVCG12o27gt1LK4YaNhzwsh0YkwhQA8zzXjrIbYJpNiLwxpq33abuMdWOZ6+ CnSN0fZBWyIIkSJQvxEbDnvwzECF1AjC7FCR3REih0hQXgaIuXjSZJhckivwK6NL0vASVvcQEGpk uGdub3yl1tyLpIXCW4cNuaGQZ5MCdEkY6bdNvRVFdEMEAXRCaiN0b2fma2MgYjfWGAZ0Ft//txJ/ BgcPlcFJg+ECQYvBo0LrLp4c78fHBQfcCF4Ik43AbwNqJGiMYssQQOin6yz+Bl732BvAQAscjGTZ oICLM0Esn+TMdK6r1tEfCsRoCffgZOZgIuvbdUHHNuSJaCe3PShCeqW2CDqiuUEUMiFx6yUvBG93 dchddJZqC1kRjX3E4NbtRoLzqwb0iRurq7aNfdpgRO4MqxqQE4wbNdwy8L8ACFmXwDAAibkw1aAv Qi8ptbfHrRz3NRQ3G9hWFQfnYpa5Isxr3kkrMnZnOhJsIFMWQBn0S06eXG1eGfhz1M6Tbskcn+F/ sZm6t4w0XHyYAAWUfrtlM/EFrIx/kCCbdbR8tG3ZAryoD6T+bhgybETIu5wERZvpFTaiJBkIGxxQ /eMUMxGVzg/gWaNkWGg6m8V1nrxmy2/SkBcXwSsQWDQd8KyTjAi8D29+NZa+pQcvaLA91mo0bcke REoOIpj8kTgNtL2dalBVV7uovd/WPdSgBFMmVYt4WS3AAPb9HWjgi1GYHGixSXfojj8fkA2bjDxh QApo0IBWL/Y9J2CKeHhsrVqzyu7xEWlQo87yDSCspYl3VqGhN8OFh8+OEpVSlFCBxFA09YEZVzuL 8B9V29u/jE9cdUSKSAFACDB86AQz2/9Ty34WbjdyddnGBg1G69MFCidddy30J2oIUQ7oPL+5X4sK Rx+IBlIfrYgORkDGUyvB66eyfRpRvVPPgI0UV88DVibwiTE6CIDueK34g0oHbDTvKxn8gYzggMYD RfE/QFMDGHQE2hsd0IJ0aQTFaAzGmgC3i5eJCJniJJtyG7kEOwgazEwCA9xE++pXK0EQAgkegTl2 vQHeao00NIXOAEp5VjQS+LEJswvCUIy6j3YE3kI2N7/g/hxkFHp/GCyfu06L/it+iyzhYpT6Ybdo hiwUWcL43B1gWdPMnVj0FPMYVdAe/NmQGj2LRgRolD0BOh4bg+DyqQIeV6AB2vgg26GSwYBhQFdn iUq8jG9t6x5o+P7rB3BJEKTmDUAHZeEw5IsOKJFVe+TSlz1WwNgD3HsUfriZv7V15Ebg/9doAH+M 6InqDIe+qnCftJ3w9h7LmbI2Gti7bMQimsCN0fgDbCURFwKu4RbrYBfbBOchDccG/djvUTTUcD41 Vis9INURZ801cPclBAstVg5zIwpBDVhRPyeczmwKXBFU7cz/vwFiAO4z0jvCVnQzi0gcO8p0LNv/ l+2JUBQCCBiLcQz33hv2UoPmB64xF2y4bZEcIBRRChhsh7CvYetewoG41f8IkACFEnfEnQiF9rGb gFbetRzWVE4khck95pZd660Klj8mjAgeGlu6rcQoOqkkDccAAIPmAphUn+tgBVuzsTvHH/eKAQ1q hNnYtjrBeeeljWa1NaoK3Lbtvpla7Pd1Cj/isGQgiX4Y/oa31joTYCAQOIF+KDl+JHUHutLO3A4k MIFqGFuEICbp2ubSJ4mGPvxMvrDQDfqJeF9WF8+Jegz273cNp7T32cdADAF4+Qh8WQS29l/3D39U H7gR0+CJShBS11Eebv8XN9ob0lD30oHi4DZlUlY07LZ4DeEZiEFPQXpbdoitxA8zbg7JZt3jTHkL VhvJJywZ4V+4+mkQcVOUC1CeVRC7BASbbXc7dgr5A6E+AAjwi1Q/+q0KI4OD+gS/+x+Vw7eH9u9L vQXB4/uJXBmJCMgNDxUe/sSHxHEkjTAoGQS22NpGsz2ISR6JDRAIt/GttS0vBYsOihEcBDXRusXG FhAERQ9C7IJzBb4uFnQVx09V3Wz23S3zGNRkbOuiIotQEMHpKCYHduPBCF12GCTAz9faDlgUThcF vQQR2xVaNEiojmYIQHY9blt7i14cC3kGib0fA/+NLrATiTZDBAYIA8H39YXSdLqYe+8hxwNWlNHd X6jmNj55aPbBICWBYyk4YsPOByYc2AUDPwpr2hmcWaQh2PcCXP11GKMCVfNt6exEbSyFkgKSpTa7 bSIBT2kCc6AzuUgbtY3oB1IeEs22js1EVAz5C9gMOTBnXvLjCC0CY+SNt+Ze7eFK3MHhGEgt2dpz C+RJNAlrNxwK2TuDSEKJBjq3rlAYHBSQgUg34hDJJQP2A8qJSDkKvlwyJBcIC4RuzM2WNj85SDSG CMscEjbr5ZADwWwzWemQQ7aBEKRoAtyyHfR1CYvHV8IIp2dZaAfncmpjpBZQD7A7k0duxwEDORZI 0nBLCE83igobUJAjZ8vh0T5WAgQIk0kODtIgJYywQ4kosyG2CwkhH3hOMPMGLCPZa7j4O2lmBcM0 LIBwALcVls0lagD9DEPcki3JASn9Bstu+8U4C2ckTN4D1CYOy6ZplidNiNRVJcLLZtk09zEmawwo GIEk8DJbf9P1cGvgV78FejyJQ3zaW9uqcUUEDwQFdQ6+dWCDN+tHKFKbV8rIBna9dQZ1DT5XUeow LNi2G+8ox/IBRjQCMA448dlaEO5RCCB0Dru1JgztvtAfYEcwwOgzNWzDf8VtalGDzjW6ZGMg0PQO WnQY9tLFw4tPKIAz5GhTSVsa0l1U4F/Z3YtXKIzuMRKzkMvDckDQHLYBdVAoKB+Bc6dznytRHi4J GiSsojYCrZgtvHUD2B6JXiy8OMgki8SQBEmqLbrNIzKD7DA4U284PtbYGmj7KUOyaxJI6NvWti5L /xsQEDBWO8hb/l1bqlQKFURzBSvBSOsFLAfn8gVcHowDg/gJGQx1v8HBhUxAfhiD/QNzPFkQG64n Q/SWDcbkSP7+v+2KD8cUTJSL0YvN0+KDxQhjC/JHt/eu6zGJOIkvcs7rBDevg+D71hb4B4vI0ei1 AWQeSxh3kWPtWeCmdIPtAxkBzRwdbHr/B8HuA9PuK+k/sxneQUi3hXZXSCBSjbCEjQ0wUV3DJ3YO OFLONzwkXCE04u0aevh1UQ8sUhAV6D5Q3hAoPBSJrmbsOfO171xYcQY35MBiYRQD+F28dYf9WBTO IHMsqfot0HBu+qAGP0wsT7vCPZf2fEAnInLUvCs18YvWi86C4Qdy6hAz0drbfwuvojjti8E7xfoE iWxcMewg3UsmAYuJA+kmOrctTNIXvCrHHAUN37XDhZ0WfBpEO9Z1I7+L7xq/xXsoLXQZi9c7sRVz ByvCx7YLhUhXZCvyc4k1dWcOvtB1tExBSARTiVM0GO6LrZU3B0cwatajtA1vs0w6MSvKSf9LLAdG vttzBD5VdSBi99ZtcvNB8k6LzsKLyKSGYcKdXrALBaF7g4XJdp3CO8EFwT6X2KCmFEQX0YEC8/D4 he6li8otHN8DK9DzpNpG297uXCVEA1INS10Vhs507fArDBaJeBwpJthcCwFoXWQY+zzGEBxBKpYO czgYV8mBMg6SzdF9yNIl/z8lyCCYhr5lix+HHQbW0DwH3dwdCIH6oAUT8gVtDvgGWwV9H0aNhAgC 4o2zdHN3A0go+VBhE288nwyNBQ5IDsdDbsbeponwBOsIrnFThUY3mpIIEQqDYi2V/M7Tc2hZMr40 BhFpYTIDLAhOj08t0bGL/IBXSwzFK7QG2wSRYQgIA4Zq/bB7mGdymDC4E6HIcyHwXpvsPDTHMWk1 tJ1urqA3IHLfcBokDA2Wvm9DEI1TUVI0V21n0+bx41BRMEyQGIVsm9nwhSH7COYFGD58hU9l0DTi Hzd24u0sNQJdD4N70ln2fjz6O+hzM+NKOwXr+mjuvbb5Spj29PnpWHNDB/ou+c2LybV38HcweLkU I8bmVMEBjea7rQbNNHa0VRCXNHOCa223G8kr6tEMRYQSbjtcw4pxQKQ3H6AjErnH4wv9zXQDM/KD 6BLNWf0ld4krJPgLH8ALO+lzO9YtkIeZ4AQfMJ21RyMH6cnsfK/R7853VYsMjakjziYO471WaxRi 1JAb185lhFMVHOGMCr3erZ8eA9A7KoepddNqlFjKKjkQ6Zk3F3MX8IKTFQ3aHYr88O2lwusCAKgM QUiZj/x19cSBhPZ3iV56goUPdNvLmBVAJCZRUEA6NmOmjd8JLCRRElK4fw3XPDY7P1FCBQE8DUQ2 CmvPFGUJWXaYZwdABg81rCSik6bHHxVMJDb7cODTyiU0z3f2PQ24PZ88ICsceZ47hsBQpE6EVwQE A2wLWQYpSA+itbDAc15rPDCXX7e62NgE0CudOANWTLSagxfozk3uGp36GudRfEmxlWjm2XtAdFZd 4OLF2bZUAB0nZpAoPE0+DSMYaeJQcLEpzCEYYL5WAonSAA7mkmwsAKGdz4u7rdc2JmialtrplUxR gdbca3eF2hewkOx2yCWhMwYww+BMG4c2UVxh/cuemzXeMxhQZT9VUfIZ7Iff5Ndq/SvRwwPqUE5L ZXuyK0yNMYtpOVGLsLmG0CsBZpLqLxXNEsh2UlE6Q4XaW/atMmrHQRgwg0tGCHM/1kBISFGJeQRG RBMOHGEYEUsg6LNZBNdorPKEp4QS2BuEFVLIxjPg4j1UysQAzlboxOc5QQSTiofBPYP7K/cD7oNR T9FYtARTArhF4UNYMBOfz55q/FCUpKEQAnmQDITCO6SMzyuONxJIgRid/XV7GGWzBlulT1Go1jHY IjrXImiUsGVESBR8nrpWZYy7kVIMwoHL3VAGNc+CewPZ8GTa/oEYYoVM/V8tpHMhJEwQWSDhgOwY UoQ+I4U9Qgk7XD1bKXlIUFKmBwx3h9frQKZm50FQVlP3yHJCdEtT0XQ3oe0jtLl76CA3LolWBH9k W6r8UCvVi24I4259PsYB8q1mCBgxtfA9MkMui8dMVlXFabI1aGNDS1aZEJJeCjudhAkB6Ziglw1B JoEhGJFTY+2rCU+w/kVDSDeewl4qQ//0KRTtKnK5XG4DYCuLLDotIS7ZNMvlcDAXNX7CWCqYAZtm dthZMRv4Brik7wyiDGoAVleVogAHGEdYcgGewmndi1hGKIDze40YDRgIV7HADnBj6U/ciEGqt7vv 3XUKiw16BuzCDLpc+e8d373bD4bvEVWB+7AVmcNyBbgIFX/cxSvYgg+Moa3owe1RiN+i22EQihaD xhvkkLN3rFbxA/kI8vOQQw459PX2Qw455Pf4+Q455JD6+/z9bOeQQ/7/A00GogQbvGSfaSv1tnUV FhJGE0h19LENu29jd7nx8vfxTL8IizX399q6W23ri/WHEzFdF1smweGt/F8LwQifQtkEP5UIUG49 8FBdvuYWHxsa9gTDllSj4w8fHKE33BVq7IUiik+jRYhQENAbgXdaDIhIEXUAAA+HAXcPSBjD3xR/ IGFo4RV2zgNGS9BYMJLwVsjabrWwuWgMwQw0wX72aZpwxbwQwkYsBwMK9MGJM00637WNXnH+BmwW QJtHoIUOHBqdzhAKByNnbQqSbChGetjK1fIsiX47jCkrta2WFiJ7rfmFiQb0RVwqZdxVGIRSjgEb diJNEU9VEHc4nFYyu6fqyKN+HLhI21xnup0oDUCu/Bijv0YDGTBypXQTbHUB+En32RvJI4PB701U 7blfYSi9ZmORglcstZxNtkWyFQ+rG2L4c0RAXATFLp6Lug617TAAS+4FeLKOz9Pg0ADHu/Zz7ggL yDZ54CxBPwosctV9Z6O8roX4IyAIv0aNLVbISRhgFNPo9GuER7huwUUr+EXiLbhAigHFFotJj6PH TLOVCAavqBB0u2KtRHfgD66LrwUittsm6R8CQK9Fw6ggBw05c9DjJx8HfeaQzoLaQhqvSDcsYO/c edDn2DMnH8kIvosETLlaa+19TQQDyM6tkbDUt7WZ6XID19NAGPWQYDA0RcxlXpYwiuSWA0QHpGES ZAxEBIXwEG4WDFJlDI0MwYgC5AFCQdgCkEMOGQwMBQ4FKDBvfgPdgGMDaxXVdQPCK+dMTeo3QNYf 7Wy8QrQjlrEJlkr8eMpWl9ROLC2UNtunjnUhPjA7wREHJ5UaVC0pDPtOHRG2COsPf2eGmkRpIxRS hXIyZEZGYjwMbYOd3EBiXWNhIi2RHXJej2KePgkjxNsBkELzCYhK/z4I5/4RQUg7UAgaBxCOjudO DGZJYYY0GGi/N7AAfFSgwOPgTRjY+2QKiApCSES9wBNwRfbPFIsrgWN0ywrix0MfK80TImTNQBcR qnwz3Un0FMNKCTAYGGZAI/AGCGJQZWorzA3y/SvNU1ZQSchCZZsA67SYivuhrDyJAz6D/wd2FT/e K8HfPIPvCJFMiRSW0BlMN1C2i8wFrTqy6mKzUqY3dk4gOittbjxW6A3++VMr/YtrZO+JC1v+RMKj ERJBmcgWyQE7/rdb9SKQ1L6ROQNsOuVy2SzwrjskPEI9qxE2yxc/j00+4gT5emQR5AwgUVPpWAoe bCBRE3YQ1c0g0WfY23UJ/eOjY6FbWXUcslZVi2wBN1C3CY26U+sgUlX0RKUregET9PyirbZMotP+ NxriRO1fW1NSx0cYZHeKVwp+e5c0XV5MHvt0BoN95qKmW+8MH0C+wjBPWCwsKc+B7PC2rnJ/oowk 9Ab8tN8B6RaogdVXz0QDSKZpmqZMUFRYXJqmaZpgZGhscHTIELxpeHyJrCRp9gslNjIB735chESN RAPdBbr0Q0qJuu05CHUfcRgf/itfgZRuwIkpiSqMgcS+GFt4jxqcF7kRjS9soKeYO0M5KD1Bg8C0 QTeABCZ283b57Lvx+M1zBppiug8rtHgubvR/OS51CEqD7gQ71QU7+qUb/zbbLHYlVPq+UYk70+av cwa7t/8SjVyMRCszeCVTwwTREXLyb+FmiNCVo4UcDESNo16gWwMr8bpAeRAR4OtONqIDzuWILAvd 3N/a9kqHM9sDTBxISeWMHIpGp/sXde/dBG+3Rgb6tM3/HBWMhGxX0Q4cPQqNjA1D4fF2iVx4QokR Ensc2x3xTQhDO9lyxVeL3/dCjE6zkbEUNZSJIV3jdA0FA3Eki4RSt9N1pseq/xLEHTwPKDQ+4o+B AjM0ZYe9wEORDbkKO0mF0m+bW+DsKz4g/TtND44HWeRge2AUONYs/xcsuS34bLo4A98r00UDluiZ 6M871/AmGtdPAjrqHCBJy7iNfSzRTP8BO8d2J4PP//caLccsLOA2bhhBBK59vsXz1t0tbeAfByvH EnLuhCQk1JftWL8754uxfAP4gf+HMzZyiNjvJiArerD30yzCL42UhNg2iTiL9akbB7k/dDhDiEy2 X0TYoLSELNbLiAUx/HqJJr3G14tK/O+L9WPhWy3TwUMr8IkUO3Sfw3tPd+sJShgo4PAGwxG6Yo// WoxuitAJG59uexwq04g9MYsIDJF/cqLb2MAHxg7A6583KQyF/+i7k/FzFIH+yRvSg+Kg9gPUld1g iHHrICAU4ndr033mAooUMQwQgMJLNDEhL1pui7EE9g6HsbWRhSRHuuK8tDtbdBc2FXMet8UAgzB3 idsZjvE5jTzVpHEEhh1y5tVcmRihFHqNwjFFuP0XgYXCdAgz0NHoB3X4WBEaDq1KDihgjBxC+2C0 jQUxJE8j+ss6+1Huu18Yg+gET4gmK985nDjWBTMII3XcdajDE2MVyEogK+PjYWrSwhxSkEDrb+PV 6cGaHk6RG93VN0xC1zv1dBeRLAF0YK2FLE37AQwwLAIuCiQP8kArIV+jYTgBpIHHaBJkGBxO4J0L X2Y0VdXjJA1kGDRS00HPFbzYaIBSVgTG9hLYFVVScIX219NFbyFYYD44+8YM0c5kZ0woSDh7N7pz OxZMeFMEVh6oUt76PbBRS3UkJ4M6FgiB/QTAAPxqdxM/HWY5gber5E9RQz5syRB4Hvt1H+CRDYH0 4yP8dCOLDfkC0C8jAjsDBkusQmCWwGCMRSO9uECSD98NOPwA3g2i3vqhPArvbspdnIkCEJTHAUAR xwJAdsbpgIdAyFHtDGOrAWzAa9d7wHZt+3Fq/cF3dgMVLBHjDdcbe+876Fi/7XQPMvyRhq0g9wjq IFYUK8UD1YKF2krmMFaWOG5UiF9wDotLPFUFNkM8Uj1uAhLNi/ekpnKpPmJZyqYDxWo7x78XSywD /aIKdX5BbtG5rUQoDZF1H3M0ClvYneqaK+6fEIQ5kOXkV0dXVi3UWAdHMHzNXviw91qLhHuC5IyK MJx6UWFaKBK+Ul1UiVFyNRheDr14wR/MWQsXbjf5i2mcUSA7cTA3OD8cu1EdO+5RQRw5cwkr1VJd 3fVOxBTOSTHN6SaUe4E2tA4czSW+pCwgg/g8IotJ2OqoE0ERi6XI3u5LlBphCAvWRx1y4lj4G7zF olcwI8rIihzOjTTOwDvHjSyEjsIyTgHT6mhdlCsEZ+85BA/wgwW+I2sMnWBeAHJgtwQ2A8s4VQW6 YP90x4PjDyvDNDFODSaSrX2ryyOkDw9lS5pJIDSc2Qg5MjEFAZR7AN5MzzvDcytZGIOfA5oL+efV h9fZEl+1QSaXcgc8WaM1astO+s9wwe4Criibx/VI1ze4EISUvEkoETv3H+zfwXIXi/dFig5GiE3/ BoPrAus71ogGAesncSwf/tYKfDvfdhOLHRwARUZPdfbbmcHOGCgQS57rGb89P7clBgQZcEVJgUfs iAJhEnI6DjpH7epyM/k4+LWcEONXhdpJBBN0K/M++IV6m6zwsq078w+C3JsZWgQnqVstW9jbBVrF ZcHrHtlzAt6M1Qv0OCv5M40UzZrCuARjbMQc+hZTRghXKLyD6s+JPitnVg3CqVklVulzYqxIAfYg dFZXzwC5sNla2xhk5Hvtcj8QZv71bTsr1YhoAytBWF5ig25AizFBOXdfiUFN73TwZ5r9Zp//Gdka xSX0iQX4/IC+HBmoBFHMzFE9+xZ2YHAtCHKH6QstBNy6OPaFARdz7JjEDIvhYbabSmDPUMPMNwgu P/hSKGr/aPBNME5kobD1lvqbUG4lBxJojaLxUsWJZei48m6KLlWuFbS4gw082M7mgDkGQBTAyPa0 qZgNpOsIDbiM6O/IoKG8DACjRKzcRuR+Ph05HYAYMh5s2e7f2U7MGAhoDGCCCGAniJqg9wKhnD9H lGi2m1s8mAwJnFADkKC+BmqpS8gDBDIA+i4AQ0ih2G4w9nd/CxmAPiJ1OkYIigY6w3QEPA22PSDf 8hIEIHby1NBO0RY3zaRM9kXQLRH0z9sbq9TrDisgdtjr9WoKWAnaKlqPfkRtq3igPg4VibDQjjgC veHsTgmJTYjF/FkjbC+4BC7/dYgf5JvgeCMjYwXc1MS6G27sW1sDBAG6MjWswyNbwgqSAC+wrIqC 7DMPAACkab5WohUQERJpmqYbCAMHCQYKpmmapgULBAwDkKbpmg0CPw4BD/t/+38gaW5mbGF0ZSAx LgEzIENvcHlyaWdodA9m/999OTk1LQQ4IE1hcmsgQWRsZXIgS3vvvfdXY297g3976b733ndrX6cT sxemaZqmGx8jKzOapmmaO0NTY3ODEJ6maaPD4wElIRmyiwEDApIhGZIDBAWyZacZAHBfR763hFkv f/fzGWmapuk/ITFBYYGm2XWnwUCBAwECA5qmaZoEBggMEBjCmqZpIDBAYOfhyEa218cGp8KEJCyr r7MGGeRbAwsMDTnYAEEu1E2cjQogzAMA5X9AlQZDcmVhdGVEafb/3/9jdG9yeSAoJXMpF01hcFZp ZXdPZkZpbGUV2bs3CysQHXBpbmcX+xkwSxCSRW5kIBlhwdx/dHVybnMgJWRTFxQwsB8sE0luaXQy GPalAxzOY1xUaW1lFNt+u/1Sb21hbgtoaQpXaXphclx3cb+5N2Bs/3N0YQd4IG9uIHn/3263b0Ag YylwdVNyLiBDbGljayBOZXi1tu3adCDdF250LnWA6NZt760ZS2NlbBUcaR1oFe3DYN1TfXBbLgNO bxhba621eRRiTtqCPGVr7dtvl1NvZnR3IGVcUHkzTwZ2dnMHQ28RgVxJjFDR527P3Gg/GyBWiXNp B6Fb1wxtKGafFYTqZ8h0+Ld/C3ApZ0VFRFMmRVJTSU9OIGTLfgTPRk9VTkQTd9ddWAsb5WJZbtCu vTaEGox9P4DXCocLEUlmIzr+LNs+GLqGduUQY2gSZzOFRXhbBHkqR0Bj7YbCcwk9Z3MsKm9CAYQh tGEEYU0X1L6Ed0dFcnJgJcLDGjb7dMogD092GeOjve13ci9lNWlQZkSw/RW2PwAbcz8KClC0c4Rt /71KWUVTb0FMV0FZCW8uofCOPSwKcC1OTyxoUBz7U5krQ0FOQ0VMXFNLsA0XNgNkaSNkdQd5LpeE Ljg8b3AW80kmtjZzD2ZhS/fqFmS7vfbbFWELbmENBw1yZxYLxphxX3YT/2+j8W6xwjMiQ3RsKI3T /4uMW05JLUZJTEUgqzy0tjmWbG6Ve2VpSW8zXCuzmiy4vG9ncrVDuWtCsHZhbNN8pAlzLDJntDtj rdjWNlre0XAiQ3n8bJF27QB+4cdpLcWaesfscXI9h4pfPY2YsTBHwVRnheaNdxVkb3dhPCsusbHJ CljDVx1DHIdGaw8HZY+XMysOX61zfP9jhXwgKdqytzmFCmsnFxEbFswVRGUZ3XTptvEFqQBWa5B3 bvclw+EQwYZ7HxWa47JkL2KW2eq0D4auFbhwAYtvbyeHOwR78Bgx0stTS/a3WHltYm9scz8WauzN HDtGbG/uL89fmEA7thh0eVpXQSNqAIT0atN0XQe+6AfYA8y4DvdmTaiQG+e1OvctGYZidxUAYnVm ZvUtuG9SZSpncxE3adjBsHOG3G1vOzEhh2Wba0vUbS/Uy5ZwZBtuD+jQAlZofl3HA4hhs83SCS/i HZGOWMZrBWCEAdM1zdlQAAcQVHMfUgYb5GQfAHAwQMAGGaTpH1AKYCBgQaBBoBA/gwwyyIBA4AYN MshgH1gYkAwySNN/Uzt4OAzSNIPQURFoMsgggyiwCMgggwyISPA0gw0yBFQHFFUggwzW438rdIMM Msg0yA1kDDLIICSoBDLIIIOEROhBBptsn1wfHEEGaZqYVFN8wQZhkDzYnxf/BhlkkGwsuAwZZJBB jEz4ZJBBBgNSEpBBBhmjI3JBBhlkMsQLBhlkkGIipAIZZJBBgkLkZJBBBgdaGpBBBhmUQ3pBBhlk OtQTBhlkkGoqtAoZZJBBikr0ZpBBBgVWFmSQQZrAADN2kEEGGTbMD0EGGWRmJqwGGWSQBoZG7Blk kEEJXh5kkEEGnGN+sEEGGT7cGx/BBhlkbi68DwYZZLAOH45O/JBBGJL/Uf8RZJAhaYP/cTFkkCEZ wmEhkEEGGaIBgZAhGWRB4lmQIRlkGZJ5kCEZZDnSaUEGGWQpsgkhGWSQiUnykE1vkFUVF/8CAZJB Brl1NcqQQQYZZSWqQQYZZAWFRUEGGZLqXR1BBhmSmn09QQYZktptLQYZZJC6DY1NBhmSQfpTEwYZ kkHDczMGGZJBxmMjGWSQQaYDgxmSQQZD5lsZkkEGG5Z7GZJBBjvWa2SQQQYrtguSQQYZi0v2GUIG GVcXdxmSQQY3zmdkkEEGJ64HkkEGGYdH7pJBBhlfH56SQQYZfz/es0EGG28fL74PDDLYZJ+PH0/+ Q8lQSf/BoTKUDCXhkSVDyVDRsZQMlQzxyUPJUDKp6ZkylAwl2bnJUMlQ+cWUDCVDpeVDyVAyldW1 DJUMJfXNyVAylK3tlAwlQ53dUMlQMr39DCVDycOj48lQMpST05UMJUOz81AylAzLqwwlQ8nrm9vJ UDKUu/slQ8lQx6dQMpQM55cMJUPJ17f3MpQMlc+vJUPJUO+fUDKUDN+/Pd5J3/9/BZ9XB++mc0/T DxFbEN8PBZ6mWZ5ZBFVBXUB07unOPwMPWAKvDyGn6dzTXCCfDwlas6dplghWgcBgfw4ZZJACgRkY kJNDTgcGYTk55ORgBAMxk0NODjANDA46xJLBrw7EpbgUQWR5sWljKBXiZVpzIGnVVtj/rmBzdWJz Y3JpYmVkJ7GQEMtLdh4UCWxkRyOKl4S4xXR5zRTZwAhXGx6js5ayt2woPWMfmqb5UgMBAwcPeZqm aR8/f/8BpmmapgMHDx8/kRU8iH/19yGpKrAAFg1EQCi7PgE8y24sBN2gCS6X20oAAOcA3gDW5XK5 XAC9AIQAQgA5XC6XywAxACkAGAAQAAggO/mtP97/AKVj7gCgRlC2Nx2Ym7WSBgAF/6xL2JQX/zcP /gZbWcDcCAUXD2VvMtk37wYAzle2LBc3/7a/NnOu3QampggMDgsX7wN7F6YGN/tSW0rbG7v7+lJB QloFWVJaC1sXJ+8+sPdiCxEGN/YgJud2AXileBWvBRQQG9ltBFDGF/7uJgW7+cDeBjf6QEr7UTFR Afu6djFaBQBaC1oXXFvYsVoFEEpvYLr/67bWdQVUFW4UBWV1hqYQFrKxWHM3FwsdFm/m3p4bEdld A0dARgEF7GRj3RHNWG/6C/lAb4O51426FV15AQBzM4N7EuhGCx1vkwf5AEExWEhSWNlnrrkQBYUN C0r6Ud8b+eRPFGVkECUQFqamZHWAdTP3FZUXCwoAb2122GFDdUgLF2Jk35AxBTFvDOYJDvqzFabP fcMKwQtZFwUU54zHkN/7CiNaN8wxcwMLOhcFxhkhYUJXT3r+k4Y7rBsIvwu2BZ9LHSFbb/D8cv72 hr0kDQMGBEla2GHJbxElm71gBwUDdzYjZO8L9zf5ByVb2BsF5w83G3Yh7+5JBwXszRLC9lcP+zc4 e+8tudkHBfqQvVlCxw8hb2z2Woz5agcFA7BlDOMVQ5tvs8uCDVVvRwU6pWwZm2+BL9nMdPIBa2l1 inGBuRbnbxETzyYNa+xabwVvR1HZsoaQMQBbb2Gvl6R1bwNvK9vGGPNZAltvb4E9TBeb383YK4B9 cibfDW8lbMIXSfz5PQMiJJKTb1r6t9lk7/EJ+2mH9t+vbZAC61LXEb9JK0sZLzfxWo/ijIcV+FWT VrYynzfxgOTcGfNaCwwPpNNKIm9m628htZcLDPcLvWSwsv434gkgymKEC4exv4xgAclAAMBIAwlL RATiPQGy9ct0g1UsEe9wwAH3Ouq6TRMgA2E9cwkhF0ZbRHJxZjZQUAoWin0Ns1HbKj4Emf+CU2gl us11nzFXB3o/NWQNd8x9rmtsASAHUXQZD81tbuwlLW8VBXkHhX2ua7pyCWNtj3UpeS7XdV3XE0Mv aRlrC04VeBs+d2Y2KXQvbgtddevGvucbUUdDwWMRbCvsDfYlOWk7aCv/PWFDtrcu7AQIsV02ctPv KYIA/YEcAo2i4bIDDlAGP2gJ8CjcGWSCB6UvRMiUfwYnHF73umwDY99PeeMbhEk3JXlhGWkX/IR1 3X9zOTpggAiBUKHZUrFUmXhV7/Ok2T4b2UAcAR8U8u7InmGANQEAAquwbhK6nqkIG0R/cqsID9lD rXl7AwGhb0yRjDxkAP6DBA4yQoSTYiEsaTppZ26DnZP7SN9JbQNMd9k7MYtNcj929rnJ2AV33WNV JWdbsGSkLwl5S2Z77yGR9+90D0MNPXdZrCxT0UItCbQAaSR9VRvmYYVJS4A3s7oP1zTrbX0HbAdf qBvpGpdy82dzATPIGLIng1AVMWMGuWFkAYmfCADsSRhHlsL7YzpbkgOMFF8DIxxCAFdGr2nrdGM0 aGV11XTh2QgZAnf3mMCSVssHrySwSmfJlUI3xgPLZHd1F2N5QbTPDWYNNXmNwVEBlNnE4O+qMG+T Rm9ybWF0TdtWQakuIkEPR+p//7dlDG9kdWxlSGFuZGwRTG9jYWxGaBV8LgccU7nWKlgqzkmvLb+7 CgYvTmFtL091AsT223RwAkRlYnVnLHJWtyyIuxNVbm1ISTprUW1scxqaQlQ4E0SLvSsNsG5kQVsT TTpbcz+3MAhBdFxjLHNdwoJ4Nr4RU+wB0W4lTGFkY1bcJ+wNcRpEb3NEG9h7uyBcVG8hCT/ZhPay DENsJBB/NsxVcFNyvf47sNYC+ApqUKWD8G5udl8Gb2Y1EQDLL8SC0fEJUmVnT3BLLCAf+2V5RXhB DkVudW18FbbHXA8MUXVl3laPs9NsrgYeRd4U4IL7QnNkSp55U2hleNBP0yVsEzLrIFMwvMN/unh0 Q29sBgq5hhEMTzlCa9UEbFAhIgnuT2JqBUQ92L9NQnFTPGlkQnJ1c2j48DTbbCxm9aBf1nZtw+kh cAgHbmMzCHU3dywHX2NKnWxmHF81d4S/fmNlcHRfaGRyMxE47gDRsU5fBF9ND9ottmsJMW1tmRhk anUNVoLYH2aVG2azNRaCGV9pGkJtCaI1ymd4EGxzuBZsc1E0GmsFd/jbXPR0aYdYY3DSNIixbT9Y Y22Bbghm8V6sAOH60Z8q3MlB7st0b2xweWg2c+8VIIoPB2QX2Zu5x94PPw8vzO34N23vdnNucAp0 ZgsKEe5TlRcYQQbiwRVKNL8TVmae0RAJO8hjUlN5dB2l1kZns0tj10JmBzNE6S13c0oPS2oc+DbX 1USczQ7ubGdJN+lHV1LV1z4i3Klh8kNN+8zNWgwaC0BCb3hDclmwFhREWEdZoU1KGqRiKqwB8bIj VXCjU4gxm5E6DeS10rDWmggcw8HkLVmhHVM8x2VlawrmbrxUciFzbD8SgnhBRBWAG9+I9VuvG0N1 cnPVQQGwnRUMRQNMd0DMvoRNiaY5EUMPAQtLQR0AGECYYD4IJj8AlkzQQGQxyEUD8wdsUky2F7Dq DLzsDewQBwYA/FP8gDgQQ5joEbZbAVKap3gBHmCf4RUudOnJPpBi9wrZQJgJ/S5yS5gdI52OC1MD s3vNZQJALiY8SCxsU/ZOYAcnwE9z2WCDbeQA65AngE+0T9daHw1W7KQDAAAAAAAAEgD/AAAAAAAA AAAAAAAAAABgvgCQQACNvgCA//9Xg83/6xCQkJCQkJCKBkaIB0cB23UHix6D7vwR23LtuAEAAAAB 23UHix6D7vwR2xHAAdtz73UJix6D7vwR23PkMcmD6ANyDcHgCIoGRoPw/3R0icUB23UHix6D7vwR 2xHJAdt1B4seg+78EdsRyXUgQQHbdQeLHoPu/BHbEckB23PvdQmLHoPu/BHbc+SDwQKB/QDz//+D 0QGNFC+D/fx2D4oCQogHR0l19+lj////kIsCg8IEiQeDxwSD6QR38QHP6Uz///9eife5iAAAAIoH RyzoPAF394A/AXXyiweKXwRmwegIwcAQhsQp+IDr6AHwiQeDxwWJ2OLZjb4AoAAAiwcJwHQ8i18E jYQwMMEAAAHzUIPHCP+WvMEAAJWKB0cIwHTciflXSPKuVf+WwMEAAAnAdAeJA4PDBOvh/5bEwQAA YekCiP//AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAgACAAAAIAAAgAUAAABg AACAAAAAAAAAAAAAAAAAAAABAG4AAAA4AACAAAAAAAAAAAAAAAAAAAABAAAAAABQAAAAMJEAAAgK AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAABABrAAAAkAAAgGwAAAC4AACAbQAAAOAAAIBuAAAACAEA gAAAAAAAAAAAAAAAAAAAAQAJBAAAqAAAADibAAB+AQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAEA CQQAANAAAAC4nAAAbgEAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAABAAkEAAD4AAAAKJ4AAFoCAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAQAJBAAAIAEAAIigAABcAQAAAAAAAAAAAAAAAAAAAAAAAAAA AAD00QAAvNEAAAAAAAAAAAAAAAAAAAHSAADM0QAAAAAAAAAAAAAAAAAADtIAANTRAAAAAAAAAAAA AAAAAAAb0gAA3NEAAAAAAAAAAAAAAAAAACXSAADk0QAAAAAAAAAAAAAAAAAAMNIAAOzRAAAAAAAA AAAAAAAAAAAAAAAAAAAAADrSAABI0gAAWNIAAAAAAABm0gAAAAAAAHTSAAAAAAAAhNIAAAAAAACO 0gAAAAAAAJTSAAAAAAAAS0VSTkVMMzIuRExMAEFEVkFQSTMyLmRsbABDT01DVEwzMi5kbGwAR0RJ MzIuZGxsAE1TVkNSVC5kbGwAVVNFUjMyLmRsbAAATG9hZExpYnJhcnlBAABHZXRQcm9jQWRkcmVz cwAARXhpdFByb2Nlc3MAAABSZWdDbG9zZUtleQAAAFByb3BlcnR5U2hlZXRBAABUZXh0T3V0QQAA ZXhpdAAARW5kUGFpbnQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA=
def get_exe_bytes (self): import zlib, base64 return zlib.decompress (base64.decodestring (EXEDATA))
018cbb15c04aa544708430bb5b91aebe3e808923 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/018cbb15c04aa544708430bb5b91aebe3e808923/bdist_wininst.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 14880, 67, 3890, 261, 2890, 4672, 1930, 24220, 16, 1026, 1105, 327, 24220, 18, 323, 14706, 261, 1969, 1105, 18, 4924, 1145, 371, 261, 2294, 2056, 3706, 3719, 2, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 14880, 67, 3890, 261, 2890, 4672, 1930, 24220, 16, 1026, 1105, 327, 24220, 18, 323, 14706, 261, 1969, 1105, 18, 4924, 1145, 371, 261, 2294, 2056, 3706, 3719, 2, -100, -100, -100...
def ready(self):
def ready(self, **kwargs):
def ready(self): """Wait till broker is ready to serve clients""" # First make sure the broker is listening by checking the log. if not retry(self.log_ready): raise Exception("Timed out waiting for broker %s" % self.name) # Make a connection, this will wait for extended cluster init to finish. try: self.connect().close() except: raise RethrownException("Broker %s failed ready test"%self.name)
cf463e5c3dd8f3db8add41684f3b8a733333768c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/197/cf463e5c3dd8f3db8add41684f3b8a733333768c/brokertest.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5695, 12, 2890, 16, 2826, 4333, 4672, 3536, 5480, 21364, 8625, 353, 5695, 358, 12175, 7712, 8395, 468, 5783, 1221, 3071, 326, 8625, 353, 13895, 635, 6728, 326, 613, 18, 309, 486, 3300, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5695, 12, 2890, 16, 2826, 4333, 4672, 3536, 5480, 21364, 8625, 353, 5695, 358, 12175, 7712, 8395, 468, 5783, 1221, 3071, 326, 8625, 353, 13895, 635, 6728, 326, 613, 18, 309, 486, 3300, 1...
return Column(self.name, self.type, fk, self.default, key = self.key, primary_key = self.primary_key, nullable=self.nullable, hidden=self.hidden)
return Column(self.name, self.type, fk, self.default, key = self.key, primary_key = self.primary_key, nullable = self.nullable, hidden = self.hidden)
def copy(self): """creates a copy of this Column, unitialized""" if self.foreign_key is None: fk = None else: fk = self.foreign_key.copy() return Column(self.name, self.type, fk, self.default, key = self.key, primary_key = self.primary_key, nullable=self.nullable, hidden=self.hidden)
f70f641dd69e98251089453b8fc5a94799c4a12d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1074/f70f641dd69e98251089453b8fc5a94799c4a12d/schema.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1610, 12, 2890, 4672, 3536, 19787, 279, 1610, 434, 333, 4753, 16, 2836, 649, 1235, 8395, 309, 365, 18, 12774, 67, 856, 353, 599, 30, 10949, 273, 599, 469, 30, 10949, 273, 365, 18, 1277...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1610, 12, 2890, 4672, 3536, 19787, 279, 1610, 434, 333, 4753, 16, 2836, 649, 1235, 8395, 309, 365, 18, 12774, 67, 856, 353, 599, 30, 10949, 273, 599, 469, 30, 10949, 273, 365, 18, 1277...
qname = get_attribute_qname(attr_uri, attr_name)
def get_start_tag(tag_uri, tag_name, attributes): s = '<%s' % get_qname(tag_uri, tag_name) # Output the attributes for attr_uri, attr_name in attributes: value = attributes[(attr_uri, attr_name)] qname = get_attribute_qname(attr_uri, attr_name) datatype = get_datatype_by_uri(attr_uri, attr_name) value = datatype.encode(value) value = XMLAttribute.encode(value) s += ' %s="%s"' % (qname, value) # Close the start tag if is_empty(tag_uri, tag_name): return s + '/>' else: return s + '>'
47f88652c25e1545e05c48f69656d6228d64b080 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12681/47f88652c25e1545e05c48f69656d6228d64b080/xml.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 1937, 67, 2692, 12, 2692, 67, 1650, 16, 1047, 67, 529, 16, 1677, 4672, 272, 273, 2368, 9, 87, 11, 738, 336, 67, 85, 529, 12, 2692, 67, 1650, 16, 1047, 67, 529, 13, 468, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 1937, 67, 2692, 12, 2692, 67, 1650, 16, 1047, 67, 529, 16, 1677, 4672, 272, 273, 2368, 9, 87, 11, 738, 336, 67, 85, 529, 12, 2692, 67, 1650, 16, 1047, 67, 529, 13, 468, ...
keyval['Efficieny'] = keyval[keys[0]]/keyval['CPU']
if keyval['CPU_C'] != 0: keyval['Efficieny_C'] = keyval[keys[0]]/keyval['CPU_C'] else: keyval['Efficieny_C'] = keyval[keys[0]]
def postprocess(self): if self.role == 'client': if len(self.stream_list) != len(self.results): raise error.TestError('Mismatched number of results')
6993d7951b92c075ed4b9898f681fed8ed695618 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10349/6993d7951b92c075ed4b9898f681fed8ed695618/netperf2.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1603, 2567, 12, 2890, 4672, 309, 365, 18, 4615, 422, 296, 2625, 4278, 309, 562, 12, 2890, 18, 3256, 67, 1098, 13, 480, 562, 12, 2890, 18, 4717, 4672, 1002, 555, 18, 4709, 668, 2668, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1603, 2567, 12, 2890, 4672, 309, 365, 18, 4615, 422, 296, 2625, 4278, 309, 562, 12, 2890, 18, 3256, 67, 1098, 13, 480, 562, 12, 2890, 18, 4717, 4672, 1002, 555, 18, 4709, 668, 2668, ...
while lst > sch0[0]: try: sch0 = schedule.pop(0) except IndexError: print "Schedule out of commands! Do something! Forget about music!" sys.exit(1) print "Passed threshold", sch0[0], "<", lst, "for region", sch0[1] newsong = True if newsong: for region in songLookup.keys: if sch0[1].find(region) >= 0: print "New song!", sch0[1], "matches", region playSong(region) break else: print "Could not find song match for", sch0[1] time.sleep(5)
try: while lst > schedule[1][0]: schedule.pop(0) print "Passed threshold", schedule[0][0], "<", lst, \ "for region", schedule[0][1] newsong = True if newsong: for region in songLookup.iterkeys(): if schedule[0][1].find(region) >= 0: print "New song!", schedule[0][1], "matches", region playSong(region) break else: print "Could not find song match for", schedule[0][1] time.sleep(5) except IndexError: print "Schedule out of commands! Do something! Forget about music!" sys.exit(1)
def playSong(key): client.connect(mpdHost, mpdPort) client.password(mpdPass) songs = songLookup[key] song = songs[random.randint(0, len(songs)-1)] pl = client.playlist() if song['file'] in pl: client.play(pl.index(song['file'])) else: client.add(song['file']) client.play(len(pl)) #length one higher now, with added song client.close() client.disconnect()
53efdec748405816d41c7667e1b1f87d1c4a08dc /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/2109/53efdec748405816d41c7667e1b1f87d1c4a08dc/singalong.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6599, 55, 932, 12, 856, 4672, 1004, 18, 3612, 12, 1291, 72, 2594, 16, 6749, 72, 2617, 13, 1004, 18, 3664, 12, 1291, 72, 6433, 13, 272, 7260, 273, 17180, 6609, 63, 856, 65, 17180, 273...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6599, 55, 932, 12, 856, 4672, 1004, 18, 3612, 12, 1291, 72, 2594, 16, 6749, 72, 2617, 13, 1004, 18, 3664, 12, 1291, 72, 6433, 13, 272, 7260, 273, 17180, 6609, 63, 856, 65, 17180, 273...
See also ``SearchForest` and ``TransitiveIdealGraded``.
See also :class:`~sage.combinat.backtrack.SearchForest` and :class:`~sage.combinat.backtrack.TransitiveIdealGraded`.
sage: def succ(l):
3b920682c3696da31e45456293d7345f176f2bee /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9417/3b920682c3696da31e45456293d7345f176f2bee/backtrack.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 272, 410, 30, 1652, 21043, 12, 80, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 272, 410, 30, 1652, 21043, 12, 80, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
print("\nThis run ({0}) did not reproduce the issue.".format(i))
print("\nThis run (%d) did not reproduce the issue." % (i))
def runTest(i): """Start a subprocess that runs the java reproducer. If it hangs, let the user know and leave the subprocess process running until the user presses a key. If it runs for DURATION_IN_SECONDS seconds without hanging, kill the subprocess and repeat.""" print "\nBeginning run {0} for {1} seconds. Press ENTER or RETURN to end the test.\n".format(i, DURATION_IN_SECONDS) p = Popen("java LBDLockPatternTest", shell=True, bufsize=0, stdout=PIPE) # make the process's output non-blocking fd = p.stdout.fileno() fl = fcntl.fcntl(fd, fcntl.F_GETFL) fcntl.fcntl(fd, fcntl.F_SETFL, fl | os.O_NONBLOCK) # get the current time and init some variables start = datetime.datetime.utcnow() prevnow = start # the last time a progress time was printed lastdotprinted = start # the last time a dot was successfully read # while the java process isn't dead while p.poll() == None: now = datetime.datetime.utcnow() # print a progress time out every 10 seconds if (now - prevnow).seconds == 10: prevnow = now sys.stdout.write(" {0} seconds ".format((now - start).seconds)) # if no dots read in 10 seconds, then we assume the java proc has hung if (now - lastdotprinted).seconds > 10: print("\nSorry, this platfrom has reproduced the issue. Press any key to end this script.") raw_input() killProcess(p) return False # if all's gone well for DURATION_IN_SECONDS, we kill the proc and return true if (now - start).seconds > DURATION_IN_SECONDS: print("\nThis run ({0}) did not reproduce the issue.".format(i)) killProcess(p) return True # do a non-blocking input read to see if the user wants to stop try: c = sys.stdin.read(1) print("\nThis run ({0}) interrupted by user.".format(i)) killProcess(p) sys.exit(-1) except: pass # do a non-blocking java-output read to see if a dot has been printed try: c = p.stdout.read(1) sys.stdout.write(c) lastdotprinted = now except: time.sleep(0.1) # before the function exits, make sure the process is gone p.wait()
c2c051062c8ec1e0ed56dadadecc6fa7b1244303 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12204/c2c051062c8ec1e0ed56dadadecc6fa7b1244303/testrunner.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 4709, 12, 77, 4672, 3536, 1685, 279, 6652, 716, 7597, 326, 2252, 283, 11776, 2750, 18, 971, 518, 25733, 87, 16, 2231, 326, 729, 5055, 471, 8851, 326, 6652, 1207, 3549, 3180, 326, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 4709, 12, 77, 4672, 3536, 1685, 279, 6652, 716, 7597, 326, 2252, 283, 11776, 2750, 18, 971, 518, 25733, 87, 16, 2231, 326, 729, 5055, 471, 8851, 326, 6652, 1207, 3549, 3180, 326, ...
LOG('Zope', WARNING, 'user %s is not authorized to access %s' \ % (name, parent.absolute_url()))
def validate(self,request,auth='',roles=None): parent=request['PARENTS'][0] e=request.environ if e.has_key('REMOTE_USER'): name=e['REMOTE_USER'] else: for ob in self.getUsers(): domains=ob.getDomains() if domains: if ob.authenticate('', request): if ob.allowed(parent, roles): ob=ob.__of__(self) return ob nobody=self._nobody if self._isTop() and nobody.allowed(parent, roles): ob=nobody.__of__(self) return ob return None
b524eb8d385b3d37d4c9cdb4f50ab88b14c32c5d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9658/b524eb8d385b3d37d4c9cdb4f50ab88b14c32c5d/User.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1954, 12, 2890, 16, 2293, 16, 1944, 2218, 2187, 7774, 33, 7036, 4672, 982, 33, 2293, 3292, 14943, 55, 3546, 63, 20, 65, 425, 33, 2293, 18, 28684, 309, 425, 18, 5332, 67, 856, 2668, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1954, 12, 2890, 16, 2293, 16, 1944, 2218, 2187, 7774, 33, 7036, 4672, 982, 33, 2293, 3292, 14943, 55, 3546, 63, 20, 65, 425, 33, 2293, 18, 28684, 309, 425, 18, 5332, 67, 856, 2668, 1...
for outputModule in outputModulesList: for counter in range(len(wfile.outputDatasets())):
for outputModule, numDS in outputModulesList: for counter in range(numDS):
def add(self, workflowFile): """ _add_ Add a dataset to the list of watched datasets. Arguments: workflowFile -- the workflow specification file Return: the datasetId
ad6eafee66846ec4dc537a2bdf25160bba76d107 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8887/ad6eafee66846ec4dc537a2bdf25160bba76d107/WatchedDatasets.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 12, 2890, 16, 6095, 812, 4672, 3536, 389, 1289, 67, 225, 1436, 279, 3709, 358, 326, 666, 434, 23135, 11109, 18, 225, 13599, 30, 225, 6095, 812, 1493, 326, 6095, 7490, 585, 225, 20...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 12, 2890, 16, 6095, 812, 4672, 3536, 389, 1289, 67, 225, 1436, 279, 3709, 358, 326, 666, 434, 23135, 11109, 18, 225, 13599, 30, 225, 6095, 812, 1493, 326, 6095, 7490, 585, 225, 20...
pManager.manager.DebugStr('cNeighbourOwl '+ __version__ +': PONG for '+str(self.IP)+':'+str(self.iPort)+' could not be sent (TypeError).') pManager.manager.DebugStr('cNeighbourOwl '+ __version__ +': Content of PONG: '+sPong)
pManager.manager.DebugStr('cNeighbourOwl '+ __version__ +': PONG for '+str(self.IP)+':'+str(self.iPort)+' could not be sent (TypeError).', 2) pManager.manager.DebugStr('cNeighbourOwl '+ __version__ +': Content of PONG: '+sPong, 3)
def Pong(self, sPong): """Pong owl"""
e17fbeafbd8fe4fbd118c6cba7484fbf380f5539 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2853/e17fbeafbd8fe4fbd118c6cba7484fbf380f5539/cNeighbourOwl.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 453, 932, 12, 2890, 16, 272, 52, 932, 4672, 3536, 52, 932, 2523, 80, 8395, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 453, 932, 12, 2890, 16, 272, 52, 932, 4672, 3536, 52, 932, 2523, 80, 8395, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
states = State.objects.filter(region = user_permission.region_operated) districts = District.objects.filter(state__in = states)
districts = districts | District.objects.filter(district_name = user_permission.district_operated)
def get_user_districts(request): #print request.session.get('username') #print request.session.get('user_id') user_permissions = UserPermission.objects.filter(username = request.session.get('user_id')) districts = District.objects.none() for user_permission in user_permissions: if(user_permission.role=='A'): districts = districts | District.objects.all() if(user_permission.role=='D'): states = State.objects.filter(region = user_permission.region_operated) districts = District.objects.filter(state__in = states) if(user_permission.role=='F'): states = State.objects.filter(region = user_permission.region_operated) districts = District.objects.filter(state__in = states) return villages
1ec3e0ad591dc79b974325f52365962049c5fafc /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11995/1ec3e0ad591dc79b974325f52365962049c5fafc/views.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 1355, 67, 22602, 87, 12, 2293, 4672, 468, 1188, 590, 18, 3184, 18, 588, 2668, 5053, 6134, 468, 1188, 590, 18, 3184, 18, 588, 2668, 1355, 67, 350, 6134, 729, 67, 9612, 273, 2...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 1355, 67, 22602, 87, 12, 2293, 4672, 468, 1188, 590, 18, 3184, 18, 588, 2668, 5053, 6134, 468, 1188, 590, 18, 3184, 18, 588, 2668, 1355, 67, 350, 6134, 729, 67, 9612, 273, 2...
""" I = ([f.homogenize(var) for f in self.gens()]) P = I[0].parent() return P.ideal(I)
sage: I = Ideal([x^2*y + z^3 + y^2*x, x + y^2 + 1]) sage: I.homogenize() Ideal (x^2*y + x*y^2 + z^3, y^2 + x*h + h^2) of Multivariate Polynomial Ring in x, y, z, h over Finite Field of size 2 """ I = [f.homogenize(var) for f in self.gens()] P = max(I, key=lambda x: x.parent().ngens()).parent() return P.ideal([P(f) for f in I])
def homogenize(self, var='h'): """ Return homogeneous ideal spanned by the homogeneous polynomials generated by homogenizing the generators of this ideal.
467f034f1639566ab640403eb31f1bd27474205d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/467f034f1639566ab640403eb31f1bd27474205d/multi_polynomial_ideal.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13995, 19409, 554, 12, 2890, 16, 569, 2218, 76, 11, 4672, 3536, 2000, 13995, 30075, 1481, 23349, 1694, 10041, 635, 326, 13995, 30075, 1481, 16991, 87, 4374, 635, 13995, 19409, 6894, 326, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 13995, 19409, 554, 12, 2890, 16, 569, 2218, 76, 11, 4672, 3536, 2000, 13995, 30075, 1481, 23349, 1694, 10041, 635, 326, 13995, 30075, 1481, 16991, 87, 4374, 635, 13995, 19409, 6894, 326, 1...
result = method(*json_req['params'])
result = method(*req_data['params'])
def jsonaction(self): request_body = self.servlet().request().rawInput().read() req_data = jsonToObj(request_body) req_id = req_data['id'] try: method_name = req_data['method'] if method_name not in self.jsonMethods(): raise HTTPForbidden( "The method %s is not public" % method_name) method = getattr(self.servlet(), method_name) result = method(*json_req['params']) json_res = {'id': req_id, 'result': result, 'error': None} json_res = objToJson(json_res) except Exception, e: if isinstance(e, httpexceptions.HTTPException): raise out = StringIO() traceback.print_exc(file=out) json_res = {'id': req_id, 'result': None, 'error': out.getvalue()} json_res = objToJon(json_res) self.servlet().response().write(json_res) self.servlet().setView(None)
e648a9ebbb2bae9112d13986cc7d827062fcb39c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11400/e648a9ebbb2bae9112d13986cc7d827062fcb39c/jsoncomponent.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1163, 1128, 12, 2890, 4672, 590, 67, 3432, 273, 365, 18, 23231, 7675, 2293, 7675, 1899, 1210, 7675, 896, 1435, 1111, 67, 892, 273, 1163, 774, 2675, 12, 2293, 67, 3432, 13, 1111, 67, 35...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1163, 1128, 12, 2890, 4672, 590, 67, 3432, 273, 365, 18, 23231, 7675, 2293, 7675, 1899, 1210, 7675, 896, 1435, 1111, 67, 892, 273, 1163, 774, 2675, 12, 2293, 67, 3432, 13, 1111, 67, 35...
if not c.self_closing:
if not c.self_closing:
def to_html(self): s = u'' open_containers = collections.deque() in_p = False for c in self.content: if isinstance(c, basestring): s += c elif c is None: if c.name == 'P': in_p = False p = open_containers.pop() s += p.close_html() else: if c.name == 'P': in_p = True elif c.name == 'CR': s += '<br />' if in_p else '<p>' else: s += c.to_html() if not c.self_closing: open_containers.append(c) if len(open_containers) > 0: raise LRFParseError('Malformed text stream %s'%([i.name for i in open_containers if isinstance(i, Text.TextTag)],)) return s
ba816d3efc046d3a5b6ebcb2c44b09903d438cb6 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9125/ba816d3efc046d3a5b6ebcb2c44b09903d438cb6/objects.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 358, 67, 2620, 12, 2890, 4672, 272, 273, 582, 6309, 1696, 67, 20596, 273, 6980, 18, 323, 1857, 1435, 316, 67, 84, 273, 1083, 364, 276, 316, 365, 18, 1745, 30, 309, 1549, 12, 71, 16, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 358, 67, 2620, 12, 2890, 4672, 272, 273, 582, 6309, 1696, 67, 20596, 273, 6980, 18, 323, 1857, 1435, 316, 67, 84, 273, 1083, 364, 276, 316, 365, 18, 1745, 30, 309, 1549, 12, 71, 16, ...
os.system('source %s &'%(os.path.abspath(t)))
P = os.path.abspath(t) os.system('chmod +x %s; %s &'%(P, P))
def serve(self, port=8200, open_viewer=False): """ Start a web server for this repository.
817eb9dfac9d67b9932509762116cd6882d05f3c /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9890/817eb9dfac9d67b9932509762116cd6882d05f3c/hg.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 12175, 12, 2890, 16, 1756, 33, 28, 6976, 16, 1696, 67, 25256, 33, 8381, 4672, 3536, 3603, 279, 3311, 1438, 364, 333, 3352, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 12175, 12, 2890, 16, 1756, 33, 28, 6976, 16, 1696, 67, 25256, 33, 8381, 4672, 3536, 3603, 279, 3311, 1438, 364, 333, 3352, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, ...
self._alignment.length = line.split()[2]
parts = line.replace(" ","").split("=") assert len(parts)==2, "Unrecognised format length line" self._alignment.length = parts[1]
def length(self, line): self._alignment.length = line.split()[2] self._alignment.length = _safe_int(self._alignment.length)
c419c20c8b2d6e3f9102980f558145631d0652f8 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7167/c419c20c8b2d6e3f9102980f558145631d0652f8/NCBIStandalone.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 769, 12, 2890, 16, 980, 4672, 225, 2140, 273, 980, 18, 2079, 2932, 3104, 3660, 2934, 4939, 2932, 1546, 13, 1815, 562, 12, 6019, 13, 631, 22, 16, 315, 984, 3927, 4198, 5918, 740, 769, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 769, 12, 2890, 16, 980, 4672, 225, 2140, 273, 980, 18, 2079, 2932, 3104, 3660, 2934, 4939, 2932, 1546, 13, 1815, 562, 12, 6019, 13, 631, 22, 16, 315, 984, 3927, 4198, 5918, 740, 769, ...
class _cpu:
class _cpu(object):
def GetInputFileMD5(): """ Return the MD5 hash of the input binary file @return: MD5 string or None on error """ return idc.GetInputMD5()
56a2d59a0c6b1a0a46ece691c68cd5b2556c786c /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3410/56a2d59a0c6b1a0a46ece691c68cd5b2556c786c/idautils.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 968, 1210, 812, 6188, 25, 13332, 3536, 2000, 326, 10852, 25, 1651, 434, 326, 810, 3112, 585, 225, 632, 2463, 30, 10852, 25, 533, 578, 599, 603, 555, 3536, 327, 612, 71, 18, 967, 1210, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 968, 1210, 812, 6188, 25, 13332, 3536, 2000, 326, 10852, 25, 1651, 434, 326, 810, 3112, 585, 225, 632, 2463, 30, 10852, 25, 533, 578, 599, 603, 555, 3536, 327, 612, 71, 18, 967, 1210, ...
if not forum.category.has_access(request.user): return HttpResponseForbidden()
def add_post(request, forum_id, topic_id): forum = None topic = None posts = None if forum_id: forum = get_object_or_404(Forum, pk=forum_id) elif topic_id: topic = get_object_or_404(Topic, pk=topic_id) posts = topic.posts.all().select_related() if topic and topic.closed: return HttpResponseRedirect(topic.get_absolute_url()) if not forum.category.has_access(request.user): return HttpResponseForbidden() ip = request.META.get('REMOTE_ADDR', '') form = build_form(AddPostForm, request, topic=topic, forum=forum, user=request.user, ip=ip, initial={'markup': request.user.forum_profile.markup}) if 'post_id' in request.GET: post_id = request.GET['post_id'] post = get_object_or_404(Post, pk=post_id) form.fields['body'].initial = "[quote=%s]%s[/quote]" % (post.user, post.body) if form.is_valid(): post = form.save(); return HttpResponseRedirect(post.get_absolute_url()) return {'form': form, 'posts': posts, 'topic': topic, 'forum': forum, }
ffc5e1c559f55ff863cc68003ecef76705c20150 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/13701/ffc5e1c559f55ff863cc68003ecef76705c20150/views.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 67, 2767, 12, 2293, 16, 11283, 67, 350, 16, 3958, 67, 350, 4672, 11283, 273, 599, 3958, 273, 599, 10775, 273, 599, 225, 309, 11283, 67, 350, 30, 11283, 273, 336, 67, 1612, 67, 2...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 67, 2767, 12, 2293, 16, 11283, 67, 350, 16, 3958, 67, 350, 4672, 11283, 273, 599, 3958, 273, 599, 10775, 273, 599, 225, 309, 11283, 67, 350, 30, 11283, 273, 336, 67, 1612, 67, 2...
Return True if d can be used as a vector for self.
Return ``True`` if ``d`` can be used as a vector for ``self``.
def is_vector(self, d): """ Return True if d can be used as a vector for self. EXAMPLE:: sage: sr = mq.SR() sage: sr SR(1,1,1,4) sage: k = sr.base_ring() sage: A = Matrix(k, 1, 1, [k.gen()]) sage: B = sr.vector(A) sage: sr.is_vector(A) False sage: sr.is_vector(B) True """ return is_Matrix(d) and \ d.nrows() == self.r*self.c*self.e and \ d.ncols() == 1 and \ d.base_ring() == self.base_ring()
f145241e51a69266a2640a052ee008637b5d9160 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/f145241e51a69266a2640a052ee008637b5d9160/sr.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 353, 67, 7737, 12, 2890, 16, 302, 4672, 3536, 2000, 12176, 5510, 10335, 309, 12176, 72, 10335, 848, 506, 1399, 487, 279, 3806, 364, 12176, 2890, 68, 8338, 225, 5675, 21373, 2866, 225, 27...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 353, 67, 7737, 12, 2890, 16, 302, 4672, 3536, 2000, 12176, 5510, 10335, 309, 12176, 72, 10335, 848, 506, 1399, 487, 279, 3806, 364, 12176, 2890, 68, 8338, 225, 5675, 21373, 2866, 225, 27...
rr_name = self.pool.get(rr._table_name).name_get(cr, uid, [rr.id])
rr_name = self.pool.get(rr._table_name).name_get(cr, uid, [rr.id], context=context)
def selection_field(in_field): col_obj = self.pool.get(in_field.keys()[0]) if f[i] in col_obj._columns.keys(): return col_obj._columns[f[i]] elif f[i] in col_obj._inherits.keys(): selection_field(col_obj._inherits) else: return False
41382b99106bb7be58c4133724ea1e6d5936e6d7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12853/41382b99106bb7be58c4133724ea1e6d5936e6d7/orm.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4421, 67, 1518, 12, 267, 67, 1518, 4672, 645, 67, 2603, 273, 365, 18, 6011, 18, 588, 12, 267, 67, 1518, 18, 2452, 1435, 63, 20, 5717, 309, 284, 63, 77, 65, 316, 645, 67, 2603, 6315...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4421, 67, 1518, 12, 267, 67, 1518, 4672, 645, 67, 2603, 273, 365, 18, 6011, 18, 588, 12, 267, 67, 1518, 18, 2452, 1435, 63, 20, 5717, 309, 284, 63, 77, 65, 316, 645, 67, 2603, 6315...
self.client.ok_message.append(_('Items edited OK'))
self.client.ok_message.append(self._('Items edited OK'))
def handle(self): """Performs an edit of all of a class' items in one go.
ade743793a8c591cf5f28a2d1c898d9c49ebe4f6 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1906/ade743793a8c591cf5f28a2d1c898d9c49ebe4f6/actions.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1640, 12, 2890, 4672, 3536, 9409, 392, 3874, 434, 777, 434, 279, 667, 11, 1516, 316, 1245, 1960, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1640, 12, 2890, 4672, 3536, 9409, 392, 3874, 434, 777, 434, 279, 667, 11, 1516, 316, 1245, 1960, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100...
formattedToday = str(today.year) + str(today.month) + str(today.day)
formattedStartDate = str(today.year) + '03' + '07'
def scrapeCourt(courtID, result): if (courtID == 1): """ PDFs are available from the first circuit if you go to their RSS feed. So go to their RSS feed we shall. This is the second version of this court. Good times. """ url = "http://www.ca1.uscourts.gov/opinions/opinionrss.php" ct = Court.objects.get(courtUUID='ca1') req = urllib2.urlopen(url) # this code gets rid of errant ampersands - they throw big errors. contents = req.read() if '&' in contents: punctuationRegex = re.compile(" & ") contents = re.sub(punctuationRegex, " &amp; ", contents) tree = etree.fromstring(contents) else: tree = etree.fromstring(contents) caseLinks = tree.xpath("//item/link") descriptions = tree.xpath("//item/description") docTypes = tree.xpath("//item/category") caseNamesAndNumbers = tree.xpath("//item/title") caseDateRegex = re.compile("(\d{2}/\d{2}/\d{4})", re.VERBOSE | re.DOTALL) caseNumberRegex = re.compile("(\d{2}-.*?\W)(.*)$") i = 0 dupCount = 0 while i < len(caseLinks): # First: docType, since we don't support them all... docType = docTypes[i].text.strip() if "unpublished" in docType.lower(): documentType = "U" elif "published" in docType.lower(): documentType = "P" elif "errata" in docType.lower(): documentType = "E" else: # something weird we don't know about, punt i += 1 continue # next, we begin with the caseLink field caseLink = caseLinks[i].text caseLink = make_url_absolute(url, caseLink) # then we download the PDF, make the hash and document myFile, doc, created, dupCount = makeDocFromURL(caseLink, ct, dupCount) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # otherwise, we continue doc.documentType = documentType # next: caseDate caseDate = caseDateRegex.search(descriptions[i].text).group(1) splitDate = caseDate.split('/') caseDate = datetime.date(int(splitDate[2]), int(splitDate[0]), int(splitDate[1])) doc.dateFiled = caseDate # next: caseNumber caseNumber = caseNumberRegex.search(caseNamesAndNumbers[i].text)\ .group(1) # next: caseNameShort caseNameShort = caseNumberRegex.search(caseNamesAndNumbers[i].text)\ .group(2) # check for dups, make the object if necessary, otherwise, get it cite, created = hasDuplicate(caseNumber, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 2): """ queries can be made on their system via HTTP POST. """ # second circuit url = "http://www.ca2.uscourts.gov/decisions" ct = Court.objects.get(courtUUID='ca2') today = datetime.date.today() formattedToday = str(today.year) + str(today.month) + str(today.day) #formattedToday = str(today.year) + '03' + '07' data = "IW_DATABASE=OPN&IW_FIELD_TEXT=*&IW_FILTER_DATE_AFTER=" +\ formattedToday + "&IW_FILTER_DATE_BEFORE=&IW_BATCHSIZE=50&" +\ "IW_SORT=-DATE" req = urllib2.Request(url, data) response = urllib2.urlopen(req) html = response.read() soup = BeautifulSoup(html) aTagsRegex = re.compile('(.*?.pdf).*?', re.IGNORECASE) aTags = soup.findAll(attrs={'href' : aTagsRegex}) caseNumRegex = re.compile('.*/(.*?)_(.*?).pdf') i = 0 dupCount = 0 while i < len(aTags): # we begin with the caseLink field caseLink = aTags[i].get('href') caseLink = aTagsRegex.search(caseLink).group(1) caseLink = make_url_absolute(url, caseLink) myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # using caseLink, we can get the caseNumber and documentType caseNum = caseNumRegex.search(caseLink).group(1) # and the docType documentType = caseNumRegex.search(caseLink).group(2) if 'opn' in documentType: # it's unpublished doc.documentType = "P" elif 'so' in documentType: doc.documentType = "U" # next, the caseNameShort (there's probably a better way to do this. caseNameShort = aTags[i].parent.parent.nextSibling.nextSibling\ .nextSibling.nextSibling.contents[0] # next, we can do the caseDate caseDate = aTags[i].parent.parent.nextSibling.nextSibling\ .nextSibling.nextSibling.nextSibling.nextSibling.contents[0]\ .replace('&nbsp;', ' ').strip() # some caseDate cleanup splitDate = caseDate.split('-') caseDate = datetime.date(int(splitDate[2]),int(splitDate[0]), int(splitDate[1])) doc.dateFiled = caseDate # check for duplicates, make the object in their absence cite, created = hasDuplicate(caseNum, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 3): """ This URL provides the latest 25 cases, so I need to pick out the new ones and only get those. I can do this efficiently by trying to do each, and then giving up once I hit one that I've done before. This will work because they are in reverse chronological order. """ # if these URLs change, the docType identification (below) will need # to be updated. It's lazy, but effective. urls = ("http://www.ca3.uscourts.gov/recentop/week/recprec.htm", "http://www.ca3.uscourts.gov/recentop/week/recnon2day.htm",) ct = Court.objects.get(courtUUID='ca3') for url in urls: html = urllib2.urlopen(url) soup = BeautifulSoup(html) # all links ending in pdf, case insensitive regex = re.compile("pdf$", re.IGNORECASE) aTags = soup.findAll(attrs={"href": regex}) # we will use these vars in our while loop, better not to compile them # each time regexII = re.compile('\d{2}/\d{2}/\d{2}') regexIII = re.compile('\d{2}-\d{4}') i = 0 dupCount = 0 while i < len(aTags): # caseLink and caseNameShort caseLink = aTags[i].get('href') myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 caseNameShort = aTags[i].contents[0] # caseDate and caseNumber junk = aTags[i].previous.previous.previous try: # this error seems to happen upon dups...not sure why yet caseDate = regexII.search(junk).group(0) caseNumber = regexIII.search(junk).group(0) except: i = i+1 continue # next up is the caseDate splitDate = caseDate.split('/') caseDate = datetime.date(int("20" + splitDate[2]),int(splitDate[0]), int(splitDate[1])) doc.dateFiled = caseDate # Make a decision about the docType. if "recprec.htm" in str(url): doc.documentType = "P" elif "recnon2day.htm" in str(url): doc.documentType = "U" cite, created = hasDuplicate(caseNumber, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 4): """The fourth circuit is THE worst form of HTML I've ever seen. It's going to break a lot, but I've done my best to clean it up, and make it reliable.""" url = "http://pacer.ca4.uscourts.gov/opinions_today.htm" ct = Court.objects.get(courtUUID='ca4') html = urllib2.urlopen(url).read() # sadly, beautifulsoup chokes on the lines lines of this file because # the HTML is so bad. Stop laughing - the HTML IS awful, but it's not # funny. Anyway, to make this thing work, we must pull out the target # attributes. And so we do. regex = re.compile("target.*>", re.IGNORECASE) html = re.sub(regex, ">", html) soup = BeautifulSoup(html) # all links ending in pdf, case insensitive regex = re.compile("pdf$", re.IGNORECASE) aTags = soup.findAll(attrs={"href": regex}) i = 0 dupCount = 0 regexII = re.compile('\d{2}/\d{2}/\d{4}') regexIII = re.compile('\d{4}(.*)') while i < len(aTags): # caseLink field, and save it caseLink = aTags[i].get('href') caseLink = make_url_absolute(url, caseLink) myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # using caselink, we can get the caseNumber and documentType fileName = caseLink.split('/')[-1] caseNumber, documentType = fileName.split('.')[0:2] # the caseNumber needs a hyphen inserted after the second digit caseNumber = caseNumber[0:2] + "-" + caseNumber[2:] doc.documentType = documentType # next, we do the caseDate and caseNameShort, so we can quit before # we get too far along. junk = aTags[i].contents[0].replace('&nbsp;', ' ').strip() try: # this error seems to happen upon dups...not sure why yet caseDate = regexII.search(junk).group(0).strip() caseNameShort = regexIII.search(junk).group(1).strip() except: i += 1 continue # some caseDate cleanup splitDate = caseDate.split('/') caseDate = datetime.date(int(splitDate[2]),int(splitDate[0]), int(splitDate[1])) doc.dateFiled = caseDate # let's check for duplicates before we proceed cite, created = hasDuplicate(caseNumber, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 5): """Fifth circuit scraper. Similar process as to elsewhere, as you might expect at this point""" url = "http://www.ca5.uscourts.gov/Opinions.aspx" ct = Court.objects.get(courtUUID='ca5') html = urllib2.urlopen(url) soup = BeautifulSoup(html) #all links ending in pdf, case insensitive aTagRegex = re.compile("pdf$", re.IGNORECASE) aTags = soup.findAll(attrs={"href": aTagRegex}) opinionRegex = re.compile(r"\opinions.*pub") unpubRegex = re.compile(r"\opinions\unpub") i = 0 dupCount = 0 while i < len(aTags): # this page has PDFs that aren't cases, we must filter them out if opinionRegex.search(str(aTags[i])) == None: # it's not an opinion, increment and punt i += 1 continue # we begin with the caseLink field caseLink = aTags[i].get('href') caseLink = make_url_absolute(url, caseLink) myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # using caseLink, we can get the caseNumber and documentType caseNumber = aTags[i].contents[0] if unpubRegex.search(str(aTags[i])) == None: # it's published, else it's unpublished documentType = "P" else: documentType = "U" doc.documentType = documentType # next, we do the caseDate caseDate = aTags[i].next.next.contents[0].contents[0] # some caseDate cleanup splitDate = caseDate.split('/') caseDate = datetime.date(int(splitDate[2]),int(splitDate[0]), int(splitDate[1])) doc.dateFiled = caseDate # next, we do the caseNameShort caseNameShort = aTags[i].next.next.next.next.next.contents[0]\ .contents[0] # now that we have the caseNumber and caseNameShort, we can dup check cite, created = hasDuplicate(caseNumber, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 6): """Results are available without an HTML POST, but those results lack a date field. Hence, we must do an HTML POST.""" url = "http://www.ca6.uscourts.gov/cgi-bin/opinions.pl" ct = Court.objects.get(courtUUID = 'ca6') today = datetime.date.today() formattedToday = str(today.month) + '/' + str(today.day) + '/' +\ str(today.year) postValues = { 'CASENUM' : '', 'TITLE' : '', 'FROMDATE' : formattedToday, 'TODATE' : formattedToday, 'OPINNUM' : '' } data = urllib.urlencode(postValues) req = urllib2.Request(url, data) response = urllib2.urlopen(req) html = response.read() soup = BeautifulSoup(html) aTagsRegex = re.compile('pdf$', re.IGNORECASE) aTags = soup.findAll(attrs={'href' : aTagsRegex}) i = 0 dupCount = 0 while i < len(aTags): # we begin with the caseLink field caseLink = aTags[i].get('href') caseLink = make_url_absolute(url, caseLink) myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # using caseLink, we can get the caseNumber and documentType caseNumber = aTags[i].next.next.next.next.next.contents[0].strip()\ .replace('&nbsp;','') # using the filename, we can determine the documentType... fileName = aTags[i].contents[0] if 'n' in fileName: # it's unpublished doc.documentType = "U" elif 'p' in fileName: doc.documentType = "P" # next, we can do the caseDate caseDate = aTags[i].next.next.next.next.next.next.next.next\ .contents[0].replace('&nbsp;', ' ').strip() # some caseDate cleanup splitDate = caseDate.split('/') caseDate = datetime.date(int(splitDate[0]),int(splitDate[1]), int(splitDate[2])) doc.dateFiled = caseDate # next, the caseNameShort (there's probably a better way to do this. caseNameShort = aTags[i].next.next.next.next.next.next.next.next\ .next.next.next # now that we have the caseNumber and caseNameShort, we can dup check cite, created = hasDuplicate(caseNumber, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 7): """another court where we need to do a post. This will be a good starting place for getting the judge field, when we're ready for that""" url = "http://www.ca7.uscourts.gov/fdocs/docs.fwx" ct = Court.objects.get(courtUUID = 'ca7') # if these strings change, check that documentType still gets set correctly. dataStrings = ("yr=&num=&Submit=Today&dtype=Opinion&scrid=Select+a+Case", "yr=&num=&Submit=Today&dtype=Nonprecedential+Disposition&scrid=Select+a+Case") for dataString in dataStrings: req = urllib2.Request(url, dataString) response = urllib2.urlopen(req) html = response.read() soup = BeautifulSoup(html) aTagsRegex = re.compile('pdf$', re.IGNORECASE) aTags = soup.findAll(attrs={'href' : aTagsRegex}) i = 0 while i < len(aTags): # we begin with the caseLink field caseLink = aTags[i].get("href") caseLink = make_url_absolute(url, caseLink) myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # using caseLink, we can get the caseNumber and documentType caseNumber = aTags[i].previous.previous.previous.previous.previous\ .previous.previous.previous.previous.previous.strip() # next up: caseDate caseDate = aTags[i].previous.previous.previous.contents[0].strip() splitDate = caseDate.split('/') caseDate = datetime.date(int(splitDate[2]), int(splitDate[0]), int(splitDate[1])) doc.dateFiled = caseDate # next up: caseNameShort caseNameShort = aTags[i].previous.previous.previous.previous\ .previous.previous.previous # next up: docStatus if "type=Opinion" in dataString: doc.documentType = 'P' elif "type=Nonprecedential+Disposition" in dataString: doc.documentType = 'U' # now that we have the caseNumber and caseNameShort, we can dup check cite, created = hasDuplicate(caseNumber, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 8): url = "http://www.ca8.uscourts.gov/cgi-bin/new/today2.pl" ct = Court.objects.get(courtUUID = 'ca8') html = urllib2.urlopen(url) soup = BeautifulSoup(html) aTagsRegex = re.compile('pdf$', re.IGNORECASE) aTags = soup.findAll(attrs={'href' : aTagsRegex}) caseNumRegex = re.compile('(\d{2})(\d{4})(u|p)', re.IGNORECASE) caseDateRegex = re.compile('(\d{2}/\d{2}/\d{4})(.*)(</b>)') i = 0 dupCount = 0 while i < len(aTags): # we begin with the caseLink field caseLink = aTags[i].get('href') caseLink = make_url_absolute(url, caseLink) myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # using caseLink, we can get the caseNumber and documentType junk = aTags[i].contents[0] caseNumber = caseNumRegex.search(junk).group(1) + "-" +\ caseNumRegex.search(junk).group(2) documentType = caseNumRegex.search(junk).group(3).upper() doc.documentType = documentType # caseDate is next on the block junk = str(aTags[i].next.next.next) caseDate = caseDateRegex.search(junk).group(1)\ .replace('&nbsp;', ' ').strip() caseNameShort = caseDateRegex.search(junk).group(2) # some caseDate cleanup splitDate = caseDate.split('/') caseDate = datetime.date(int(splitDate[2]),int(splitDate[0]), int(splitDate[1])) doc.dateFiled = caseDate # now that we have the caseNumber and caseNameShort, we can dup check cite, created = hasDuplicate(caseNumber, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 9): """This court, by virtue of having a javascript laden website, was very hard to parse properly. BeautifulSoup couldn't handle it at all, so lxml has to be used. lxml seems pretty useful, but it was a pain to learn.""" urls = ("http://www.ca9.uscourts.gov/opinions/?o_mode=view&amp;o_sort_field=19&amp;o_sort_type=DESC&o_page_size=100", "http://www.ca9.uscourts.gov/memoranda/?o_mode=view&amp;o_sort_field=21&amp;o_sort_type=DESC&o_page_size=100",) ct = Court.objects.get(courtUUID = 'ca9') for url in urls: req = urllib2.urlopen(url).read() tree = fromstring(req) if url == urls[0]: caseLinks = tree.xpath('//table[3]/tbody/tr/td/a') caseNumbers = tree.xpath('//table[3]/tbody/tr/td[2]/label') caseDates = tree.xpath('//table[3]/tbody/tr/td[6]/label') elif url == urls[1]: caseLinks = tree.xpath('//table[3]/tbody/tr/td/a') caseNumbers = tree.xpath('//table[3]/tbody/tr/td[2]/label') caseDates = tree.xpath('//table[3]/tbody/tr/td[7]/label') i = 0 dupCount = 0 while i < len(caseLinks): # we begin with the caseLink field caseLink = caseLinks[i].get('href') caseLink = make_url_absolute(url, caseLink) # special case if 'no memos filed' in caseLink.lower(): i += 1 continue myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # next, we'll do the caseNumber caseNumber = caseNumbers[i].text # next up: document type (static for now) if 'memoranda' in url: doc.documentType = "U" elif 'opinions' in url: doc.documentType = "P" # next up: caseDate splitDate = caseDates[i].text.split('/') caseDate = datetime.date(int(splitDate[2]), int(splitDate[0]), int(splitDate[1])) doc.dateFiled = caseDate #next up: caseNameShort caseNameShort = titlecase(caseLinks[i].text.lower()) # now that we have the caseNumber and caseNameShort, we can dup check cite, created = hasDuplicate(caseNumber, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 10): url = "http://www.ck10.uscourts.gov/opinions/new/daily_decisions.rss" ct = Court.objects.get(courtUUID = 'ca10') req = urllib2.urlopen(url) # this code gets rid of errant ampersands - they throw big errors. contents = req.read() if '&' in contents: punctuationRegex = re.compile(" & ") contents = re.sub(punctuationRegex, " &amp; ", contents) tree = etree.fromstring(contents) else: tree = etree.fromstring(contents) caseLinks = tree.xpath("//item/link") descriptions = tree.xpath("//item/description") docTypes = tree.xpath("//item/category") caseNames = tree.xpath("//item/title") caseDateRegex = re.compile("(\d{2}/\d{2}/\d{4})", re.VERBOSE | re.DOTALL) caseNumberRegex = re.compile("(\d{2}-\d{4})(.*)$") i = 0 dupCount = 0 while i < len(caseLinks): # we begin with the caseLink field caseLink = caseLinks[i].text caseLink = make_url_absolute(url, caseLink) myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # next: docType (this order of if statements IS correct) docType = docTypes[i].text.strip() if "unpublished" in docType.lower(): doc.documentType = "U" elif "published" in docType.lower(): doc.documentType = "P" else: # it's an errata, or something else we don't care about i += 1 continue # next: caseDate caseDate = caseDateRegex.search(descriptions[i].text).group(1) splitDate = caseDate.split('/') caseDate = datetime.date(int(splitDate[2]), int(splitDate[0]), int(splitDate[1])) doc.dateFiled = caseDate # next: caseNumber caseNumber = caseNumberRegex.search(descriptions[i].text)\ .group(1) # next: caseNameShort caseNameShort = caseNames[i].text # check for dups, make the object if necessary, otherwise, get it cite, created = hasDuplicate(caseNumber, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 11): """Trying out an RSS feed this time, since the feed looks good. This court lacks a feed for unpublished opinions, so we don't fetch them at present""" url = "http://www.ca11.uscourts.gov/rss/pubopnsfeed.php" ct = Court.objects.get(courtUUID = 'ca11') req = urllib2.urlopen(url) # this code gets rid of errant ampersands - they throw big errors. contents = req.read() if '&' in contents: punctuationRegex = re.compile(" & ") contents = re.sub(punctuationRegex, " &amp; ", contents) tree = etree.fromstring(contents) else: tree = etree.fromstring(contents) caseLinks = tree.xpath('//item/link') description = tree.xpath('//item/description') caseNames = tree.xpath('//item/title') # some regexes caseNumRegex = re.compile(''' case # the word case .*? # some junk, not greedy (\d{2}-\d{5}) # two digits a hyphen then five more ''', re.IGNORECASE | re.VERBOSE) caseDateRegex = re.compile(''' date # the word date .*? # some junk, not greedy (\d{2}-\d{2}-\d{4}) # two digits - two digits - four digits ''', re.IGNORECASE | re.VERBOSE) i = 0 dupCount = 0 while i < len(caseLinks): # we begin with the caseLink field caseLink = caseLinks[i].text caseLink = make_url_absolute(url, caseLink) myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # these are only published opinions, unpublished lack a feed (boo!) doc.documentType = "P" # next, we'll do the caseNumber caseNumber = caseNumRegex.search(description[i].text).group(1) # next up: caseDate caseDate = caseDateRegex.search(description[i].text).group(1) splitDate = caseDate.split('-') caseDate = datetime.date(int(splitDate[2]), int(splitDate[0]), int(splitDate[1])) doc.dateFiled = caseDate #next up: caseNameShort caseNameShort = caseNames[i].text # now that we have the caseNumber and caseNameShort, we can dup check cite, created = hasDuplicate(caseNumber, caseNameShort) # if that goes well, we save to the DB doc.citation = cite # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 12): url = "http://www.cadc.uscourts.gov/bin/opinions/allopinions.asp" ct = Court.objects.get(courtUUID = 'cadc') html = urllib2.urlopen(url) soup = BeautifulSoup(html) aTagsRegex = re.compile('pdf$', re.IGNORECASE) aTags = soup.findAll(attrs={'href' : aTagsRegex}) caseNumRegex = re.compile("(\d{2}-\d{4})") i = 0 dupCount = 0 while i < len(aTags): # we begin with the caseLink field caseLink = aTags[i].get('href') caseLink = make_url_absolute(url, caseLink) myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # using caseLink, we can get the caseNumber caseNumber = caseNumRegex.search(caseLink).group(1) # we can hard-code this b/c the D.C. Court paywalls all # unpublished opinions. doc.documentType = "P" # caseDate is next on the block caseDate = datetime.date.today() doc.dateFiled = caseDate caseNameShort = aTags[i].next.next.next # now that we have the caseNumber and caseNameShort, we can dup check cite, created = hasDuplicate(caseNumber, caseNameShort) # if that goes well, we save to the DB doc.citation = cite # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result elif (courtID == 13): url = "http://www.cafc.uscourts.gov/dailylog.html" ct = Court.objects.get(courtUUID = "cafc") html = urllib2.urlopen(url) soup = BeautifulSoup(html) aTagsRegex = re.compile('pdf$', re.IGNORECASE) trTags = soup.findAll('tr') # start on the second row, since the first is headers. i = 1 dupCount = 0 while i <= 50: #stop at 50, if no triple dups first. try: caseLink = trTags[i].td.nextSibling.nextSibling.nextSibling\ .nextSibling.nextSibling.nextSibling.a.get('href').strip('.') caseLink = make_url_absolute(url, caseLink) if 'opinion' not in caseLink: # we have a non-case PDF. punt i += 1 continue except: # the above fails when things get funky, in that case, we punt i += 1 continue myFile, doc, created = makeDocFromURL(caseLink, ct) if not created: # it's an oldie, punt! result += "Duplicate found at " + str(i) + "<br>" dupCount += 1 if dupCount == 3: # third dup in a a row. BREAK! break i += 1 continue else: dupCount = 0 # next: caseNumber caseNumber = trTags[i].td.nextSibling.nextSibling.contents[0]\ .strip('.pdf') # next: dateFiled dateFiled = trTags[i].td.contents splitDate = dateFiled[0].split("/") dateFiled = datetime.date(int(splitDate[0]), int(splitDate[1]), int(splitDate[2])) doc.dateFiled = dateFiled # next: caseNameShort caseNameShort = trTags[i].td.nextSibling.nextSibling.nextSibling\ .nextSibling.nextSibling.nextSibling.a.contents[0] # next: documentType documentType = trTags[i].td.nextSibling.nextSibling.nextSibling\ .nextSibling.nextSibling.nextSibling.nextSibling.nextSibling\ .contents[0].contents[0] # normalize the result for our internal purposes... if documentType == 'N': documentType = 'U' doc.documentType = documentType # now that we have the caseNumber and caseNameShort, we can dup check cite, created = hasDuplicate(caseNumber, caseNameShort) # last, save evrything (pdf, citation and document) doc.citation = cite doc.local_path.save(caseNameShort + ".pdf", myFile) doc.save() i += 1 return result if (courtID == 14): # we do SCOTUS pass
1cb7f4014a5290c6300b47a862966c4595810970 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6762/1cb7f4014a5290c6300b47a862966c4595810970/views.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 888, 25360, 29328, 88, 12, 71, 477, 88, 734, 16, 563, 4672, 309, 261, 71, 477, 88, 734, 422, 404, 4672, 3536, 12667, 87, 854, 2319, 628, 326, 1122, 12937, 309, 1846, 1960, 358, 3675, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 888, 25360, 29328, 88, 12, 71, 477, 88, 734, 16, 563, 4672, 309, 261, 71, 477, 88, 734, 422, 404, 4672, 3536, 12667, 87, 854, 2319, 628, 326, 1122, 12937, 309, 1846, 1960, 358, 3675, ...
case_id = self.create(cr, uid, val) ids.append(case_id)
case_id = self.create(cr, uid, val) ids.append(case_id)
def import_cal(self, cr, uid, data, context={}): file_content = base64.decodestring(data) event_obj = self.pool.get('basic.calendar.event') event_obj.__attribute__.update(self.__attribute__) vals = event_obj.import_ical(cr, uid, file_content) ids = [] for val in vals: is_exists = common.uid2openobjectid(cr, val['id'], self._name) if val.has_key('create_date'): val.pop('create_date') val['caldav_url'] = context.get('url') or '' val.pop('id') if is_exists: self.write(cr, uid, [is_exists], val) ids.append(is_exists) else: case_id = self.create(cr, uid, val) ids.append(case_id) return ids
75258015cae6ae0d493e7eb6204a16eafed73cd9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/75258015cae6ae0d493e7eb6204a16eafed73cd9/crm_meeting.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1930, 67, 771, 12, 2890, 16, 4422, 16, 4555, 16, 501, 16, 819, 12938, 4672, 585, 67, 1745, 273, 1026, 1105, 18, 4924, 1145, 371, 12, 892, 13, 871, 67, 2603, 273, 365, 18, 6011, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1930, 67, 771, 12, 2890, 16, 4422, 16, 4555, 16, 501, 16, 819, 12938, 4672, 585, 67, 1745, 273, 1026, 1105, 18, 4924, 1145, 371, 12, 892, 13, 871, 67, 2603, 273, 365, 18, 6011, 18, ...
else:
if doc == 'geoserver':
def build(): # fix with pushd(download_path): with pushd(docs_path): for doc in config.options(section): info("Build docs for %s" % doc) app_doc = path(doc) with pushd(app_doc): if doc == 'geoext': sh("sphinx-build -bhtml . html") else: sh("sphinx-build -bhtml source html")
9e70da551d35668cfe2fd1051e7da5e5e2be3221 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3279/9e70da551d35668cfe2fd1051e7da5e5e2be3221/pavement.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1361, 13332, 468, 2917, 598, 1817, 72, 12, 7813, 67, 803, 4672, 598, 1817, 72, 12, 8532, 67, 803, 4672, 364, 997, 316, 642, 18, 2116, 12, 3464, 4672, 1123, 2932, 3116, 3270, 364, 738, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1361, 13332, 468, 2917, 598, 1817, 72, 12, 7813, 67, 803, 4672, 598, 1817, 72, 12, 8532, 67, 803, 4672, 364, 997, 316, 642, 18, 2116, 12, 3464, 4672, 1123, 2932, 3116, 3270, 364, 738, ...
\begin_inset
\begin_inset
def revert_listings_inset(document): r''' Revert listings inset to \lstinline or \begin, \end lstlisting, translate
dec8e81ebdf0e6978560ba351aaedd7878e8a79b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7514/dec8e81ebdf0e6978560ba351aaedd7878e8a79b/lyx_1_5.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 15226, 67, 1098, 899, 67, 267, 542, 12, 5457, 4672, 436, 26418, 868, 1097, 666, 899, 316, 542, 358, 521, 16923, 10047, 578, 521, 10086, 16, 521, 409, 9441, 21228, 16, 4204, 2, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 15226, 67, 1098, 899, 67, 267, 542, 12, 5457, 4672, 436, 26418, 868, 1097, 666, 899, 316, 542, 358, 521, 16923, 10047, 578, 521, 10086, 16, 521, 409, 9441, 21228, 16, 4204, 2, -100, -1...
heappushpop(result, elem)
_heappushpop(result, elem)
def nlargest(n, iterable): """Find the n largest elements in a dataset. Equivalent to: sorted(iterable, reverse=True)[:n] """ it = iter(iterable) result = list(islice(it, n)) if not result: return result heapify(result) _heappushpop = heappushpop for elem in it: heappushpop(result, elem) result.sort(reverse=True) return result
005691263df8cc222d90402c57ad1b63a53deee9 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8125/005691263df8cc222d90402c57ad1b63a53deee9/heapq.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7741, 11475, 12, 82, 16, 6087, 4672, 3536, 3125, 326, 290, 12756, 2186, 316, 279, 3709, 18, 225, 31208, 358, 30, 225, 3115, 12, 15364, 16, 4219, 33, 5510, 13, 10531, 82, 65, 3536, 518,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 7741, 11475, 12, 82, 16, 6087, 4672, 3536, 3125, 326, 290, 12756, 2186, 316, 279, 3709, 18, 225, 31208, 358, 30, 225, 3115, 12, 15364, 16, 4219, 33, 5510, 13, 10531, 82, 65, 3536, 518,...
self.write(cr, uid, ids, {'state':'confirmed'}) i=0 def create_chained_picking(self,cr,uid,moves,context): new_moves=[]
self.write(cr, uid, ids, {'state': 'confirmed'}) i = 0 def create_chained_picking(self, cr, uid, moves, context): new_moves = []
def action_confirm(self, cr, uid, ids, context={}):
369221b47101072e094ad2d02fe2edd2b47690aa /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7397/369221b47101072e094ad2d02fe2edd2b47690aa/stock.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1301, 67, 10927, 12, 2890, 16, 4422, 16, 4555, 16, 3258, 16, 819, 12938, 4672, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1301, 67, 10927, 12, 2890, 16, 4422, 16, 4555, 16, 3258, 16, 819, 12938, 4672, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -1...
opts, args = getopt.getopt(sys.argv[1:], 'lfd:')
opts, args = getopt.getopt(sys.argv[1:], 'lfd:x:')
def main(): """Script main program.""" import getopt try: opts, args = getopt.getopt(sys.argv[1:], 'lfd:') except getopt.error, msg: print msg print "usage: compileall [-l] [-f] [-d destdir] [directory ...]" print "-l: don't recurse down" print "-f: force rebuild even if timestamps are up-to-date" print "-d destdir: purported directory name for error messages" print "if no directory arguments, -l sys.path is assumed" sys.exit(2) maxlevels = 10 ddir = None force = 0 for o, a in opts: if o == '-l': maxlevels = 0 if o == '-d': ddir = a if o == '-f': force = 1 if ddir: if len(args) != 1: print "-d destdir require exactly one directory argument" sys.exit(2) success = 1 try: if args: for dir in args: success = success and compile_dir(dir, maxlevels, ddir, force) else: success = compile_path() except KeyboardInterrupt: print "\n[interrupt]" success = 0 return success
59776b466a8795b66b9c91444afb05cff217b099 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/59776b466a8795b66b9c91444afb05cff217b099/compileall.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 3536, 3651, 2774, 5402, 12123, 1930, 336, 3838, 775, 30, 1500, 16, 833, 273, 336, 3838, 18, 588, 3838, 12, 9499, 18, 19485, 63, 21, 30, 6487, 296, 80, 8313, 30, 92, 2497, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 3536, 3651, 2774, 5402, 12123, 1930, 336, 3838, 775, 30, 1500, 16, 833, 273, 336, 3838, 18, 588, 3838, 12, 9499, 18, 19485, 63, 21, 30, 6487, 296, 80, 8313, 30, 92, 2497, ...
ppcOnly = False elif platform.processor() == "powerpc": print "Building a PPC-only binary" ppcOnly = True
preferPPC = False
def addDataFiles(dataFiles, fromDir, toSubDir=None, inclHiddenDirs=False): """Find data files and format data for the data_files argument of setup. In/Out: - dataFiles: a list to which is appended zero or more of these elements: [subDir, list of paths to resource files] Inputs: - fromDir: path to root directory of existing resource files - toSubDir: relative path to resources in package; if omitted then the final dir of fromDir is used - inclHiddenDirs: if True, the contents of directories whose names start with "." are included Returns a list of the following elements: """ lenFromDir = len(fromDir) if toSubDir == None: toSubDir = os.path.split(fromDir)[1] for (dirPath, dirNames, fileNames) in os.walk(fromDir): if not inclHiddenDirs: dirNamesCopy = dirNames[:] for ii in range(len(dirNamesCopy)-1, -1, -1): if dirNamesCopy[ii].startswith("."): del(dirNames[ii]) if not dirPath.startswith(fromDir): raise RuntimeError("Cannot deal with %r files; %s does not start with %r" %\ (resBase, dirPath, fromDir)) toPath = os.path.join(toSubDir, dirPath[lenFromDir+1:]) filePaths = [os.path.join(dirPath, fileName) for fileName in fileNames] dataFiles.append((toPath, filePaths))
8ca36d123e004717e35e40dcd11394fcbb58f096 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/6236/8ca36d123e004717e35e40dcd11394fcbb58f096/setup.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 21958, 2697, 12, 892, 2697, 16, 628, 1621, 16, 358, 1676, 1621, 33, 7036, 16, 2823, 9345, 9872, 33, 8381, 4672, 3536, 3125, 501, 1390, 471, 740, 501, 364, 326, 501, 67, 2354, 1237, 434...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 21958, 2697, 12, 892, 2697, 16, 628, 1621, 16, 358, 1676, 1621, 33, 7036, 16, 2823, 9345, 9872, 33, 8381, 4672, 3536, 3125, 501, 1390, 471, 740, 501, 364, 326, 501, 67, 2354, 1237, 434...
print 'find_bzr', path
def find_bzr(path=None): """ Return bzr root directory path, or None. """ if not path: path = os.getcwd() if not os.path.exists(path): return None while path: print 'find_bzr', path if os.path.exists(os.path.join(path, '.bzr')) or \ os.path.exists(os.path.join(path, '.bzrignore')): return os.path.abspath(path) else: pth = path path = os.path.dirname(path) if path == pth: return None return None
7df0fcf451721354710b28f5d16d4a7153c0c6da /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12610/7df0fcf451721354710b28f5d16d4a7153c0c6da/repo.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1104, 67, 25292, 86, 12, 803, 33, 7036, 4672, 3536, 2000, 24788, 86, 1365, 1867, 589, 16, 578, 599, 18, 3536, 309, 486, 589, 30, 589, 273, 1140, 18, 588, 11089, 1435, 309, 486, 1140, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1104, 67, 25292, 86, 12, 803, 33, 7036, 4672, 3536, 2000, 24788, 86, 1365, 1867, 589, 16, 578, 599, 18, 3536, 309, 486, 589, 30, 589, 273, 1140, 18, 588, 11089, 1435, 309, 486, 1140, ...
element_class = Set_generic
def _element_constructor_(self, x): """ TESTS:: sage: S3 = Subsets(3); S3([1,2]) {1, 2} sage: S3([0,1,2]) Traceback (most recent call last): ... ValueError: [0, 1, 2] not in Subsets of {1, 2, 3} """ return Set(x) element_class = Set_object_enumerated
def unrank(self, r): """ Returns the subset of s that has rank k. EXAMPLES:: sage: Subsets(3).unrank(0) {} sage: Subsets([2,4,5]).unrank(1) {2} sage: s = Subsets([2,4,5]) """
dc716b96cd454a10479e28fc7daa50f2ef780219 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9890/dc716b96cd454a10479e28fc7daa50f2ef780219/subset.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 640, 11500, 12, 2890, 16, 436, 4672, 3536, 2860, 326, 7931, 434, 272, 716, 711, 6171, 417, 18, 225, 5675, 8900, 11386, 2866, 225, 272, 410, 30, 2592, 4424, 12, 23, 2934, 318, 11500, 12...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 640, 11500, 12, 2890, 16, 436, 4672, 3536, 2860, 326, 7931, 434, 272, 716, 711, 6171, 417, 18, 225, 5675, 8900, 11386, 2866, 225, 272, 410, 30, 2592, 4424, 12, 23, 2934, 318, 11500, 12...
def iter_packages(self, packages, selected, hint=False):
def iter_packages(self, packages, selected, hint=False, nuninst=None):
def iter_packages(self, packages, selected, hint=False): """Iter on the list of actions and apply them one-by-one
ab0576817151497fdda52a6927ea2e47cb997bd6 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2784/ab0576817151497fdda52a6927ea2e47cb997bd6/britney.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1400, 67, 10308, 12, 2890, 16, 5907, 16, 3170, 16, 7380, 33, 8381, 16, 290, 318, 8591, 33, 7036, 4672, 3536, 2360, 603, 326, 666, 434, 4209, 471, 2230, 2182, 1245, 17, 1637, 17, 476, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1400, 67, 10308, 12, 2890, 16, 5907, 16, 3170, 16, 7380, 33, 8381, 16, 290, 318, 8591, 33, 7036, 4672, 3536, 2360, 603, 326, 666, 434, 4209, 471, 2230, 2182, 1245, 17, 1637, 17, 476, ...
curObj = None if self._curObjRef is None: self._curObjRef = self._baseObjRef
if curObjRef is None: startRefWorkingList = workingListSelector.next() while True: yield None try: curObjRef = startRefWorkingList.refGen.next() break except StopIteration: if len(startRefWorkingList.source) == 0: break for flywheel in makeFlywheelGen( startRefWorkingList.source.values(), countFunc=lambda x: self.getStartObjAffinity(x), scale=.05): yield None startRefWorkingList.refGen = flywheel if curObjRef is None: continue if type(curObjRef) in (types.IntType, types.LongType): startId = curObjRef curObjRef = None try: for containerRef in self._leakDetector.getContainerByIdGen(startId): yield None except: self.notify.debug('invalid startRef, stored as id %s' % startId) self._leakDetector.removeContainerById(startId) continue curObjRef = containerRef
def run(self): try: while True: # yield up here instead of at the end, since we skip back to the # top of the while loop from various points yield None #import pdb;pdb.set_trace() curObj = None if self._curObjRef is None: self._curObjRef = self._baseObjRef try: for result in self._curObjRef.getContainer(): yield None curObj = result except: self.notify.debug('lost current container: %s' % self._curObjRef) # that container is gone, try again self._curObjRef = None continue self.notify.debug('--> %s' % self._curObjRef)
d9560fa19bec29d4ea08e0b58cd1b7384659f9ff /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7242/d9560fa19bec29d4ea08e0b58cd1b7384659f9ff/ContainerLeakDetector.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 4672, 775, 30, 1323, 1053, 30, 468, 2824, 731, 2674, 3560, 434, 622, 326, 679, 16, 3241, 732, 2488, 1473, 358, 326, 468, 1760, 434, 326, 1323, 2798, 628, 11191, 3143, 2...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 4672, 775, 30, 1323, 1053, 30, 468, 2824, 731, 2674, 3560, 434, 622, 326, 679, 16, 3241, 732, 2488, 1473, 358, 326, 468, 1760, 434, 326, 1323, 2798, 628, 11191, 3143, 2...
print "Group order now ",n1*n2,"=",n1,"*",n2
print "Subgroup order now ",n1*n2,"=",n1,"*",n2
def abelian_group(self, debug=False): r""" Returns the abelian group structure of the group of points on this elliptic curve.
7e0308bfa96d392f0529a28f30a5ec4774dcfbb8 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/7e0308bfa96d392f0529a28f30a5ec4774dcfbb8/ell_finite_field.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1223, 292, 2779, 67, 1655, 12, 2890, 16, 1198, 33, 8381, 4672, 436, 8395, 2860, 326, 1223, 292, 2779, 1041, 3695, 434, 326, 1041, 434, 3143, 603, 333, 415, 549, 21507, 8882, 18, 2, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1223, 292, 2779, 67, 1655, 12, 2890, 16, 1198, 33, 8381, 4672, 436, 8395, 2860, 326, 1223, 292, 2779, 1041, 3695, 434, 326, 1041, 434, 3143, 603, 333, 415, 549, 21507, 8882, 18, 2, -10...
if xsource.find('Recommended').text == 'true':
if xsource.find('Recommended').text in ['True', 'true']:
def source_from_xml(xsource): ret = dict([('rawurl', False), ('url', False)]) for key, tag in [('groups', 'Group'), ('components', 'Component'), ('arches', 'Arch'), ('blacklist', 'Blacklist')]: ret[key] = [item.text for item in xsource.findall(tag)] # version and component need to both contain data for sources to work try: ret['version'] = xsource.find('Version').text except: ret['version'] = 'placeholder' if ret['components'] == []: ret['components'] = ['placeholder'] try: if xsource.find('Recommended').text == 'true': ret['recommended'] = True except: ret['recommended'] = False if xsource.find('RawURL') is not None: ret['rawurl'] = xsource.find('RawURL').text if not ret['rawurl'].endswith('/'): ret['rawurl'] += '/' else: ret['url'] = xsource.find('URL').text if not ret['url'].endswith('/'): ret['url'] += '/' return ret
c688c4592be717cb28dabd13d6441c4c7c5bca06 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11867/c688c4592be717cb28dabd13d6441c4c7c5bca06/Packages.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1084, 67, 2080, 67, 2902, 12, 92, 3168, 4672, 325, 273, 2065, 3816, 2668, 1899, 718, 2187, 1083, 3631, 7707, 718, 2187, 1083, 13, 5717, 364, 498, 16, 1047, 316, 306, 2668, 4650, 2187, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1084, 67, 2080, 67, 2902, 12, 92, 3168, 4672, 325, 273, 2065, 3816, 2668, 1899, 718, 2187, 1083, 3631, 7707, 718, 2187, 1083, 13, 5717, 364, 498, 16, 1047, 316, 306, 2668, 4650, 2187, ...
paragraph_element = content_context.get_element_list('//text:p')
paragraph_element = content_context.get_element_list('//text:p[1]')
def setUp(self): container = get_odf_container('samples/example.odt') self.container = container content_context = odf_context('content', container) self.content_context = content_context # TODO make "//text:p[0]" pass paragraph_element = content_context.get_element_list('//text:p') self.paragraph_element = paragraph_element[0]
c9486ff7a8d654b48e9af6077ea60fa81b536bfe /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10612/c9486ff7a8d654b48e9af6077ea60fa81b536bfe/test_context.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24292, 12, 2890, 4672, 1478, 273, 336, 67, 369, 74, 67, 3782, 2668, 7319, 19, 8236, 18, 369, 88, 6134, 365, 18, 3782, 273, 1478, 913, 67, 2472, 273, 320, 2180, 67, 2472, 2668, 1745, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24292, 12, 2890, 4672, 1478, 273, 336, 67, 369, 74, 67, 3782, 2668, 7319, 19, 8236, 18, 369, 88, 6134, 365, 18, 3782, 273, 1478, 913, 67, 2472, 273, 320, 2180, 67, 2472, 2668, 1745, ...
a(formatstring((header for k in months), colwidth, c).rstrip())
headers = (header for k in months) a(formatstring(headers, colwidth, c).rstrip())
def formatyear(self, theyear, w=2, l=1, c=6, m=3): """ Returns a year's calendar as a multi-line string. """ w = max(2, w) l = max(1, l) c = max(2, c) colwidth = (w + 1) * 7 - 1 v = [] a = v.append a(repr(theyear).center(colwidth*m+c*(m-1)).rstrip()) a('\n'*l) header = self.formatweekheader(w) for (i, row) in enumerate(self.yeardays2calendar(theyear, m)): # months in this row months = xrange(m*i+1, min(m*(i+1)+1, 13)) a('\n'*l) a(formatstring((month_name[k] for k in months), colwidth, c).rstrip()) a('\n'*l) a(formatstring((header for k in months), colwidth, c).rstrip()) a('\n'*l) # max number of weeks for this row height = max(len(cal) for cal in row) for j in xrange(height): weeks = [] for cal in row: if j >= len(cal): weeks.append('') else: weeks.append(self.formatweek(cal[j], w)) a(formatstring(weeks, colwidth, c).rstrip()) a('\n' * l) return ''.join(v)
243815ee949d8494ecadb24f0872307909b05fc8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/243815ee949d8494ecadb24f0872307909b05fc8/calendar.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 740, 6874, 12, 2890, 16, 2898, 2091, 16, 341, 33, 22, 16, 328, 33, 21, 16, 276, 33, 26, 16, 312, 33, 23, 4672, 3536, 2860, 279, 3286, 1807, 5686, 487, 279, 3309, 17, 1369, 533, 18,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 740, 6874, 12, 2890, 16, 2898, 2091, 16, 341, 33, 22, 16, 328, 33, 21, 16, 276, 33, 26, 16, 312, 33, 23, 4672, 3536, 2860, 279, 3286, 1807, 5686, 487, 279, 3309, 17, 1369, 533, 18,...
for i in l + [n[1]]:
for i in l: self.ui.debug("narrowing %d:%d %s\n" % (f, len(l), short(i)))
def getchangegroup(self, remote): m = self.changelog.nodemap search = [] fetch = [] seen = {} seenbranch = {}
3832fb88ce2f811c47b456520408be17f4aae77b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11312/3832fb88ce2f811c47b456520408be17f4aae77b/hg.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 3427, 1655, 12, 2890, 16, 2632, 4672, 312, 273, 365, 18, 24083, 12970, 18, 2159, 1458, 1623, 273, 5378, 2158, 273, 5378, 5881, 273, 2618, 5881, 7500, 273, 2618, 2, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 3427, 1655, 12, 2890, 16, 2632, 4672, 312, 273, 365, 18, 24083, 12970, 18, 2159, 1458, 1623, 273, 5378, 2158, 273, 5378, 5881, 273, 2618, 5881, 7500, 273, 2618, 2, -100, -100, -100,...
if verify: flag = gas.solve(crval) else: flag = gas.solve()
xy = afwImage.PointD(1,1) rd = afwImage.PointD(334.2762566, -17.2097954) self.checkMatch(rd, xy)
def solveOrVerify(self, starlist, crval, crpix, plateScale=0, verify=False): """Test the solve() function Input: starlist List of objects as returned by loadXYFromFile crval lsst.afw.image.PointD ra/dec of a known position on the image crpix lsst.afw.image.PointD xy of a known position on the image plateScale Size of image in arcsec/pixel. Specifing this dramatically improves search time verify If True, crval is passed to to solve() to speed up the match, otherwise a solution is found with no inital guess at the position """ #Set plate scale if plateScale > 0: gas.setImageScaleArcsecPerPixel(plateScale) #Set starlist starlist = loadXYFromFile(starlist) gas.setStarlist(starlist) #Run solver if verify: flag = gas.solve(crval) else: flag = gas.solve()
4f02d21d4da7291cf5f3245909c3dc640b3fce4d /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6522/4f02d21d4da7291cf5f3245909c3dc640b3fce4d/GlobalAstrometrySolution.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 12439, 1162, 8097, 12, 2890, 16, 10443, 1098, 16, 4422, 1125, 16, 4422, 14861, 16, 225, 21972, 5587, 33, 20, 16, 3929, 33, 8381, 4672, 3536, 4709, 326, 12439, 1435, 445, 225, 2741, 30, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 12439, 1162, 8097, 12, 2890, 16, 10443, 1098, 16, 4422, 1125, 16, 4422, 14861, 16, 225, 21972, 5587, 33, 20, 16, 3929, 33, 8381, 4672, 3536, 4709, 326, 12439, 1435, 445, 225, 2741, 30, ...
def killed(self):
def killed(self):
def killed(self): #bruce 041029; totally revised by bruce 050702 """ (Public method) Report whether an atom has been killed. """ # Note: some "friend code" inlines this method for speed # (and omits the debug code). To find it, search for _Atom__killed # (the mangled version of __killed). [bruce 071018 comment] if debug_flags.atom_debug: # this cond is for speed mol = self.molecule from chunk import _nullMol better_alive_answer = mol is not None and self.key in mol.atoms and mol is not _nullMol ##e and mol is not killed??? if (not not better_alive_answer) != (not self.__killed): if debug_flags.atom_debug: #bruce 060414 re bug 1779, but it never printed for it (worth keeping in for other bugs) #bruce 071018 fixed typo of () after debug_flags.atom_debug -- could that be why it never printed it?!? print "debug: better_alive_answer is %r but (not self.__killed) is %r" % (better_alive_answer , not self.__killed) return self.__killed
ed409134e77b5e0b4516d0060e8f298611249859 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/11221/ed409134e77b5e0b4516d0060e8f298611249859/chem.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24859, 12, 2890, 4672, 468, 2848, 3965, 16486, 2163, 5540, 31, 9997, 1230, 283, 26779, 635, 5186, 3965, 374, 3361, 27, 3103, 3536, 261, 4782, 707, 13, 8706, 2856, 392, 3179, 711, 2118, 2...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 24859, 12, 2890, 4672, 468, 2848, 3965, 16486, 2163, 5540, 31, 9997, 1230, 283, 26779, 635, 5186, 3965, 374, 3361, 27, 3103, 3536, 261, 4782, 707, 13, 8706, 2856, 392, 3179, 711, 2118, 2...
if (updater_stat == "T" or updater_stat == "X" or updater_stat == "T"):
if (updater_stat == "T" or updater_stat == "X" or updater_stat == "Z"):
def nmclient_rawsay(nmip, nmport, *args): fullresponse = nmclient_rawcommunicate(nmip, nmport, *args) try: (response, status) = fullresponse.rsplit('\n',1) except KeyError: raise NMClientException, "Communication error '"+fullresponse+"'" if status == 'Success': return response elif status == 'Error': raise NMClientException, "Node Manager error '"+response+"'" elif status == 'Warning': raise NMClientException, "Node Manager warning '"+response+"'" else: raise NMClientException, "Unknown status '"+fullresponse+"'"
1492264581c4f4eef2330103bb89d205a710a2ed /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7263/1492264581c4f4eef2330103bb89d205a710a2ed/testprocess.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8442, 2625, 67, 1899, 87, 528, 12, 15094, 625, 16, 8442, 655, 16, 380, 1968, 4672, 1983, 2740, 273, 8442, 2625, 67, 1899, 5702, 318, 2659, 12, 15094, 625, 16, 8442, 655, 16, 380, 1968,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8442, 2625, 67, 1899, 87, 528, 12, 15094, 625, 16, 8442, 655, 16, 380, 1968, 4672, 1983, 2740, 273, 8442, 2625, 67, 1899, 5702, 318, 2659, 12, 15094, 625, 16, 8442, 655, 16, 380, 1968,...
lon, lat = float(site.lon), float(site.lat)
try: lon, lat = float(site.lon), float(site.lat) except TypeError: raise GeocodeError
def __init__(self, data=None, site=None, *args, **kwargs): lon, lat = float(site.lon), float(site.lat) super(OrderSettingsForm, self).__init__(data, *args, **kwargs) self.fields['delivery_area'].widget = EditableMap(options={ 'geometry': 'polygon', 'isCollection': True, 'layers': ['google.streets'], 'default_lat': lat, 'default_lon': lon, 'defaultZoom': 13, 'map_options': { 'controls': ['Navigation', 'PanZoom'] } }) self.fields['email'].initial = site.email self.fields['fax'].initial = site.fax_number self.site = site
d5f358dec9268b887716daf4572652e4007f5baf /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9903/d5f358dec9268b887716daf4572652e4007f5baf/forms.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 501, 33, 7036, 16, 2834, 33, 7036, 16, 380, 1968, 16, 2826, 4333, 4672, 775, 30, 4281, 16, 2516, 273, 1431, 12, 4256, 18, 9379, 3631, 1431, 12, 4256, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 501, 33, 7036, 16, 2834, 33, 7036, 16, 380, 1968, 16, 2826, 4333, 4672, 775, 30, 4281, 16, 2516, 273, 1431, 12, 4256, 18, 9379, 3631, 1431, 12, 4256, 1...
self.rules[path]['event_producer_thread'].stop()
o = self.rules.pop(path) o['event_producer_thread'].stop()
def remove_rule(self, path): if path in self.rules: self.rules[path]['event_producer_thread'].stop()
19f48676341f4f13ee6fd272f4be1c3471a41d87 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14346/19f48676341f4f13ee6fd272f4be1c3471a41d87/PollingWatcher.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1206, 67, 5345, 12, 2890, 16, 589, 4672, 309, 589, 316, 365, 18, 7482, 30, 365, 18, 7482, 63, 803, 23962, 2575, 67, 11776, 2750, 67, 5930, 29489, 5681, 1435, 2, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1206, 67, 5345, 12, 2890, 16, 589, 4672, 309, 589, 316, 365, 18, 7482, 30, 365, 18, 7482, 63, 803, 23962, 2575, 67, 11776, 2750, 67, 5930, 29489, 5681, 1435, 2, -100, -100, -100, -100,...
matt[n].specular_hardness=materials[idx_mat][3][4]
if raise_small_hardness and (materials[idx_mat][3][4]<default_hardeness): matt[n].specular_hardness=default_hardeness else: matt[n].specular_hardness=materials[idx_mat][3][4]
def create_materials(groups,materials,textures,orbiterpath): #counting material/texture combinations print("Creating materials") matpairset=set() matpair=[] # [(mat,tex),[mgroups...]] Unique mat+tex and corresponding groups for n in range(len(groups)): l=(groups[n][1],groups[n][2]) #print(l) if l not in matpairset: matpairset.add(l) matpair.append([l,[]]) #fill unique mat+tex combination for n in range(len(groups)): l=(groups[n][1],groups[n][2]) for i in range(len(matpair)): if l==matpair[i][0]: matpair[i][1].append(n) #fill array of corresponding groups print("\nUnique pairs:\n",matpairset) print(matpair) #create textures #TODO: upper/lower case handling in Linux tx=[] tex_load_fails=0 orbiter_path_ok=os.access(orbiterpath,os.F_OK) if not(orbiter_path_ok): print("Orbiter path is wrong! path=",orbiterpath) print("creating textures") for n in range(len(textures)): tx.append(bpy.data.textures.new(textures[n][1],"IMAGE")) if orbiter_path_ok: v=ntpath.split(textures[n][0]) print(v); fpath=os.path.join(orbiterpath,"Textures") for i in v: fpath=os.path.join(fpath,i) print (fpath) #Trying to load data try: img=bpy.data.images.load(fpath) except: print("!!!!!Can not load image: ",fpath) tex_load_fails=tex_load_fails+1 continue else: tex_load_fails=tex_load_fails+1 continue tx[n].image=img tx[n].use_alpha=True print("creating materials") n=0 matt=[] for pair in matpair: #create material object idx_mat=pair[0][0]-1 idx_tex=pair[0][1]-1 print("idx_mat=",idx_mat) print("mat_name=",materials[idx_mat][0]) print("diff=",materials[idx_mat][1][:3]) print("tex=",textures[idx_tex][1],"idx=",idx_tex) matt.append(bpy.data.materials.new(materials[idx_mat][0])) #diffuse component matt[n].diffuse_color=materials[idx_mat][1][:3] matt[n].alpha=materials[idx_mat][1][3] if materials[idx_mat][1][3]<1.0: matt[n].use_transparency=True #specular component matt[n].specular_color=materials[idx_mat][3][:3] matt[n].specular_alpha=materials[idx_mat][3][3] if len(materials[idx_mat][3])==5: matt[n].specular_hardness=materials[idx_mat][3][4] #there aren''t different ambient and emissive color component in blender #ambient is very often equal to diffuse, it's like amb=1.0 in blender #Emmissive component: import_emmissive=True; if import_emmissive: emm_c=materials[idx_mat][4][:3] matt[n].emit=(emm_c[0]+emm_c[1]+emm_c[2])/3 #Adding texture to material if idx_tex>=0: mtex=matt[n].texture_slots.add() mtex.texture=tx[idx_tex] mtex.texture_coords="UV" #mtex.map_colordiff = True #mtex.map_alpha = True #mtex.map_coloremission = True #mtex.map_density = True #mtex.mapping = 'FLAT' for grp_idx in pair[1]: groups[grp_idx][5].data.materials.append(matt[n]) n=n+1 print("=============Materials creation summary:=================") print("Created ",n," materials,") print("Loaded ",len(tx)-tex_load_fails," textures.") if not(orbiter_path_ok): print("WARNING! Orbiter path is wrong or not accessible, textures cannot be loaded!") print("Wrong path=",orbiterpath) if tex_load_fails>0: print("WARNING! ",tex_load_fails," of ",len(tx)," textures aren't loaded, possibly wrong file name(s)!")
86e35fd95f0c62cd139674c6ebaa7ac78d3174ba /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10993/86e35fd95f0c62cd139674c6ebaa7ac78d3174ba/io_import_msh.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 67, 20542, 87, 12, 4650, 16, 20542, 87, 16, 955, 1823, 16, 16640, 2165, 803, 4672, 468, 1883, 310, 9390, 19, 955, 594, 17265, 1172, 2932, 11092, 31824, 7923, 4834, 6017, 542, 33, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 752, 67, 20542, 87, 12, 4650, 16, 20542, 87, 16, 955, 1823, 16, 16640, 2165, 803, 4672, 468, 1883, 310, 9390, 19, 955, 594, 17265, 1172, 2932, 11092, 31824, 7923, 4834, 6017, 542, 33, ...
alphabet = _consensus_alphabet([rec.seq.alphabet for rec in sequences \
alphabet = _consensus_alphabet([rec.seq.alphabet \ for rec in sequences \
def to_alignment(sequences, alphabet=None, strict=True): """Returns a multiple sequence alignment (OBSOLETE). - sequences -An iterator that returns SeqRecord objects, or simply a list of SeqRecord objects. All the record sequences must be the same length. - alphabet - Optional alphabet. Stongly recommended. - strict - Optional, defaults to True. Should error checking be done? Using this function is now discouraged. Rather doing this: >>> from Bio import SeqIO >>> handle = open("Clustalw/protein.aln") >>> alignment = SeqIO.to_alignment(SeqIO.parse(handle, "clustal")) >>> handle.close() You are now encouraged to use Bio.AlignIO instead, e.g. >>> from Bio import AlignIO >>> handle = open("Clustalw/protein.aln") >>> alignment = AlignIO.read(handle, "clustal") >>> handle.close() """ #TODO - Move this functionality into the Alignment class instead? from Bio.Alphabet import generic_alphabet from Bio.Alphabet import _consensus_alphabet if alphabet is None: sequences = list(sequences) alphabet = _consensus_alphabet([rec.seq.alphabet for rec in sequences \ if rec.seq is not None]) if not (isinstance(alphabet, Alphabet) or isinstance(alphabet, AlphabetEncoder)): raise ValueError("Invalid alphabet") alignment_length = None alignment = Alignment(alphabet) for record in sequences: if strict: if alignment_length is None: alignment_length = len(record.seq) elif alignment_length != len(record.seq): raise ValueError("Sequences must all be the same length") assert isinstance(record.seq.alphabet, Alphabet) \ or isinstance(record.seq.alphabet, AlphabetEncoder), \ "Sequence does not have a valid alphabet" #TODO - Move this alphabet comparison code into the Alphabet module/class? #TODO - Is a normal alphabet "ungapped" by default, or does it just mean #undecided? if isinstance(record.seq.alphabet, Alphabet) \ and isinstance(alphabet, Alphabet): #Comparing two non-gapped alphabets if not isinstance(record.seq.alphabet, alphabet.__class__): raise ValueError("Incompatible sequence alphabet " \ + "%s for %s alignment" \ % (record.seq.alphabet, alphabet)) elif isinstance(record.seq.alphabet, AlphabetEncoder) \ and isinstance(alphabet, Alphabet): raise ValueError("Sequence has a gapped alphabet, alignment does not") elif isinstance(record.seq.alphabet, Alphabet) \ and isinstance(alphabet, Gapped): #Sequence isn't gapped, alignment is. if not isinstance(record.seq.alphabet, alphabet.alphabet.__class__): raise ValueError("Incompatible sequence alphabet " \ + "%s for %s alignment" \ % (record.seq.alphabet, alphabet)) else: #Comparing two gapped alphabets if not isinstance(record.seq.alphabet, alphabet.__class__): raise ValueError("Incompatible sequence alphabet " \ + "%s for %s alignment" \ % (record.seq.alphabet, alphabet)) if record.seq.alphabet.gap_char != alphabet.gap_char: raise ValueError("Sequence gap characters != alignment gap char") #ToDo, additional checks on the specified alignment... #Should we look at the alphabet.contains() method? if record.seq is None: raise TypeError("SeqRecord (id=%s) has None for its sequence." % record.id) #This is abusing the "private" records list, #we should really have a method like add_sequence #but which takes SeqRecord objects. See also Bug 1944 alignment._records.append(record) return alignment
a2855b4f95e94e4ef816422674d48a89ec6bbe18 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7167/a2855b4f95e94e4ef816422674d48a89ec6bbe18/__init__.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 358, 67, 14409, 12, 17190, 16, 10877, 33, 7036, 16, 5490, 33, 5510, 4672, 3536, 1356, 279, 3229, 3102, 8710, 261, 5704, 3584, 4311, 2934, 225, 300, 8463, 300, 979, 2775, 716, 1135, 14367...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 358, 67, 14409, 12, 17190, 16, 10877, 33, 7036, 16, 5490, 33, 5510, 4672, 3536, 1356, 279, 3229, 3102, 8710, 261, 5704, 3584, 4311, 2934, 225, 300, 8463, 300, 979, 2775, 716, 1135, 14367...
self.dbh.execute("update msn_users set session_id=%r where session_id=-1",m.session_id)
try: self.dbh.execute("update msn_users set session_id=%r where session_id=-1",m.session_id) except Exception: pass self.dbh.execute("delete from msn_users where session_id=-1")
def process_stream(self, stream, factories): ports = dissect.fix_ports("MSN") if stream.src_port in ports or stream.dest_port in ports: #Keep track of the streams we have combined so we don't process reverse again #self.processed is a ring buffer so it doesn't get too big if not self.processed.has_key(stream.con_id): #Haven't processed this one. #This returns the forward and reverse stream associated with the MSN port (uses port numbers in .pyflagrc). forward_stream, reverse_stream = self.stream_to_server(stream, "MSN") #We need both streams otherwise this won't work if not reverse_stream or not forward_stream: return
cf7fac3ff0c735f3f514d4f41c9e494e2787883c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5568/cf7fac3ff0c735f3f514d4f41c9e494e2787883c/MSN.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1207, 67, 3256, 12, 2890, 16, 1407, 16, 17456, 4672, 9048, 273, 1015, 5709, 18, 904, 67, 4363, 2932, 3537, 50, 7923, 225, 309, 1407, 18, 4816, 67, 655, 316, 9048, 578, 1407, 18, 10488,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1207, 67, 3256, 12, 2890, 16, 1407, 16, 17456, 4672, 9048, 273, 1015, 5709, 18, 904, 67, 4363, 2932, 3537, 50, 7923, 225, 309, 1407, 18, 4816, 67, 655, 316, 9048, 578, 1407, 18, 10488,...
include "meta.rnc"
def buildSchemaDrivers(): schemaDir = os.path.join(buildRoot, "syntax", "relaxng") legacyRnc = os.path.join(os.path.join(buildRoot, "validator", "schema", "legacy", "legacy.rnc")) buildSchemaDriverHtmlCore(schemaDir) buildSchemaDriverHtml5(schemaDir) buildSchemaDriverHtml5Aria(schemaDir) buildSchemaDriverHtml5AriaRdfa(schemaDir) buildSchemaDriverXhtmlCore(schemaDir) buildSchemaDriverXhtmlCorePlusWf2(schemaDir) buildSchemaDriverXhtml5html(schemaDir) buildSchemaDriverXhtml5xhtml(schemaDir) buildSchemaDriverXhtml5Aria(schemaDir) buildSchemaDriverXhtml5AriaRdfa(schemaDir) removeIfExists(os.path.join(schemaDir, "legacy.rnc")) shutil.copy(legacyRnc, schemaDir)
79071eb95d29311580797e9db095992d6fc2d2ee /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/13711/79071eb95d29311580797e9db095992d6fc2d2ee/build.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1361, 3078, 26310, 13332, 1963, 1621, 273, 1140, 18, 803, 18, 5701, 12, 3510, 2375, 16, 315, 14308, 3113, 315, 2878, 651, 3368, 7923, 8866, 54, 14202, 273, 1140, 18, 803, 18, 5701, 12, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1361, 3078, 26310, 13332, 1963, 1621, 273, 1140, 18, 803, 18, 5701, 12, 3510, 2375, 16, 315, 14308, 3113, 315, 2878, 651, 3368, 7923, 8866, 54, 14202, 273, 1140, 18, 803, 18, 5701, 12, ...
design_size = m.group (1) prefixes = [("cmtt", "Computer Modern Typewriter Text"), ("cmvtt", "Computer Modern Variable Width Typewriter Text"), ("cmss", "Computer Modern Sans Serif"), ("cm", "Computer Modern")]
design_size = string.atoi (m.group (1)) fontinfo['DesignSize'] = design_size prefixes = [("cmtt", "CMTypewriter"), ("cmvtt", "CMVariableWidthTypewriter"), ("cmss", "CMSans"), ("cm", "CM")]
def cm_guess_font_info (filename): """this function fills in sensible values for fontinfo.
dfdd47ac4957b0e437c434b79c7589ca35658d7b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11856/dfdd47ac4957b0e437c434b79c7589ca35658d7b/mftrace.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5003, 67, 20885, 67, 5776, 67, 1376, 261, 3459, 4672, 3536, 2211, 445, 21652, 316, 15390, 1523, 924, 364, 3512, 1376, 18, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 5003, 67, 20885, 67, 5776, 67, 1376, 261, 3459, 4672, 3536, 2211, 445, 21652, 316, 15390, 1523, 924, 364, 3512, 1376, 18, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -...
<dt>Enable for:</dt> <dd>To make the rank method visible in the search interface for a collection, choose the wanted collection and press the "Enable" button.</dd> <dt>Disable for:</dt> <dd>To hide the rank method from the search interface for a collection, choose the wanted collection and press the "Disable" button.</dd> <dt>'%s' is currently enabled for the following collections:</dt>
<dt>The rank method is currently enabled for these collections:</dt>
def perform_modifycollection(rnkID='', ln=cdslang, func='', colID='', confirm=0): """Modify which collections the rank method is visible to""" output = "" subtitle = "" if rnkID: rnkNAME = get_current_name(rnkID, ln, get_rnk_nametypes()[0][0], "rnkMETHOD")[0][1] if func in ["0", 0] and confirm in ["1", 1]: finresult = attach_col_rnk(rnkID, colID) elif func in ["1", 1] and confirm in ["1", 1]: finresult = detach_col_rnk(rnkID, colID) if colID: colNAME = get_current_name(colID, ln, get_col_nametypes()[0][0], "collection")[0][1] subtitle = """Step 1 - Select collection to enable/disable rank method '%s' for""" % rnkNAME output = """ <dl> <dt>Enable for:</dt> <dd>To make the rank method visible in the search interface for a collection, choose the wanted collection and press the "Enable" button.</dd> <dt>Disable for:</dt> <dd>To hide the rank method from the search interface for a collection, choose the wanted collection and press the "Disable" button.</dd> <dt>'%s' is currently enabled for the following collections:</dt> <dd> """ % rnkNAME col_list = get_rnk_col(rnkID, ln) if not col_list: output += """No collections""" else: for (id, name) in col_list: output += """%s, """ % name output += """</dd> </dl> """ col_list = get_current_name('', ln, get_col_nametypes()[0][0], "collection") col_rnk = dict(get_rnk_col(rnkID)) col_list = filter(lambda x: not col_rnk.has_key(x[0]), col_list) if col_list: text = """ <span class="adminlabel">Enable for:</span> <select name="colID" class="admin_w200"> <option value="">- select collection -</option> """ for (id, name) in col_list: text += """<option value="%s" %s>%s</option>""" % (id, (func in ["0", 0] and confirm in ["0", 0] and colID and int(colID) == int(id)) and 'selected="selected"' or '' , name) text += """</select>""" output += createhiddenform(action="modifycollection", text=text, button="Enable", rnkID=rnkID, ln=ln, func=0, confirm=0) if confirm in ["0", 0] and func in ["0", 0] and colID: subtitle = "Step 2 - Confirm to enable rank method for the chosen collection" text = "<b><p>Please confirm to enable rank method '%s' for the collection '%s'</p></b>" % (rnkNAME, colNAME) output += createhiddenform(action="modifycollection", text=text, button="Confirm", rnkID=rnkID, ln=ln, colID=colID, func=0, confirm=1) elif confirm in ["1", 1] and func in ["0", 0] and colID: subtitle = "Step 3 - Result" if finresult: output += """<b><span class="info">Rank method '%s' enabled for collection '%s'</span></b>""" % (rnkNAME, colNAME) else: output += """<b><span class="info">Rank method '%s' could not be enabled for collection '%s'</span></b>""" % (rnkNAME[0], colNAME) col_list = get_rnk_col(rnkID, ln) if col_list: text = """ <span class="adminlabel">Disable for:</span> <select name="colID" class="admin_w200"> <option value="">- select collection -</option> """ for (id, name) in col_list: text += """<option value="%s" %s>%s</option>""" % (id, (func in ["1", 1] and confirm in ["0", 0] and colID and int(colID) == int(id)) and 'selected="selected"' or '' , name) text += """</select>""" output += createhiddenform(action="modifycollection", text=text, button="Disable", rnkID=rnkID, ln=ln, func=1, confirm=0) if confirm in ["0", 0] and func in ["1", 1] and colID: subtitle = "Step 2 - Confirm to disable rank method for collection" text = "<b><p>Please confirm to disable rank method '%s' for collection '%s'</p></b>" % (rnkNAME, colNAME) output += createhiddenform(action="modifycollection", text=text, button="Confirm", rnkID=rnkID, ln=ln, colID=colID, func=1, confirm=1) elif confirm in ["1", 1] and func in ["1", 1] and colID: subtitle = "Step 3 - Result" if finresult: output += """<b><span class="info">Rank method '%s' disabled for collection '%s'</span></b>""" % (rnkNAME, colNAME) else: output += """<b><span class="info">Rank method '%s' could not be disabled for collection '%s'</span></b>""" % (rnkNAME, colNAME) try: body = [output, extra] except NameError: body = [output] return addadminbox(subtitle + """&nbsp;&nbsp&nbsp;<small>[<a href="%s/admin/bibrank/guide.html#mc">?</a>]</small>""" % weburl, body)
a6a2a3f836d914026219bf3610ec841b23ae03ce /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12027/a6a2a3f836d914026219bf3610ec841b23ae03ce/bibrankadminlib.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3073, 67, 17042, 5548, 12, 27639, 79, 734, 2218, 2187, 7211, 33, 4315, 2069, 539, 16, 1326, 2218, 2187, 645, 734, 2218, 2187, 6932, 33, 20, 4672, 3536, 11047, 1492, 6980, 326, 6171, 707,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 3073, 67, 17042, 5548, 12, 27639, 79, 734, 2218, 2187, 7211, 33, 4315, 2069, 539, 16, 1326, 2218, 2187, 645, 734, 2218, 2187, 6932, 33, 20, 4672, 3536, 11047, 1492, 6980, 326, 6171, 707,...
for f,r in self.replace_paths:
for f, r in self.replace_paths:
def replace_paths_in_code(self, co): new_filename = original_filename = os.path.normpath(co.co_filename) for f,r in self.replace_paths: if original_filename.startswith(f): new_filename = r+original_filename[len(f):] break
e29310a2b3e50958b00c0e92b142992599cf7ef1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/e29310a2b3e50958b00c0e92b142992599cf7ef1/modulefinder.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1453, 67, 4481, 67, 267, 67, 710, 12, 2890, 16, 1825, 4672, 394, 67, 3459, 273, 2282, 67, 3459, 273, 1140, 18, 803, 18, 7959, 803, 12, 2894, 18, 2894, 67, 3459, 13, 364, 284, 16, 4...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1453, 67, 4481, 67, 267, 67, 710, 12, 2890, 16, 1825, 4672, 394, 67, 3459, 273, 2282, 67, 3459, 273, 1140, 18, 803, 18, 7959, 803, 12, 2894, 18, 2894, 67, 3459, 13, 364, 284, 16, 4...
frameCustom.pack(side=LEFT,padx=5,pady=10,expand=TRUE,fill=BOTH) frameTheme.pack(side=LEFT,padx=5,pady=10,fill=Y)
frameCustom.pack(side=LEFT,padx=5,pady=5,expand=TRUE,fill=BOTH) frameTheme.pack(side=LEFT,padx=5,pady=5,fill=Y)
def CreatePageHighlight(self): self.builtinTheme=StringVar(self) self.customTheme=StringVar(self) self.fgHilite=BooleanVar(self) self.colour=StringVar(self) self.fontName=StringVar(self) self.themeIsBuiltin=BooleanVar(self) self.highlightTarget=StringVar(self) ##widget creation #body frame frame=self.tabPages.pages['Highlighting']['page'] #body section frames frameCustom=Frame(frame,borderwidth=2,relief=GROOVE) frameTheme=Frame(frame,borderwidth=2,relief=GROOVE) #frameCustom self.textHighlightSample=Text(frameCustom,relief=SOLID,borderwidth=1, font=('courier',12,''),cursor='hand2',width=21,height=10, takefocus=FALSE,highlightthickness=0,wrap=NONE) text=self.textHighlightSample text.bind('<Double-Button-1>',lambda e: 'break') text.bind('<B1-Motion>',lambda e: 'break') textAndTags=(('#you can click here','comment'),('\n','normal'), ('#to choose items','comment'),('\n','normal'),('def','keyword'), (' ','normal'),('func','definition'),('(param):','normal'), ('\n ','normal'),('"""string"""','string'),('\n var0 = ','normal'), ("'string'",'string'),('\n var1 = ','normal'),("'selected'",'hilite'), ('\n var2 = ','normal'),("'found'",'hit'), ('\n var3 = ','normal'),('list', 'builtin'), ('(','normal'), ('None', 'builtin'),(')\n\n','normal'), (' error ','error'),(' ','normal'),('cursor |','cursor'), ('\n ','normal'),('shell','console'),(' ','normal'),('stdout','stdout'), (' ','normal'),('stderr','stderr'),('\n','normal')) for txTa in textAndTags: text.insert(END,txTa[0],txTa[1]) for element in self.themeElements.keys(): text.tag_bind(self.themeElements[element][0],'<ButtonPress-1>', lambda event,elem=element: event.widget.winfo_toplevel() .highlightTarget.set(elem)) text.config(state=DISABLED) self.frameColourSet=Frame(frameCustom,relief=SOLID,borderwidth=1) frameFgBg=Frame(frameCustom) labelCustomTitle=Label(frameCustom,text='Set Custom Highlighting') buttonSetColour=Button(self.frameColourSet,text='Choose Colour for :', command=self.GetColour,highlightthickness=0) self.optMenuHighlightTarget=DynOptionMenu(self.frameColourSet, self.highlightTarget,None,highlightthickness=0)#,command=self.SetHighlightTargetBinding self.radioFg=Radiobutton(frameFgBg,variable=self.fgHilite, value=1,text='Foreground',command=self.SetColourSampleBinding) self.radioBg=Radiobutton(frameFgBg,variable=self.fgHilite, value=0,text='Background',command=self.SetColourSampleBinding) self.fgHilite.set(1) buttonSaveCustomTheme=Button(frameCustom, text='Save as New Custom Theme',command=self.SaveAsNewTheme) #frameTheme labelThemeTitle=Label(frameTheme,text='Select a Highlighting Theme') labelTypeTitle=Label(frameTheme,text='Select : ') self.radioThemeBuiltin=Radiobutton(frameTheme,variable=self.themeIsBuiltin, value=1,command=self.SetThemeType,text='a Built-in Theme') self.radioThemeCustom=Radiobutton(frameTheme,variable=self.themeIsBuiltin, value=0,command=self.SetThemeType,text='a Custom Theme') self.optMenuThemeBuiltin=DynOptionMenu(frameTheme, self.builtinTheme,None,command=None) self.optMenuThemeCustom=DynOptionMenu(frameTheme, self.customTheme,None,command=None) self.buttonDeleteCustomTheme=Button(frameTheme,text='Delete Custom Theme', command=self.DeleteCustomTheme) ##widget packing #body frameCustom.pack(side=LEFT,padx=5,pady=10,expand=TRUE,fill=BOTH) frameTheme.pack(side=LEFT,padx=5,pady=10,fill=Y) #frameCustom labelCustomTitle.pack(side=TOP,anchor=W,padx=5,pady=5) self.frameColourSet.pack(side=TOP,padx=5,pady=5,expand=TRUE,fill=X) frameFgBg.pack(side=TOP,padx=5,pady=0) self.textHighlightSample.pack(side=TOP,padx=5,pady=5,expand=TRUE, fill=BOTH) buttonSetColour.pack(side=TOP,expand=TRUE,fill=X,padx=8,pady=4) self.optMenuHighlightTarget.pack(side=TOP,expand=TRUE,fill=X,padx=8,pady=3) self.radioFg.pack(side=LEFT,anchor=E) self.radioBg.pack(side=RIGHT,anchor=W) buttonSaveCustomTheme.pack(side=BOTTOM,fill=X,padx=5,pady=5) #frameTheme labelThemeTitle.pack(side=TOP,anchor=W,padx=5,pady=5) labelTypeTitle.pack(side=TOP,anchor=W,padx=5,pady=5) self.radioThemeBuiltin.pack(side=TOP,anchor=W,padx=5) self.radioThemeCustom.pack(side=TOP,anchor=W,padx=5,pady=2) self.optMenuThemeBuiltin.pack(side=TOP,fill=X,padx=5,pady=5) self.optMenuThemeCustom.pack(side=TOP,fill=X,anchor=W,padx=5,pady=5) self.buttonDeleteCustomTheme.pack(side=TOP,fill=X,padx=5,pady=5) return frame
cb3fac2e967e9ef77ede3f9945cfed8702f7e0a5 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/3187/cb3fac2e967e9ef77ede3f9945cfed8702f7e0a5/configDialog.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1788, 1964, 16205, 12, 2890, 4672, 365, 18, 24553, 8335, 33, 780, 1537, 12, 2890, 13, 365, 18, 3662, 8335, 33, 780, 1537, 12, 2890, 13, 365, 18, 2137, 44, 330, 1137, 33, 5507, 1537, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1788, 1964, 16205, 12, 2890, 4672, 365, 18, 24553, 8335, 33, 780, 1537, 12, 2890, 13, 365, 18, 3662, 8335, 33, 780, 1537, 12, 2890, 13, 365, 18, 2137, 44, 330, 1137, 33, 5507, 1537, ...
self._completeDelayedCallback = FrameDelayedCallback(
self._completeDelayedCallback = FrameDelayedCall(
def sendEvent(): messenger.send(self.getAllInterestsCompleteEvent()) for callback in self._allInterestsCompleteCallbacks: callback() self._allInterestsCompleteCallbacks = []
0ad3bac423d39b998827d489a97d947c4e85743a /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8543/0ad3bac423d39b998827d489a97d947c4e85743a/DoInterestManager.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1366, 1133, 13332, 31086, 18, 4661, 12, 2890, 18, 588, 1595, 2465, 25563, 6322, 1133, 10756, 364, 1348, 316, 365, 6315, 454, 2465, 25563, 6322, 10617, 30, 1348, 1435, 365, 6315, 454, 2465,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1366, 1133, 13332, 31086, 18, 4661, 12, 2890, 18, 588, 1595, 2465, 25563, 6322, 1133, 10756, 364, 1348, 316, 365, 6315, 454, 2465, 25563, 6322, 10617, 30, 1348, 1435, 365, 6315, 454, 2465,...
BonjourFile(file, 'resolution').close()
BonjourFile(file, 'resolution').close()
def _resolve_cb(self, file, flags, interface_index, error_code, fullname, host_target, port, txtrecord): notification_center = NotificationCenter() settings = SIPSimpleSettings() if error_code == bonjour.kDNSServiceErr_NoError: txt = bonjour.TXTRecord.parse(txtrecord) contact = txt['contact'].strip('<>') if 'contact' in txt else None if contact: display_name = txt['name'].decode('utf-8') if 'name' in txt else None host = re.match(r'^(.*?)(\.local)?\.?$', host_target).group(1) uri = FrozenSIPURI.parse(contact) transport = uri.parameters.get('transport', 'udp') supported_transport = transport in settings.sip.transport_list and (transport!='tls' or self.account.tls.certificate is not None) if supported_transport and uri != self.account.contact[transport] and uri not in self._neighbours: self._neighbours.add(uri) notification_center.post_notification('BonjourAccountDidAddNeighbour', sender=self.account, data=TimestampedNotificationData(display_name=display_name, host=host, uri=uri)) BonjourFile(file, 'resolution').close()
e155e163c29dab32cacafb785a44e3adb5d577b4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3449/e155e163c29dab32cacafb785a44e3adb5d577b4/account.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 10828, 67, 7358, 12, 2890, 16, 585, 16, 2943, 16, 1560, 67, 1615, 16, 555, 67, 710, 16, 13321, 16, 1479, 67, 3299, 16, 1756, 16, 6463, 3366, 4672, 3851, 67, 5693, 273, 8050, 844...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 10828, 67, 7358, 12, 2890, 16, 585, 16, 2943, 16, 1560, 67, 1615, 16, 555, 67, 710, 16, 13321, 16, 1479, 67, 3299, 16, 1756, 16, 6463, 3366, 4672, 3851, 67, 5693, 273, 8050, 844...
self.assertEquals(map(lambda x:x.text, fragList), [u'Hello ',u'\xc2\xa9',u' copyright'])
self.assertEquals(map(lambda x:x.text, fragList), [u'Hello ',u'\xa9',u' copyright'])
def testEntityUnicode(self): "Numeric entities should be unescaped by parser" txt = u"Hello &#169; copyright" fragList = ParaParser().parse(txt, self.style)[1] self.assertEquals(map(lambda x:x.text, fragList), [u'Hello ',u'\xc2\xa9',u' copyright'])
55d7869f3b6dcd4bb6bef80704fcd9716369ea6b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3878/55d7869f3b6dcd4bb6bef80704fcd9716369ea6b/test_platypus_paraparser.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 1943, 16532, 12, 2890, 4672, 315, 9902, 5140, 1410, 506, 27869, 635, 2082, 6, 6463, 273, 582, 6, 18601, 6523, 26035, 31, 22245, 6, 10111, 682, 273, 2280, 69, 2678, 7675, 2670, 12, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 1943, 16532, 12, 2890, 4672, 315, 9902, 5140, 1410, 506, 27869, 635, 2082, 6, 6463, 273, 582, 6, 18601, 6523, 26035, 31, 22245, 6, 10111, 682, 273, 2280, 69, 2678, 7675, 2670, 12, ...
browsed_words = browse_in_bibxxx(p, f, rg) while not browsed_words:
browsed_phrases = get_nearest_terms_in_bibxxx(p, f, rg, 1) while not browsed_phrases:
def browse_pattern(req, colls, p, f, rg): """Browse either biliographic phrases or words indexes, and display it.""" ## do we search in words indexes? if not f: return browse_in_bibwords(req, p, f) ## prepare collection urlargument for later printing: p_orig = p urlarg_colls = "" for coll in colls: urlarg_colls += "&c=%s" % urllib.quote(coll) ## okay, "real browse" follows: browsed_words = browse_in_bibxxx(p, f, rg) while not browsed_words: # try again and again with shorter and shorter pattern: try: p = p[:-1] browsed_words = browse_in_bibxxx(p, f, rg) except: # probably there are no hits at all: req.write("<p>No values found.") return ## try to check hits in these particular collection selection: browsed_words_in_colls = [] for word,nbhits in browsed_words: word_hitlist = HitList() word_hitlists = search_pattern("", word, f, colls, 'e') for coll in colls: word_hitlist.union(word_hitlists[coll]) word_hitlist.calculate_nbhits() if word_hitlist._nbhits > 0: # okay, this word has some hits in colls, so add it: browsed_words_in_colls.append([word, word_hitlist._nbhits]) ## were there hits in collections? if browsed_words_in_colls == []: if browsed_words != []: print_warning(req, """<p>No match close to <em>%s</em> found in given collections. Please try different term.<p>Displaying matches in any collection...""" % p_orig, "") browsed_words_in_colls = browsed_words ## display results now: out = """<table class="searchresultsbox"> <thead> <tr> <th class="searchresultsboxheader" align="left"> hits </th> <th class="searchresultsboxheader" width="15"> &nbsp; </th> <th class="searchresultsboxheader" align="left"> %s </th> </tr> </thead> <tbody>""" % f if len(browsed_words_in_colls) == 1: # one hit only found: word, nbhits = browsed_words_in_colls[0][0], browsed_words_in_colls[0][1] out += """<tr> <td class="searchresultsboxbody" align="right"> %s </td> <td class="searchresultsboxbody" width="15"> &nbsp; </td> <td class="searchresultsboxbody" align="left"> <a href="%s/search.py?p=%%22%s%%22&f=%s%s">%s</a> </td> </tr>""" % (nbhits, weburl, urllib.quote(word), urllib.quote(f), urlarg_colls, word) elif len(browsed_words_in_colls) > 1: # first display what was found but the last one: for word, nbhits in browsed_words_in_colls[:-1]: out += """<tr> <td class="searchresultsboxbody" align="right"> %s </td> <td class="searchresultsboxbody" width="15"> &nbsp; </td> <td class="searchresultsboxbody" align="left"> <a href="%s/search.py?p=%%22%s%%22&f=%s%s">%s</a> </td> </tr>""" % (nbhits, weburl, urllib.quote(word), urllib.quote(f), urlarg_colls, word) # now display last hit as "next term": word, nbhits = browsed_words_in_colls[-1] out += """<tr><td colspan="2" class="normal"> &nbsp; </td> <td class="normal"> <img src="%s/img/sn.gif" alt="" border="0"> next %s: <a href="%s/search.py?search=Browse&p=%s&f=%s%s">%s</a> </td> </tr>""" % (weburl, f, weburl, urllib.quote(word), urllib.quote(f), urlarg_colls, word) out += """</tbody> </table>""" req.write(out) return
535ed6a8d836745c0ddd7f1016fbe145d7c7bebd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2139/535ed6a8d836745c0ddd7f1016fbe145d7c7bebd/search_engine.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 21670, 67, 4951, 12, 3658, 16, 645, 3251, 16, 293, 16, 284, 16, 14524, 4672, 3536, 27304, 3344, 324, 15700, 16983, 31413, 578, 4511, 5596, 16, 471, 2562, 518, 12123, 7541, 741, 732, 1623...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 21670, 67, 4951, 12, 3658, 16, 645, 3251, 16, 293, 16, 284, 16, 14524, 4672, 3536, 27304, 3344, 324, 15700, 16983, 31413, 578, 4511, 5596, 16, 471, 2562, 518, 12123, 7541, 741, 732, 1623...
t = self.font_switcher.fontifyText(xmlescape(''.join(txt).strip()).encode('utf-8'))
t = self.font_switcher.fontifyText(xmlescape(''.join(txt).strip())).encode('utf-8')
def writeLink(self,obj): """ Link nodes are intra wiki links """
5b13d94f47a43c6f34f2d8bd5ccb703e70557fd2 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12390/5b13d94f47a43c6f34f2d8bd5ccb703e70557fd2/rlwriter.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1045, 2098, 12, 2890, 16, 2603, 4672, 3536, 4048, 2199, 854, 316, 2033, 9050, 4716, 3536, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1045, 2098, 12, 2890, 16, 2603, 4672, 3536, 4048, 2199, 854, 316, 2033, 9050, 4716, 3536, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100,...
if s3path == self._prefix:
if s3path == "/" or s3path == self._prefix:
def isdir(self,path): """Check whether a path exists and is a directory.""" s3path = self._s3path(path) + self._separator # Root is always a directory if s3path == self._prefix: return True # Use a list request so that we return true if there are any files # in that directory. This avoids requiring a special file for the # the directory itself, which other tools may not create. ks = self._s3bukt.list(prefix=s3path,delimiter=self._separator) for k in ks: return True return False
a97fab46dd4f11ad2fd10368707968cd764a18d9 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5579/a97fab46dd4f11ad2fd10368707968cd764a18d9/s3fs.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 9783, 12, 2890, 16, 803, 4672, 3536, 1564, 2856, 279, 589, 1704, 471, 353, 279, 1867, 12123, 272, 23, 803, 273, 365, 6315, 87, 23, 803, 12, 803, 13, 397, 365, 6315, 11287, 468, 7450, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 9783, 12, 2890, 16, 803, 4672, 3536, 1564, 2856, 279, 589, 1704, 471, 353, 279, 1867, 12123, 272, 23, 803, 273, 365, 6315, 87, 23, 803, 12, 803, 13, 397, 365, 6315, 11287, 468, 7450, ...
key = string.upper(key) self.data[key] = item
self.data[key.upper()] = item
def __setitem__(self, key, item): putenv(key, item) key = string.upper(key) self.data[key] = item
1e3c380cd4e6e569be13bdaa54da1f9acbd3c0d0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/1e3c380cd4e6e569be13bdaa54da1f9acbd3c0d0/os.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 542, 1726, 972, 12, 2890, 16, 498, 16, 761, 4672, 1378, 3074, 12, 856, 16, 761, 13, 498, 273, 533, 18, 5797, 12, 856, 13, 365, 18, 892, 63, 856, 65, 273, 761, 2, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 542, 1726, 972, 12, 2890, 16, 498, 16, 761, 4672, 1378, 3074, 12, 856, 16, 761, 13, 498, 273, 533, 18, 5797, 12, 856, 13, 365, 18, 892, 63, 856, 65, 273, 761, 2, -100, -100, ...
debug('cleaning up worker %d' % w.pid)
debug('cleaning up worker %s' % w.pid)
def _terminate_pool(cls, taskqueue, inqueue, outqueue, pool, worker_handler, task_handler, result_handler, cache): # this is guaranteed to only be called once debug('finalizing pool')
89db55254387b4cae708c5477e87a6a3b76344aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8125/89db55254387b4cae708c5477e87a6a3b76344aa/pool.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 30017, 67, 6011, 12, 6429, 16, 1562, 4000, 16, 316, 4000, 16, 596, 4000, 16, 2845, 16, 4322, 67, 4176, 16, 1562, 67, 4176, 16, 563, 67, 4176, 16, 1247, 4672, 468, 333, 353, 1540...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 30017, 67, 6011, 12, 6429, 16, 1562, 4000, 16, 316, 4000, 16, 596, 4000, 16, 2845, 16, 4322, 67, 4176, 16, 1562, 67, 4176, 16, 563, 67, 4176, 16, 1247, 4672, 468, 333, 353, 1540...
ttf = _py_getFont(ttfn) t1f = _py_getFont(t1fn)
ttf = getFont(ttfn) t1f = getFont(t1fn)
def test_instanceStringWidth(self): from reportlab.pdfbase.pdfmetrics import registerFont, _py_getFont from reportlab.pdfbase.ttfonts import TTFont ttfn = 'Luxi-Serif' t1fn = 'Times-Roman' registerFont(TTFont(ttfn, "luxiserif.ttf")) ttf = _py_getFont(ttfn) t1f = _py_getFont(t1fn) testCp1252 = 'copyright %s trademark %s registered %s ReportLab! Ol%s!' % (chr(169), chr(153),chr(174), chr(0xe9)) enc='cp1252' senc = 'utf8' ts = 'ABCDEF\xce\x91\xce\xb2G' utext = 'ABCDEF\xce\x91\xce\xb2G'.decode(senc) fontSize = 12. def tfunc(f,ts,fontSize,enc): w1 = f.stringWidth(ts,fontSize,enc) w2 = f._py_stringWidth(ts,fontSize,enc) assert abs(w1-w2)<1e-10,"f(%r).stringWidthU(%r,%s,%r)-->%r != f._py_stringWidth(...)-->%r" % (f,ts,fontSize,enc,w1,w2) tfunc(t1f,testCp1252,fontSize,enc) tfunc(t1f,ts,fontSize,senc) tfunc(t1f,utext,fontSize,senc) tfunc(ttf,ts,fontSize,senc) tfunc(ttf,testCp1252,fontSize,enc) tfunc(ttf,utext,fontSize,senc)
12e8e9cc55671e5cc6f4a8cba83406f53d6c3513 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3878/12e8e9cc55671e5cc6f4a8cba83406f53d6c3513/test_rl_accel.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 1336, 780, 2384, 12, 2890, 4672, 628, 2605, 7411, 18, 7699, 1969, 18, 7699, 9597, 1930, 1744, 5711, 16, 389, 2074, 67, 588, 5711, 628, 2605, 7411, 18, 7699, 1969, 18, 748, 22...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1842, 67, 1336, 780, 2384, 12, 2890, 4672, 628, 2605, 7411, 18, 7699, 1969, 18, 7699, 9597, 1930, 1744, 5711, 16, 389, 2074, 67, 588, 5711, 628, 2605, 7411, 18, 7699, 1969, 18, 748, 22...
noshell = 0
noshell = 0
def main(self, argv, noshell): cmd = None edit = 0 debug = 0 startup = 0 try: opts, args = getopt.getopt(argv, "c:deist:") except getopt.error, msg: sys.stderr.write("Error: %s\n" % str(msg)) sys.stderr.write(usage_msg) sys.exit(2) for o, a in opts: noshell = 0 if o == '-c': cmd = a if o == '-d': debug = 1 if o == '-e': edit = 1 if o == '-s': startup = 1 if o == '-t': PyShell.shell_title = a if noshell: edit=1 for i in range(len(sys.path)): sys.path[i] = os.path.abspath(sys.path[i]) pathx = [] if edit: for filename in args: pathx.append(os.path.dirname(filename)) elif args and args[0] != "-": pathx.append(os.path.dirname(args[0])) else: pathx.append(os.curdir) for dir in pathx: dir = os.path.abspath(dir) if not dir in sys.path: sys.path.insert(0, dir)
e1ea5a6a5cba72909c5bd271708c70ad1391d343 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/e1ea5a6a5cba72909c5bd271708c70ad1391d343/PyShell.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 12, 2890, 16, 5261, 16, 290, 17636, 1165, 4672, 1797, 273, 599, 3874, 273, 374, 1198, 273, 374, 11850, 273, 374, 225, 775, 30, 1500, 16, 833, 273, 336, 3838, 18, 588, 3838, 12, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 12, 2890, 16, 5261, 16, 290, 17636, 1165, 4672, 1797, 273, 599, 3874, 273, 374, 1198, 273, 374, 11850, 273, 374, 225, 775, 30, 1500, 16, 833, 273, 336, 3838, 18, 588, 3838, 12, ...
if len(geometries) > 0:
if len(geometries) > 1:
def __call__(self, f): xyz_reader = XYZReader(f) molecule = xyz_reader.get_first_molecule()
f5e1005bc70646586504f98cb0f3c655ceed081b /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/11052/f5e1005bc70646586504f98cb0f3c655ceed081b/xyz.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 1991, 972, 12, 2890, 16, 284, 4672, 14779, 67, 10530, 273, 30742, 2514, 12, 74, 13, 13661, 273, 14779, 67, 10530, 18, 588, 67, 3645, 67, 81, 10545, 1435, 2, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 1991, 972, 12, 2890, 16, 284, 4672, 14779, 67, 10530, 273, 30742, 2514, 12, 74, 13, 13661, 273, 14779, 67, 10530, 18, 588, 67, 3645, 67, 81, 10545, 1435, 2, -100, -100, -100, -10...
if subDirsRemoved and filesRemoved:
if subDirsRemoved and allFilesRemoved:
def __removeDir(self,directory): """ Black magic to recursively remove the directory and sub dirs. Repeatedly calls itself to delete recursively. """ filesRemoved = 0 sizeRemoved = 0 res = self.listDirectory(directory) if not res['OK']: return S_ERROR("Failed to list directory") if not res['Value']['Successful'].has_key(directory): return S_ERROR("Failed to list directory") filesRemoved = False surlsDict = res['Value']['Successful'][directoryPath]['Files'] subDirsDict = res['Value']['Successful'][directoryPath]['SubDirs'] res = storage.removeFile(surlsDict.keys()) if res['OK']: for removedSurl in res['Value']['Successful'].keys(): filesRemoved += 1 sizeRemoved += files[removedSurl]['Size'] if len(res['Value']['Failed'].keys()) == 0: filesRemoved = True # Remove the sub directories found subDirsRemoved = True for subDir in subDirsDict.keys(): res = self.__removeDir(subDir) if not res['OK']: subDirsRemoved = False if not res['Value']['AllRemoved']: subDirsRemoved = False filesRemoved += res['Value']['Files'] sizeRemoved += res['Value']['Size'] if subDirsRemoved and filesRemoved: allRemoved = True else: allRemoved = False resDict = {'AllRemoved':allRemoved,'Files':filesRemoved,'Size':sizeRemoved} return S_OK(resDict)
0c142939ddc410eebdc2cef5627d4332317e31b1 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12864/0c142939ddc410eebdc2cef5627d4332317e31b1/SRM2Storage.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 4479, 1621, 12, 2890, 16, 5149, 4672, 3536, 22467, 8146, 358, 8536, 1206, 326, 1867, 471, 720, 7717, 18, 868, 9061, 715, 4097, 6174, 358, 1430, 8536, 18, 3536, 1390, 10026, 273, 37...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 4479, 1621, 12, 2890, 16, 5149, 4672, 3536, 22467, 8146, 358, 8536, 1206, 326, 1867, 471, 720, 7717, 18, 868, 9061, 715, 4097, 6174, 358, 1430, 8536, 18, 3536, 1390, 10026, 273, 37...
for i, row in enumerate(self.model):
rows = self.get_all_file_rows() for i, row in enumerate(rows):
def get_next_bean_by_UUID(self, repeat_all=False): UUID = self.active_UUID for i, row in enumerate(self.model): if row[self.is_file[0]] and row[self.UUID[0]] == UUID: if i + 1 < len(self.model): next_row = self.model[i + 1] if next_row[self.is_file[0]]: return self.get_bean_from_row(next_row) if repeat_all: for row in self.model: if not row[self.is_file[0]]: continue return self.get_bean_from_row(row) else: return None
a9e8da818dbcefbfc3c0172d631c37b3e9ee048a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14563/a9e8da818dbcefbfc3c0172d631c37b3e9ee048a/common_tree.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 4285, 67, 14496, 67, 1637, 67, 5562, 12, 2890, 16, 7666, 67, 454, 33, 8381, 4672, 5866, 273, 365, 18, 3535, 67, 5562, 225, 2595, 273, 365, 18, 588, 67, 454, 67, 768, 67, 3...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 67, 4285, 67, 14496, 67, 1637, 67, 5562, 12, 2890, 16, 7666, 67, 454, 33, 8381, 4672, 5866, 273, 365, 18, 3535, 67, 5562, 225, 2595, 273, 365, 18, 588, 67, 454, 67, 768, 67, 3...
action = True featured = wikipedia.translate(mysite, featured_articles) ref = wikipedia.Page(wikipedia.getSite(), featured) gen = pagegenerators.ReferringPageGenerator(ref) generator = pagegenerators.NamespaceFilterPageGenerator(gen, [0]) for page in generator: workon(page)
featured = True
def workon(page): try: text = page.get() except wikipedia.IsRedirectPage: return wikipedia.output(u"\n\n>>> \03{lightpurple}%s\03{default} <<<" % page.title()) links = page.linkedPages() wikipedia.getall(mysite,links) for page2 in links: try: target = page2.getRedirectTarget() except (wikipedia.Error,wikipedia.SectionError): continue text = treat(text, page2, target) if text != page.get(): comment = wikipedia.translate(mysite, msg) page.put(text, comment)
89a93e7c410791c4b730c6fda9f2d91608b2c69d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/4404/89a93e7c410791c4b730c6fda9f2d91608b2c69d/fixing_redirects.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1440, 265, 12, 2433, 4672, 775, 30, 977, 273, 1363, 18, 588, 1435, 1335, 21137, 18, 2520, 5961, 1964, 30, 327, 21137, 18, 2844, 12, 89, 12691, 82, 64, 82, 23012, 521, 4630, 95, 5099, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1440, 265, 12, 2433, 4672, 775, 30, 977, 273, 1363, 18, 588, 1435, 1335, 21137, 18, 2520, 5961, 1964, 30, 327, 21137, 18, 2844, 12, 89, 12691, 82, 64, 82, 23012, 521, 4630, 95, 5099, ...
return Hammie(storage.open_storage((filename, mode), useDB)
return Hammie(storage.open_storage((filename, mode), useDB))
def open(filename, useDB=True, mode='r'): """Open a file, returning a Hammie instance. If usedb is False, open as a pickle instead of a DBDict. mode is used as the flag to open DBDict objects. 'c' for read-write (create if needed), 'r' for read-only, 'w' for read-write. """ return Hammie(storage.open_storage((filename, mode), useDB)
b1bba2e09369c91c9ae0bfe08540efc9c68536be /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9857/b1bba2e09369c91c9ae0bfe08540efc9c68536be/hammie.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1696, 12, 3459, 16, 999, 2290, 33, 5510, 16, 1965, 2218, 86, 11, 4672, 3536, 3678, 279, 585, 16, 5785, 279, 670, 301, 81, 1385, 791, 18, 225, 971, 1399, 70, 353, 1083, 16, 1696, 487,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1696, 12, 3459, 16, 999, 2290, 33, 5510, 16, 1965, 2218, 86, 11, 4672, 3536, 3678, 279, 585, 16, 5785, 279, 670, 301, 81, 1385, 791, 18, 225, 971, 1399, 70, 353, 1083, 16, 1696, 487,...
def update_idx(): exe.add((cfg.bin.git, 'update-index', '--refresh'))
def update_idx(): exe.add((cfg.bin.git, 'update-index', '--refresh'), block=True)
def update_idx(): exe.add((cfg.bin.git, 'update-index', '--refresh'))
1eb490ecf022d10b9baff3f98342b64ecf7ff733 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11594/1eb490ecf022d10b9baff3f98342b64ecf7ff733/git.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 67, 3465, 13332, 15073, 18, 1289, 12443, 7066, 18, 4757, 18, 6845, 16, 296, 2725, 17, 1615, 2187, 3534, 9144, 19899, 1203, 33, 5510, 13, 282, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 67, 3465, 13332, 15073, 18, 1289, 12443, 7066, 18, 4757, 18, 6845, 16, 296, 2725, 17, 1615, 2187, 3534, 9144, 19899, 1203, 33, 5510, 13, 282, 2, -100, -100, -100, -100, -100, -100,...
for domain in domains:
for domain in self._domains[level]:
def _log(self, level, newline, raw, format, *args): # log level higher than logging level? if level > self._level: return
3d7d63966162f8881fad278dc8cc728a7e5a1f77 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1143/3d7d63966162f8881fad278dc8cc728a7e5a1f77/logger.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 1330, 12, 2890, 16, 1801, 16, 9472, 16, 1831, 16, 740, 16, 380, 1968, 4672, 468, 613, 1801, 10478, 2353, 2907, 1801, 35, 309, 1801, 405, 365, 6315, 2815, 30, 327, 2, 0, 0, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 1330, 12, 2890, 16, 1801, 16, 9472, 16, 1831, 16, 740, 16, 380, 1968, 4672, 468, 613, 1801, 10478, 2353, 2907, 1801, 35, 309, 1801, 405, 365, 6315, 2815, 30, 327, 2, -100, -100, ...
assert u"%(x)s, %()s" % {'x':u"abc", u''.encode('utf-8'):"def"} == u'abc, def'
try: value = u"%(x)s, %()s" % {'x':u"abc", u''.encode('utf-8'):"def"} except KeyError: print '*** formatting failed for "%s"' % "u'abc, def'" else: assert value == u'abc, def'
test('capwords', u'abc\t def \nghi', u'Abc Def Ghi')
6b096315daae01cf10bf6b6f3074de1bb0c20fec /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/6b096315daae01cf10bf6b6f3074de1bb0c20fec/test_unicode.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1842, 2668, 5909, 3753, 2187, 582, 11, 18947, 64, 88, 282, 1652, 225, 521, 3368, 12266, 2187, 582, 11, 5895, 71, 10922, 611, 12266, 6134, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1842, 2668, 5909, 3753, 2187, 582, 11, 18947, 64, 88, 282, 1652, 225, 521, 3368, 12266, 2187, 582, 11, 5895, 71, 10922, 611, 12266, 6134, 2, -100, -100, -100, -100, -100, -100, -100, -100, -10...
if(self.target_name.count("kunit") == 1): assert self.scale_factor == 1
if(self.target_name.count("kunit") == 1 or self.target_name.count("ask") == 1): assert self.scale_factor != 0.0
def target_functor_w(self, selection = None): if(selection is None): f_obs = self.f_obs_w weights = self.weights_w abcd = self.abcd_w else: assert selection.size() == self.f_obs_w.data().size() f_obs = self.f_obs_w.select(selection) if(self.weights_w is not None): weights = self.weights_w.select(selection) else: weights = self.weights_w if(self.abcd_w is not None): abcd = self.abcd_w.select(selection) else: abcd = self.abcd_w if(self.target_name.count("k1") == 1): assert self.scale_factor == 0 return ls_k1(f_obs = f_obs, weights = weights, scale_factor = self.scale_factor, fix_scale_factor = False) if(self.target_name.count("k2") == 1): assert self.scale_factor == 0 return ls_k2(f_obs = f_obs, weights = weights, scale_factor = self.scale_factor, fix_scale_factor = False) if(self.target_name.count("kunit") == 1): assert self.scale_factor == 1 return ls_k1(f_obs = f_obs, weights = weights, scale_factor = self.scale_factor, fix_scale_factor = True) if(self.target_name == "ml"): return maximum_likelihood_criterion(f_obs = f_obs) if(self.target_name == "mlhl"): return maximum_likelihood_criterion_hl( f_obs = f_obs, abcd = abcd.data())
082c0769ffd6e9832016293c38a1c2d981e8374a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/696/082c0769ffd6e9832016293c38a1c2d981e8374a/target_functors.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1018, 67, 74, 6931, 280, 67, 91, 12, 2890, 16, 4421, 273, 599, 4672, 309, 12, 10705, 353, 599, 4672, 284, 67, 10992, 282, 273, 365, 18, 74, 67, 10992, 67, 91, 5376, 273, 365, 18, 9...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1018, 67, 74, 6931, 280, 67, 91, 12, 2890, 16, 4421, 273, 599, 4672, 309, 12, 10705, 353, 599, 4672, 284, 67, 10992, 282, 273, 365, 18, 74, 67, 10992, 67, 91, 5376, 273, 365, 18, 9...
opts, args = getopt.getopt(sys.argv[1:], "-p:-h", ["port", "host"])
opts, args = getopt.getopt(sys.argv[1:], '-p:-h', ['port', 'host'])
def opts(): try: opts, args = getopt.getopt(sys.argv[1:], "-p:-h", ["port", "host"]) except getopt.GetoptError, err: print str(err) sys.exit() global host global port host = "localhost" port = 1984 for o, a in opts: if o == "-p": port = int(a) if o == "-h": host = a
2c08272f70137ab264e2522f23535b6805965cfd /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/13081/2c08272f70137ab264e2522f23535b6805965cfd/BaseX.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1500, 13332, 775, 30, 1500, 16, 833, 273, 336, 3838, 18, 588, 3838, 12, 9499, 18, 19485, 63, 21, 30, 6487, 2400, 84, 30, 17, 76, 2187, 10228, 655, 2187, 296, 2564, 19486, 1335, 336, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1500, 13332, 775, 30, 1500, 16, 833, 273, 336, 3838, 18, 588, 3838, 12, 9499, 18, 19485, 63, 21, 30, 6487, 2400, 84, 30, 17, 76, 2187, 10228, 655, 2187, 296, 2564, 19486, 1335, 336, ...
if day > self.dayOfWeek: daysToWait = (day - self.dayOfWeek) - 1 self.rolloverAt = self.rolloverAt + (daysToWait * (60 * 60 * 24)) if day < self.dayOfWeek: daysToWait = (6 - self.dayOfWeek) + day
if day != self.dayOfWeek: if day < self.dayOfWeek: daysToWait = self.dayOfWeek - day - 1 else: daysToWait = 6 - day + self.dayOfWeek
def __init__(self, filename, when='h', interval=1, backupCount=0, encoding=None): BaseRotatingHandler.__init__(self, filename, 'a', encoding) self.when = string.upper(when) self.backupCount = backupCount # Calculate the real rollover interval, which is just the number of # seconds between rollovers. Also set the filename suffix used when # a rollover occurs. Current 'when' events supported: # S - Seconds # M - Minutes # H - Hours # D - Days # midnight - roll over at midnight # W{0-6} - roll over on a certain day; 0 - Monday # # Case of the 'when' specifier is not important; lower or upper case # will work. currentTime = int(time.time()) if self.when == 'S': self.interval = 1 # one second self.suffix = "%Y-%m-%d_%H-%M-%S" elif self.when == 'M': self.interval = 60 # one minute self.suffix = "%Y-%m-%d_%H-%M" elif self.when == 'H': self.interval = 60 * 60 # one hour self.suffix = "%Y-%m-%d_%H" elif self.when == 'D' or self.when == 'MIDNIGHT': self.interval = 60 * 60 * 24 # one day self.suffix = "%Y-%m-%d" elif self.when.startswith('W'): self.interval = 60 * 60 * 24 * 7 # one week if len(self.when) != 2: raise ValueError("You must specify a day for weekly rollover from 0 to 6 (0 is Monday): %s" % self.when) if self.when[1] < '0' or self.when[1] > '6': raise ValueError("Invalid day specified for weekly rollover: %s" % self.when) self.dayOfWeek = int(self.when[1]) self.suffix = "%Y-%m-%d" else: raise ValueError("Invalid rollover interval specified: %s" % self.when)
5d512fa2e42642e6e4e143967a5d72b71cdec5e4 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8546/5d512fa2e42642e6e4e143967a5d72b71cdec5e4/handlers.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 1544, 16, 1347, 2218, 76, 2187, 3673, 33, 21, 16, 5114, 1380, 33, 20, 16, 2688, 33, 7036, 4672, 3360, 8570, 1776, 1503, 16186, 2738, 972, 12, 2890, 16, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2738, 972, 12, 2890, 16, 1544, 16, 1347, 2218, 76, 2187, 3673, 33, 21, 16, 5114, 1380, 33, 20, 16, 2688, 33, 7036, 4672, 3360, 8570, 1776, 1503, 16186, 2738, 972, 12, 2890, 16, ...
[NOTE: this test currently FAILS due to trac \ essentially correct though. Should be revisited when that bug is resolved.]
def padic_sigma(self, p, N=20, E2=None, check=False, check_hypotheses=True): r""" Computes the p-adic sigma function with respect to the standard invariant differential $dx/(2y + a_1 x + a_3)$, as defined by Mazur and Tate, as a power series in the usual uniformiser $t$ at the origin.
e901b45b607d92ed1b9df94a3e96aead8ff614d6 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9890/e901b45b607d92ed1b9df94a3e96aead8ff614d6/ell_rational_field.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4627, 335, 67, 13201, 12, 2890, 16, 293, 16, 423, 33, 3462, 16, 512, 22, 33, 7036, 16, 866, 33, 8381, 16, 866, 67, 76, 879, 10370, 281, 281, 33, 5510, 4672, 436, 8395, 14169, 281, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4627, 335, 67, 13201, 12, 2890, 16, 293, 16, 423, 33, 3462, 16, 512, 22, 33, 7036, 16, 866, 33, 8381, 16, 866, 67, 76, 879, 10370, 281, 281, 33, 5510, 4672, 436, 8395, 14169, 281, ...
return self.execute('xls2csv -d UTF-8 -q 0 "%s" 2> nul:' % tmp_name)
html = self.execute('xlhtml "%s"' % tmp_name) return html2text(html, ignore_tags=('img',), indent_width=4, page_width=80)
def xls_convert(self, doc): """Convert Excel document to raw text""" tmp_name = self.saveFile(doc) if sys.platform == 'win32': return self.execute('xls2csv -d UTF-8 -q 0 "%s" 2> nul:' % tmp_name) else: return self.execute('xls2csv -d UTF-8 -q 0 "%s" 2> /dev/null' % tmp_name)
48abb591653ffe58326fac6e17b9dcd2287ff3bc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3287/48abb591653ffe58326fac6e17b9dcd2287ff3bc/__init__.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 619, 3251, 67, 6283, 12, 2890, 16, 997, 4672, 3536, 2723, 17674, 1668, 358, 1831, 977, 8395, 1853, 67, 529, 273, 365, 18, 5688, 812, 12, 2434, 13, 309, 2589, 18, 9898, 422, 296, 8082, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 619, 3251, 67, 6283, 12, 2890, 16, 997, 4672, 3536, 2723, 17674, 1668, 358, 1831, 977, 8395, 1853, 67, 529, 273, 365, 18, 5688, 812, 12, 2434, 13, 309, 2589, 18, 9898, 422, 296, 8082, ...
skip_chars = decoded_chars - len(self._decoded_text)
skip_chars = self._decoded_text_offset
def tell(self): if not self._seekable: raise IOError("underlying stream is not seekable") if not self._telling: raise IOError("telling position disabled by next() call") self.flush() position = self.buffer.tell() decoder = self._decoder if decoder is None or self._snapshot is None: if self._decoded_text: # This should never happen. raise AssertionError("pending decoded text") return position
dbe28e573e5c1227777db9dd630fac7732b81374 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8546/dbe28e573e5c1227777db9dd630fac7732b81374/io.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 9276, 12, 2890, 4672, 309, 486, 365, 6315, 16508, 429, 30, 1002, 8340, 2932, 9341, 6291, 1407, 353, 486, 6520, 429, 7923, 309, 486, 365, 6315, 88, 1165, 310, 30, 1002, 8340, 2932, 88, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 9276, 12, 2890, 4672, 309, 486, 365, 6315, 16508, 429, 30, 1002, 8340, 2932, 9341, 6291, 1407, 353, 486, 6520, 429, 7923, 309, 486, 365, 6315, 88, 1165, 310, 30, 1002, 8340, 2932, 88, ...
parser.parse("wu-blastp.out")
parser.parse(outf)
def main(): from xml.sax import saxutils parser = blastp.make_parser(debug_level = 0) parser.setContentHandler(saxutils.XMLGenerator()) parser.parse("wu-blastp.out") #parser.parse("wu-sh_blastp.out")
0c2f7f68124acbc5695851db91d462c21e90eef1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7167/0c2f7f68124acbc5695851db91d462c21e90eef1/wublast.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 628, 2025, 18, 87, 651, 1930, 20319, 5471, 2082, 273, 19047, 84, 18, 6540, 67, 4288, 12, 4148, 67, 2815, 273, 374, 13, 2082, 18, 542, 1350, 1503, 12, 87, 651, 5471, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 628, 2025, 18, 87, 651, 1930, 20319, 5471, 2082, 273, 19047, 84, 18, 6540, 67, 4288, 12, 4148, 67, 2815, 273, 374, 13, 2082, 18, 542, 1350, 1503, 12, 87, 651, 5471, 18, ...
return filter(None, map(lambda n, e=self.entries: e.get(n), names))
return map(lambda n, e=self.entries: e[_my_normcase(n)], names)
def _glob1(self, pattern, ondisk=True, source=False, strings=False): """ Globs for and returns a list of entry names matching a single pattern in this directory.
2958b4895e2270f8e56503dbda89ab0dced75be0 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12817/2958b4895e2270f8e56503dbda89ab0dced75be0/FS.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 10581, 21, 12, 2890, 16, 1936, 16, 603, 10863, 33, 5510, 16, 1084, 33, 8381, 16, 2064, 33, 8381, 4672, 3536, 18901, 87, 364, 471, 1135, 279, 666, 434, 1241, 1257, 3607, 279, 2202,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 10581, 21, 12, 2890, 16, 1936, 16, 603, 10863, 33, 5510, 16, 1084, 33, 8381, 16, 2064, 33, 8381, 4672, 3536, 18901, 87, 364, 471, 1135, 279, 666, 434, 1241, 1257, 3607, 279, 2202,...
info, " /' < ", __tmpfilename,
info, " /' < ", __tmpfilename,
def addinfoToEachRaw(targetfile="",info=""): """ add info to each raw. """ __funcname="addinfoToEachRaw" __errmsg="Error"+" in "+__funcname if (targetfile is ""): print __errmsg, "Enough numbers of arguments are not given." sys.exit() if os.path.isfile(targetfile) is False: print __errmsg, targetfile, "does not exit!!" sys.exit()
a6565b4090de43bfb103a990979db74b9b2cddb1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5758/a6565b4090de43bfb103a990979db74b9b2cddb1/combiner_v2.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 1376, 774, 3442, 4809, 12, 3299, 768, 1546, 3113, 1376, 1546, 6, 4672, 3536, 527, 1123, 358, 1517, 1831, 18, 3536, 1001, 644, 529, 1546, 1289, 1376, 774, 3442, 4809, 6, 1001, 24409,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 527, 1376, 774, 3442, 4809, 12, 3299, 768, 1546, 3113, 1376, 1546, 6, 4672, 3536, 527, 1123, 358, 1517, 1831, 18, 3536, 1001, 644, 529, 1546, 1289, 1376, 774, 3442, 4809, 6, 1001, 24409,...
g2list = [gen for gen in self.generators if gen.HandlesEntry(entry, metadata)]
g2list = [gen for gen in self.generators if gen.HandlesEntry(entry, metadata)]
def Bind(self, entry, metadata): '''Bind an entry using the appropriate generator''' if 'altsrc' in entry.attrib: oldname = entry.get('name') entry.set('name', entry.get('altsrc')) entry.set('realname', oldname) del entry.attrib['altsrc'] try: ret = self.Bind(entry, metadata) entry.set('name', oldname) del entry.attrib['realname'] return ret except: entry.set('name', oldname) logger.error("Failed binding entry %s:%s with altsrc %s" \ % (entry.tag, entry.get('name'), entry.get('altsrc'))) logger.error("Falling back to %s:%s" % (entry.tag, entry.get('name')))
4d1544016635254d1c51ec77e3e26699d64b1d26 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11867/4d1544016635254d1c51ec77e3e26699d64b1d26/Core.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6936, 12, 2890, 16, 1241, 16, 1982, 4672, 9163, 3357, 392, 1241, 1450, 326, 5505, 4456, 26418, 309, 296, 2390, 4816, 11, 316, 1241, 18, 14588, 30, 1592, 529, 273, 1241, 18, 588, 2668, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 6936, 12, 2890, 16, 1241, 16, 1982, 4672, 9163, 3357, 392, 1241, 1450, 326, 5505, 4456, 26418, 309, 296, 2390, 4816, 11, 316, 1241, 18, 14588, 30, 1592, 529, 273, 1241, 18, 588, 2668, ...
print 'm=',m
def getentry(self, which): self._check() if which == 'in': cmd = IN_ENTRY_SENSE elif which == 'out': cmd = OUT_ENTRY_SENSE self.replycmd(cmd) h = self._getnumber(2) print 'h=',h m = self._getnumber(2) print 'm=',m s = self._getnumber(2) print 's=',s f = self._getnumber(2) print 'f=',f return (h, m, s, f)
0b424aeee61cf467f726567c4a8deb7458742940 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/0b424aeee61cf467f726567c4a8deb7458742940/VCR.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 4099, 12, 2890, 16, 1492, 4672, 365, 6315, 1893, 1435, 309, 1492, 422, 296, 267, 4278, 1797, 273, 2120, 67, 19083, 67, 1090, 50, 1090, 1327, 1492, 422, 296, 659, 4278, 1797, 273, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 336, 4099, 12, 2890, 16, 1492, 4672, 365, 6315, 1893, 1435, 309, 1492, 422, 296, 267, 4278, 1797, 273, 2120, 67, 19083, 67, 1090, 50, 1090, 1327, 1492, 422, 296, 659, 4278, 1797, 273, ...
if type(self.L[0]) in [list, tuple]: return type(self.L[0][index]) else: return type(self.L[0])
return self.__types[index]
def on_get_column_type(self, index): if type(self.L[0]) in [list, tuple]: return type(self.L[0][index]) else: return type(self.L[0])
cf5341a61a2498b093430a22411daa9f62f4e076 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8273/cf5341a61a2498b093430a22411daa9f62f4e076/pythonlistmodel.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 603, 67, 588, 67, 2827, 67, 723, 12, 2890, 16, 770, 4672, 309, 618, 12, 2890, 18, 48, 63, 20, 5717, 316, 306, 1098, 16, 3193, 14542, 327, 618, 12, 2890, 18, 48, 63, 20, 6362, 1615,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 603, 67, 588, 67, 2827, 67, 723, 12, 2890, 16, 770, 4672, 309, 618, 12, 2890, 18, 48, 63, 20, 5717, 316, 306, 1098, 16, 3193, 14542, 327, 618, 12, 2890, 18, 48, 63, 20, 6362, 1615,...
class TestRq(NumpyTestCase):
class TestRQ(NumpyTestCase):
def check_random_complex(self): n = 20 for k in range(2): a = random([n,n])+1j*random([n,n]) q,r = qr(a) assert_array_almost_equal(dot(conj(transpose(q)),q),identity(n)) assert_array_almost_equal(dot(q,r),a)
ed190d870bbfc8f800477027b23dcafc112a646e /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12971/ed190d870bbfc8f800477027b23dcafc112a646e/test_decomp.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 866, 67, 9188, 67, 14259, 12, 2890, 4672, 290, 273, 4200, 364, 417, 316, 1048, 12, 22, 4672, 279, 273, 2744, 3816, 82, 16, 82, 5717, 15, 21, 78, 14, 9188, 3816, 82, 16, 82, 5717, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 866, 67, 9188, 67, 14259, 12, 2890, 4672, 290, 273, 4200, 364, 417, 316, 1048, 12, 22, 4672, 279, 273, 2744, 3816, 82, 16, 82, 5717, 15, 21, 78, 14, 9188, 3816, 82, 16, 82, 5717, 1...
print " * Generating includer..." sourceOutput = settingsStr srcEol = "";
if options.sourceTemplateInputFile: options.sourceTemplateInputFile = os.path.normpath(options.sourceTemplateInputFile) if options.sourceTemplateOutputFile: options.sourceTemplateOutputFile = os.path.normpath(options.sourceTemplateOutputFile) print " * Generating script block..." sourceLineFeed = "";
def execute(fileDb, moduleDb, options, pkgid="", names=[]): additionalOutput = [] ###################################################################### # SORT OF INCLUDE LIST ###################################################################### print print " SORT OF INCLUDE LIST:" print "----------------------------------------------------------------------------" if options.verbose: print " * Include (with dependencies): %s" % options.includeWithDeps print " * Include (without dependencies): %s" % options.includeWithoutDeps print " * Exclude (with dependencies): %s" % options.excludeWithDeps print " * Exclude (without dependencies): %s" % options.excludeWithoutDeps print " * Sorting %s classes..." % len(fileDb) sortedIncludeList = loader.getSortedList(options, fileDb, moduleDb) if len(sortedIncludeList) == 0: print " - No class files to include. Exciting!" sys.exit(1) else: print " - Including %s classes" % len(sortedIncludeList) if options.printIncludes: print print " PRINT OF INCLUDE ORDER:" print "----------------------------------------------------------------------------" print " * The files will be included in this order:" for fileId in sortedIncludeList: print " - %s" % fileId if options.printDeps: print print " OUTPUT OF DEPENDENCIES:" print "----------------------------------------------------------------------------" print " * These are all included files with their dependencies:" for fileId in sortedIncludeList: print " - %s" % fileId if len(fileDb[fileId]["loadtimeDeps"]) > 0: print " - Loadtime: " for depEntry in fileDb[fileId]["loadtimeDeps"]: print " - %s" % depEntry if len(fileDb[fileId]["afterDeps"]) > 0: print " - After: " for depEntry in fileDb[fileId]["afterDeps"]: print " - %s" % depEntry if len(fileDb[fileId]["runtimeDeps"]) > 0: print " - Runtime: " for depEntry in fileDb[fileId]["runtimeDeps"]: print " - %s" % depEntry if len(fileDb[fileId]["beforeDeps"]) > 0: print " - Before: " for depEntry in fileDb[fileId]["beforeDeps"]: print " - %s" % depEntry if len(fileDb[fileId]["optionalDeps"]) > 0: print " - Optional: " for depEntry in fileDb[fileId]["optionalDeps"]: print " - %s" % depEntry ###################################################################### # SOURCE MIGRATION ###################################################################### if options.migrateSource: print print " SOURCE MIGRATION:" print "----------------------------------------------------------------------------" print " * Migrate Source Code..." migrator.handle(sortedIncludeList, fileDb, options) # Return after migration: Ignore other jobs return ###################################################################### # GENERATION OF PRETTY PRINTED CODE ###################################################################### if options.fixSource: print print " FIX SOURCE CODE:" print "----------------------------------------------------------------------------" if options.verbose: print " * Fixing code..." else: print " * Fixing code: ", for fileId in sortedIncludeList: if options.verbose: print " - Reading %s" % fileId fileEntry = fileDb[fileId] filePath = fileEntry["path"] fileEncoding = fileEntry["encoding"] fileContent = filetool.read(filePath, fileEncoding) fixedContent = textutil.removeTrailingSpaces(textutil.tab2Space(textutil.any2Unix(fileContent), 2)) if fixedContent != fileContent: if options.verbose: print " - Storing modifications..." else: sys.stdout.write("!") sys.stdout.flush() filetool.save(filePath, fixedContent, fileEncoding) elif not options.verbose: sys.stdout.write(".") sys.stdout.flush() if not options.verbose: print # Return after fixing: Ignore other jobs return ###################################################################### # GENERATION OF PRETTY PRINTED CODE ###################################################################### if options.prettyPrint: print print " GENERATION OF PRETTY PRINTED CODE:" print "----------------------------------------------------------------------------" if options.verbose: print " * Pretty printing..." else: print " * Pretty printing: ", for fileId in sortedIncludeList: if options.verbose: print " - Compiling %s" % fileId else: sys.stdout.write(".") sys.stdout.flush() prettyFileContent = compiler.compile(loader.getTree(fileDb, fileId, options), True) if not prettyFileContent.endswith("\n"): prettyFileContent += "\n" filetool.save(fileDb[fileId]["path"], prettyFileContent) if not options.verbose: print # Return after pretty print: Ignore other jobs return ###################################################################### # STRING OPTIMIZATION ###################################################################### if options.optimizeStrings: print print " STRING OPTIMIZATION:" print "----------------------------------------------------------------------------" if options.verbose: print " * Searching strings..." else: print " * Searching strings: ", stringMap = {} for fileId in sortedIncludeList: if options.verbose: print " - %s" % fileId else: sys.stdout.write(".") sys.stdout.flush() localMap = loader.getStrings(fileDb, fileId, options) for value in localMap: if value in stringMap: stringMap[value] += localMap[value] else: stringMap[value] = localMap[value] if not options.verbose: print counter = 0 for value in stringMap: counter += stringMap[value] stringList = stringoptimizer.sort(stringMap) print " * Found %s strings (used %s times)" % (len(stringMap), counter) if options.verbose: print " * Replacing strings..." else: print " * Replacing strings: ", for fileId in sortedIncludeList: if options.verbose: print " - %s" % fileId else: sys.stdout.write(".") sys.stdout.flush() stringoptimizer.replace(loader.getTree(fileDb, fileId, options), stringList, "$" + pkgid, options.verbose) if not options.verbose: print print " * Generating replacement..." additionalOutput.append(stringoptimizer.replacement(stringList, "$" + pkgid)) ###################################################################### # LOCAL VARIABLE OPTIMIZATION ###################################################################### if options.optimizeVariables: print print " LOCAL VARIABLE OPTIMIZATION:" print "----------------------------------------------------------------------------" if options.verbose: print " * Optimizing variables..." else: print " * Optimizing variables: ", for fileId in sortedIncludeList: if options.verbose: print " - %s" % fileId else: sys.stdout.write(".") sys.stdout.flush() variableoptimizer.search(loader.getTree(fileDb, fileId, options), [], 0, "$", skipPrefix = options.optimizeVariablesSkipPrefix, debug = options.enableDebug) if not options.verbose: print ###################################################################### # NAME OBFUSCATION ###################################################################### if options.obfuscateIdentifiers: print print " OBFUSCATE IDENTIFIERS:" print "----------------------------------------------------------------------------" if options.verbose: print " * Obfuscating identifiers..." else: print " * Obfuscating identifiers: ", counter = 0 for fileId in sortedIncludeList: if options.verbose: print " - %s" % fileId else: sys.stdout.write(".") sys.stdout.flush() counter += obfuscator.update(loader.getTree(fileDb, fileId, options), names, "$$") if not options.verbose: print print " * Updated %s names" % counter ###################################################################### # TOKEN STORAGE ###################################################################### if options.storeTokens: print print " TOKEN STORAGE:" print "----------------------------------------------------------------------------" if options.tokenOutputDirectory == None: print " * You must define the token output directory!" sys.exit(1) if options.verbose: print " * Storing tokens..." else: print " * Storing tokens: ", for fileId in sortedIncludeList: tokenString = tokenizer.convertTokensToString(loader.getTokens(fileDb, fileId, options)) if options.verbose: print " * writing tokens for %s (%s KB)..." % (fileIdm, len(tokenString) / 1000.0) else: sys.stdout.write(".") sys.stdout.flush() filetool.save(os.path.join(filetool.normalize(options.tokenOutputDirectory), fileId + config.TOKENEXT), tokenString) if not options.verbose: print ###################################################################### # TREE STORAGE ###################################################################### if options.storeTree: print print " TREE STORAGE:" print "----------------------------------------------------------------------------" if options.treeOutputDirectory == None: print " * You must define the tree output directory!" sys.exit(1) if options.verbose: print " * Storing tree..." else: print " * Storing tree: ", for fileId in sortedIncludeList: treeString = "<?xml version=\"1.0\" encoding=\"" + options.xmlOutputEncoding + "\"?>\n" + tree.nodeToXmlString(loader.getTree(fileDb, fileId, options)) if options.verbose: print " * writing tree for %s (%s KB)..." % (fileId, len(treeString) / 1000.0) else: sys.stdout.write(".") sys.stdout.flush() filetool.save(os.path.join(filetool.normalize(options.treeOutputDirectory), fileId + config.XMLEXT), treeString) if not options.verbose: print ###################################################################### # GENERATION OF API ###################################################################### if options.generateApiDocumentation: print print " GENERATION OF API:" print "----------------------------------------------------------------------------" if options.apiDocumentationJsonFile == None and options.apiDocumentationXmlFile == None: print " * You must define one of JSON or XML API documentation file!" docTree = None if options.verbose: print " * Generating API tree..." else: print " * Generating API tree: ", for fileId in sortedIncludeList: if options.verbose: print " - %s" % fileId else: sys.stdout.write(".") sys.stdout.flush() docTree = api.createDoc(loader.getTree(fileDb, fileId, options), docTree) if not options.verbose: print if docTree: print " * Finalizing tree..." api.postWorkPackage(docTree, docTree) if options.apiDocumentationXmlFile != None: print " * Writing XML API file to %s" % options.apiDocumentationXmlFile xmlContent = "<?xml version=\"1.0\" encoding=\"" + options.xmlOutputEncoding + "\"?>\n" if options.addNewLines: xmlContent += "\n" + tree.nodeToXmlString(docTree) else: xmlContent += tree.nodeToXmlString(docTree, "", "", "") filetool.save(options.apiDocumentationXmlFile, xmlContent, options.xmlOutputEncoding) if options.apiDocumentationJsonFile != None: print " * Writing JSON API file to %s" % options.apiDocumentationJsonFile if options.addNewLines: jsonContent = tree.nodeToJsonString(docTree) else: jsonContent = tree.nodeToJsonString(docTree, "", "", "") filetool.save(options.apiDocumentationJsonFile, jsonContent, options.scriptOutputEncoding) ###################################################################### # CREATE COPY OF RESOURCES ###################################################################### if options.copyResources: print print " CREATE COPY OF RESOURCES:" print "----------------------------------------------------------------------------" resources.copy(options, sortedIncludeList, fileDb) ###################################################################### # GENERATION OF SETTINGS ###################################################################### if options.generateSourceScript or options.generateCompiledScript: settingsStr = "" if len(options.defineRuntimeSetting) != 0: print print " GENERATION OF SETTINGS:" print "----------------------------------------------------------------------------" print " * Processing input data..." settingsStr = settings.generate(options) if options.settingsScriptFile: print " * Storing result to %s" % options.settingsScriptFile filetool.save(options.settingsScriptFile, settingsStr) # clear settings for build and source settingsStr = "" ###################################################################### # GENERATION OF SOURCE VERSION ###################################################################### if options.generateSourceScript: print print " GENERATION OF SOURCE SCRIPT:" print "----------------------------------------------------------------------------" if options.sourceScriptFile == None: print " * You must define the source script file!" sys.exit(1) else: options.sourceScriptFile = os.path.normpath(options.sourceScriptFile) print " * Generating includer..." sourceOutput = settingsStr srcEol = ""; if options.addNewLines: srcEol = "\n"; if sourceOutput != "": settingsStr += srcEol # Define javascript loaders jsLoaders = {} # HTML-only: create <script> tags using document.write() jsLoaders["docwrite"] = """var includeJs=function(src){document.write('<script type="text/javascript" src="'+src+'"></script>')};""" # XHTML-compatible: load scripts using XMLHttpRequest and eval() them jsLoaders["xhrequest"] = """{var xhr=null;if(window.XMLHttpRequest)xhr=new XMLHttpRequest;else if(window.ActiveXObject){var s=["MSXML2.XMLHTTP.3.0","MSXML2.XMLHTTP.6.0","MSXML2.XMLHTTP","Microsoft.XMLHTTP"];for(var i=0;i<s.length;++i)try{xhr=new ActiveXObject(s[i])}catch(e){}};""" + srcEol + """if(!xhr){alert("Sorry, you need support for XMLHttpRequest in order to\\nload source builds into XHTML documents.");return}""" + srcEol + """var includeJs=function(src){xhr.open("GET", src, false);xhr.send(null);window.eval(xhr.responseText)}}""" # Source loader closure sourceOutput += """(function(sources){""" + srcEol # Differentiate source build sourceOutput += """qx.SOURCE_BUILD=true;""" + srcEol # Detect the node we are being called from sourceOutput += """var parentNode=document.getElementsByTagName('body')[0]||document.getElementsByTagName('head')[0];""" + srcEol # Autoselect a loader based on document type if options.sourceLoaderType == "auto": # Detect if document type is XHTML sourceOutput += """var isXHTML=false;""" + srcEol sourceOutput += """if(document.doctype&&document.doctype.publicId.match(/-\/\/W3C\/\/DTD XHTML/))isXHTML=true;""" + srcEol sourceOutput += """else if(parentNode.namespaceURI&&parentNode.namespaceURI.match(/\/xhtml/))isXHTML=true;""" + srcEol # Select the loader based on document type sourceOutput += """if(isXHTML)""" + jsLoaders["xhrequest"]+ srcEol sourceOutput += """else """ + jsLoaders["docwrite"] + srcEol # Use a fixed loader else: sourceOutput += jsLoaders[options.sourceLoaderType] + srcEol # Loading loop sourceOutput += """for(var i=0;i<sources.length;++i)includeJs(sources[i])""" + srcEol + """})""" sources = "" for fileId in sortedIncludeList: if fileDb[fileId]["classUri"] == None: print " * Missing source path definition for script input %s. Could not create source script file!" % fileDb[fileId]["classPath"] sys.exit(1) sources += srcEol + '"%s%s",' % (os.path.join(fileDb[fileId]["classUri"], fileDb[fileId]["pathId"].replace(".", os.sep)), config.JSEXT) # Pass the array with source files to include sourceOutput += "([" + sources[:-1] + srcEol + "]);" + srcEol print " * Storing output as %s..." % options.sourceScriptFile filetool.save(options.sourceScriptFile, sourceOutput, options.scriptOutputEncoding) ###################################################################### # GENERATION OF COMPILED VERSION ###################################################################### if options.generateCompiledScript: print print " GENERATION OF COMPILED SCRIPT:" print "----------------------------------------------------------------------------" compiledOutput = settingsStr + "".join(additionalOutput) if options.compiledScriptFile == None: print " * You must define the compiled script file!" sys.exit(1) if options.verbose: print " * Compiling..." else: print " * Compiling: ", for fileId in sortedIncludeList: if options.verbose: print " - Compiling %s" % fileId else: sys.stdout.write(".") sys.stdout.flush() compiledFileContent = compiler.compile(loader.getTree(fileDb, fileId, options), False, options.addNewLines, options.enableDebug) if options.addFileIds: compiledOutput += "\n\n\n/* ID: " + fileId + " */\n" + compiledFileContent + "\n" else: compiledOutput += compiledFileContent if not compiledOutput.endswith(";") and not compiledOutput.endswith("\n"): compiledOutput += ";" if not options.verbose: print print " * Storing output as %s..." % options.compiledScriptFile filetool.save(options.compiledScriptFile, compiledOutput, options.scriptOutputEncoding)
c747d776efd38fcc589afb4462ed3b74c1fc7c1b /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/5718/c747d776efd38fcc589afb4462ed3b74c1fc7c1b/generator.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1836, 12, 768, 4331, 16, 1605, 4331, 16, 702, 16, 3475, 350, 1546, 3113, 1257, 33, 8526, 4672, 225, 3312, 1447, 273, 5378, 282, 28256, 1970, 7, 468, 225, 11530, 15932, 28062, 15130, 2825...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1836, 12, 768, 4331, 16, 1605, 4331, 16, 702, 16, 3475, 350, 1546, 3113, 1257, 33, 8526, 4672, 225, 3312, 1447, 273, 5378, 282, 28256, 1970, 7, 468, 225, 11530, 15932, 28062, 15130, 2825...
Abelian variety factor of dimension 2 of J0(125), Abelian variety factor of dimension 2 of J0(125), Abelian variety factor of dimension 4 of J0(125)
Simple abelian subvariety 125a(1,125) of dimension 2 of J0(125), Simple abelian subvariety 125b(1,125) of dimension 2 of J0(125), Simple abelian subvariety 125c(1,125) of dimension 4 of J0(125)
def __getslice__(self, i, j): """ The slice i:j of decompositions of self. EXAMPLES: sage: J = J0(125); J.decomposition() [ Abelian variety factor of dimension 2 of J0(125), Abelian variety factor of dimension 2 of J0(125), Abelian variety factor of dimension 4 of J0(125) ] sage: J[:2] [ Abelian variety factor of dimension 2 of J0(125), Abelian variety factor of dimension 2 of J0(125) ] """ return self.decomposition()[i:j]
f30f87afcce1c095974cd5259aa7272b60928586 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/f30f87afcce1c095974cd5259aa7272b60928586/abvar.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 588, 6665, 972, 12, 2890, 16, 277, 16, 525, 4672, 3536, 1021, 2788, 277, 30, 78, 434, 9862, 12388, 434, 365, 18, 225, 5675, 8900, 11386, 30, 272, 410, 30, 804, 273, 804, 20, 12...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 588, 6665, 972, 12, 2890, 16, 277, 16, 525, 4672, 3536, 1021, 2788, 277, 30, 78, 434, 9862, 12388, 434, 365, 18, 225, 5675, 8900, 11386, 30, 272, 410, 30, 804, 273, 804, 20, 12...
recs.append(106, self.timestamp)
recs.append((106, self.timestamp))
def update(self, mi): recs = [] try: from calibre.ebooks.conversion.config import load_defaults prefs = load_defaults('mobi_output') pas = prefs.get('prefer_author_sort', False) except: pas = False if mi.author_sort and pas: authors = mi.author_sort recs.append((100, authors.encode(self.codec, 'replace'))) elif mi.authors: authors = '; '.join(mi.authors) recs.append((100, authors.encode(self.codec, 'replace'))) if mi.publisher: recs.append((101, mi.publisher.encode(self.codec, 'replace'))) if mi.comments: recs.append((103, mi.comments.encode(self.codec, 'replace'))) if mi.isbn: recs.append((104, mi.isbn.encode(self.codec, 'replace'))) if mi.tags: subjects = '; '.join(mi.tags) recs.append((105, subjects.encode(self.codec, 'replace')))
ffb1d3d770bd1f7764e724a7da0caa39a75038fb /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9125/ffb1d3d770bd1f7764e724a7da0caa39a75038fb/mobi.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 12, 2890, 16, 12837, 4672, 27255, 273, 5378, 775, 30, 628, 26139, 266, 18, 73, 12567, 18, 20990, 18, 1425, 1930, 1262, 67, 7606, 15503, 273, 1262, 67, 7606, 2668, 81, 30875, 67, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1089, 12, 2890, 16, 12837, 4672, 27255, 273, 5378, 775, 30, 628, 26139, 266, 18, 73, 12567, 18, 20990, 18, 1425, 1930, 1262, 67, 7606, 15503, 273, 1262, 67, 7606, 2668, 81, 30875, 67, ...
tableName = self.__getTableName( "bucket", typeName )
tableName = _getTableName( "bucket", typeName )
def __updateBucket( self, typeName, startTime, bucketLength, keyValues, bucketValues, proportion, connObj = False ): """ Update a bucket when coming from the raw insert """ tableName = self.__getTableName( "bucket", typeName ) cmd = "UPDATE `%s` SET " % tableName sqlValList = [] for pos in range( len( self.dbCatalog[ typeName ][ 'values' ] ) ): valueField = self.dbCatalog[ typeName ][ 'values' ][ pos ] value = bucketValues[ pos ] fullFieldName = "`%s`.`%s`" % ( tableName, valueField ) sqlValList.append( "%s=%s+(%s*%s)" % ( fullFieldName, fullFieldName, value, proportion ) ) sqlValList.append( "`%s`.`entriesInBucket`=`%s`.`entriesInBucket`+(%s*%s)" % ( tableName, tableName, bucketValues[-1], proportion ) ) cmd += ", ".join( sqlValList ) cmd += " WHERE `%s`.`startTime`='%s' AND `%s`.`bucketLength`='%s' AND " % ( tableName, startTime, tableName, bucketLength ) cmd += self.__generateSQLConditionForKeys( typeName, keyValues ) return self._update( cmd, conn = connObj )
5644734dd5f9b2176cf1b7afc4f6ab55941c88f0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/5644734dd5f9b2176cf1b7afc4f6ab55941c88f0/AccountingDB.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2725, 4103, 12, 365, 16, 8173, 16, 8657, 16, 2783, 1782, 16, 498, 1972, 16, 2783, 1972, 16, 23279, 16, 1487, 2675, 273, 1083, 262, 30, 3536, 2315, 279, 2783, 1347, 19283, 628, 32...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1001, 2725, 4103, 12, 365, 16, 8173, 16, 8657, 16, 2783, 1782, 16, 498, 1972, 16, 2783, 1972, 16, 23279, 16, 1487, 2675, 273, 1083, 262, 30, 3536, 2315, 279, 2783, 1347, 19283, 628, 32...
sage: P1xP1 = ToricVariety(fan) sage: P1xP1.inject_coefficients() The last command does nothing, since ``P1xP1`` is defined over `\QQ`. Let's construct a toric variety over a more complicated field::
def inject_coefficients(self, scope=None, verbose=True): r""" Inject generators of the base field of ``self`` into ``scope``.
bf8155d5074543bbf066ffe38ee5541218f84088 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9417/bf8155d5074543bbf066ffe38ee5541218f84088/toric_variety.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4590, 67, 8075, 12344, 12, 2890, 16, 2146, 33, 7036, 16, 3988, 33, 5510, 4672, 436, 8395, 20085, 13327, 434, 326, 1026, 652, 434, 12176, 2890, 10335, 1368, 12176, 4887, 68, 8338, 2, 0, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 4590, 67, 8075, 12344, 12, 2890, 16, 2146, 33, 7036, 16, 3988, 33, 5510, 4672, 436, 8395, 20085, 13327, 434, 326, 1026, 652, 434, 12176, 2890, 10335, 1368, 12176, 4887, 68, 8338, 2, -100...
self.pmonitors[i].start(int(duration / self.rate) + 1)
self.pmonitors[k].start(int(duration / self.rate) + 1)
def start(self, duration=None): for i in range(len(self.cbs)): self.pmonitors[i] = PeriodicMonitor(self.cbs[i], self.rate)
8c860a65b7d3d0c4154b79977b181d0ece64761d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8117/8c860a65b7d3d0c4154b79977b181d0ece64761d/perception_monitor.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 787, 12, 2890, 16, 3734, 33, 7036, 4672, 364, 277, 316, 1048, 12, 1897, 12, 2890, 18, 71, 2038, 3719, 30, 365, 18, 84, 2586, 13704, 63, 77, 65, 273, 12698, 335, 7187, 12, 2890, 18, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 787, 12, 2890, 16, 3734, 33, 7036, 4672, 364, 277, 316, 1048, 12, 1897, 12, 2890, 18, 71, 2038, 3719, 30, 365, 18, 84, 2586, 13704, 63, 77, 65, 273, 12698, 335, 7187, 12, 2890, 18, ...
-1.0
0.5 + 0.866025403784*I
def _complex_double_(self, field): """ EXAMPLES: sage: CDF((-1)^(1/3)) 0.5 + 0.866025403784*I
91ce32de726922bd909433810b61f39b42ddf8c5 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9417/91ce32de726922bd909433810b61f39b42ddf8c5/calculus.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 14259, 67, 9056, 67, 12, 2890, 16, 652, 4672, 3536, 5675, 8900, 11386, 30, 272, 410, 30, 31971, 12443, 17, 21, 13, 29020, 21, 19, 23, 3719, 374, 18, 25, 397, 374, 18, 5292, 4848...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 389, 14259, 67, 9056, 67, 12, 2890, 16, 652, 4672, 3536, 5675, 8900, 11386, 30, 272, 410, 30, 31971, 12443, 17, 21, 13, 29020, 21, 19, 23, 3719, 374, 18, 25, 397, 374, 18, 5292, 4848...
(text, create_nearest_terms_box(req.args, pi, fi)), "")
(text, create_nearest_terms_box(req.args, pi, fi, ti)))
def search_pattern(req, p=None, f=None, colls=[], m=None, hit_hints=0): """Searches for pattern 'p' and field 'f' and returns dict of recIDs HitLists per each collection in 'colls'. - Optionally, the function accepts the match type argument 'm'. If it is set (e.g. from advanced search interface), then it performs this kind of matching. If it is not set, then a guess is made. - Calls search_in_bibwords() and/or search_in_bibxxx() functions. - If hit_hints is set, than print lots of hints on current search. Useful for debugging and/or when search gives zero hits. - Called by main().""" ## create empty output results set: results_out = {} for coll in colls: results_out[coll] = HitList() results_out[coll]._nbhits = 0 ## if p is not defined, return all hits in given collections: if not p: for coll in colls: results_out[coll]._set = collrecs_cache[coll]._set results_out[coll]._nbhits = collrecs_cache[coll]._nbhits return results_out ## now we are sure to have some p: good. ## create search units: opft_items = create_opft_search_units(req, p, f, m) #req.write("<p>%s" % opft_items) ## search regardless of collection: results_in_any_collection = HitList() results_in_any_collection_empty = 1 # to be able to skip first set operation for opft_item in opft_items: results_for_opft_item = {} oi, pi, fi, ti = opft_item[0], opft_item[1], opft_item[2], opft_item[3] # firstly, launch search for this pattern item: if ti == 'a' or ti == 'r': # we are doing either direct bibxxx search or phrase search or regexp search results_for_opft_item = search_in_bibxxx(req, pi, fi, ti) elif ti == 'w': # we are doing bibwords search results_for_opft_item = search_in_bibwords(req, pi, fi) else: print_warning(req, "Search type '%s' is not defined." % ti, "Error") return None if hit_hints: results_for_opft_item.calculate_nbhits() if results_for_opft_item._nbhits == 0: text = "Search term <em>%s</em>" % pi if fi: text += " inside <em>%s</em> " % fi print_warning(req, "%s did not match any record. Nearest terms in any collection are: %s" % (text, create_nearest_terms_box(req.args, pi, fi)), "") return results_out # empty if dbg: print_warning(req, "Item <strong>%s:%s</strong> gave %d hits." % (fi, pi, results_for_opft_item._nbhits), "") # secondly, apply the set intersect/union/disjunction functions: if dbg: t1 = os.times()[4] if results_in_any_collection_empty: results_in_any_collection = results_for_opft_item results_in_any_collection_empty = 0 else: if oi == '+': results_in_any_collection.intersect(results_for_opft_item) elif oi == '-': results_in_any_collection.difference(results_for_opft_item) elif oi == '|': results_in_any_collection.union(results_for_opft_item) else: print_warning(req, "Invalid set operation %s." % oi, "Error") if dbg: t2 = os.times()[4] print_warning(req, "Set operation '%s' took %.2f seconds." % (oi, (t2 - t1)), "Profile") ## before doing collection intersection, see how many hits we have now: results_in_any_collection.calculate_nbhits() ## return all hits if no collection list specified (useful for WebSearch Admin to determine collection ## recID universe): if colls == []: return results_in_any_collection ## were there no hits at all before doing collection intersection? if results_in_any_collection._nbhits == 0: # pattern not found in any public/private collection: if hit_hints: text = """All search terms matched but boolean query returned no hits. Please combine your search terms differently.""" url_args = req.args text += """<blockquote><table class="nearesttermsbox" cellpadding="0" cellspacing="0" border="0">""" for opft_item in opft_items: oi, pi, fi, ti = opft_item[0], opft_item[1], opft_item[2], opft_item[3] url_args_new = re.sub(r'(^|\&)p=.*?(\&|$)', r'\1p='+urllib.quote(pi)+r'\2', url_args) url_args_new = re.sub(r'(^|\&)f=.*?(\&|$)', r'\1f='+urllib.quote(fi)+r'\2', url_args_new) text += """<tr><td class="nearesttermsboxbody" align="right">%s</td><td class="nearesttermsboxbody" width="15">&nbsp;</td><td class="nearesttermsboxbody" align="left"><a class="nearestterms" href="%s/search.py?%s">%s</a></td></tr>""" % \ (get_word_nbhits(pi, fi), weburl, url_args_new, pi) text += """</table></blockquote>""" print_warning(req, text, "") return results_out # still empty ## intersect with collection universe: if dbg: t1 = os.times()[4] if colls: for coll in colls: results_out[coll]._set = Numeric.bitwise_and(results_in_any_collection._set, collrecs_cache[coll]._set) if dbg: t2 = os.times()[4] print_warning(req, "Intersecting with collection hitlist universe took %.2f seconds." % ((t2 - t1)), "Profile") ## count number of hits in given collections: results_out_nb_hits_total = 0 for coll in colls: results_out[coll].calculate_nbhits() results_out_nb_hits_total += results_out[coll]._nbhits if results_out_nb_hits_total == 0: # pattern had been found but not in the collections chosen by the user: if hit_hints: # try to search in Home: results_in_Home = HitList() results_in_Home._set = Numeric.bitwise_and(results_in_any_collection._set, collrecs_cache[cdsname]._set) results_in_Home.calculate_nbhits() if results_in_Home._nbhits > 0: # some hits found in Home, so propose this search: url_args = req.args url_args = re.sub(r'(^|\&)cc=.*?(\&|$)', r'\2', url_args) url_args = re.sub(r'(^|\&)c=.*?(\&[^c]+=|$)', r'\2', url_args) url_args = re.sub(r'^\&+', '', url_args) url_args = re.sub(r'\&+$', '', url_args) print_warning(req, """Match found in other public collections: <a class="nearestterms" href="%s/search.py?%s">%d hits</a>.""" % (weburl, url_args, results_in_Home._nbhits), "") else: # no hits found in Home, recommend different search terms: text = """No public collection matched your query. If you were looking for a non-public document, please choose the desired restricted collection first.""" print_warning(req, text, "") return results_out
535ed6a8d836745c0ddd7f1016fbe145d7c7bebd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2139/535ed6a8d836745c0ddd7f1016fbe145d7c7bebd/search_engine.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1623, 67, 4951, 12, 3658, 16, 293, 33, 7036, 16, 284, 33, 7036, 16, 645, 3251, 22850, 6487, 312, 33, 7036, 16, 6800, 67, 24598, 33, 20, 4672, 3536, 16294, 364, 1936, 296, 84, 11, 471...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1623, 67, 4951, 12, 3658, 16, 293, 33, 7036, 16, 284, 33, 7036, 16, 645, 3251, 22850, 6487, 312, 33, 7036, 16, 6800, 67, 24598, 33, 20, 4672, 3536, 16294, 364, 1936, 296, 84, 11, 471...
cr, uid, ids, context=context)
cr, uid, ids, context=context)
def unlink(self, cr, uid, ids, context=None): for move in self.browse(cr, uid, ids, context=context): if move.state != 'draft': raise osv.except_osv(_('UserError'), _('You can only delete draft moves.')) return super(stock_move, self).unlink( cr, uid, ids, context=context)
18be8e52469417aa8e0ec02e033c0942d0ad805f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/18be8e52469417aa8e0ec02e033c0942d0ad805f/stock.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8255, 12, 2890, 16, 4422, 16, 4555, 16, 3258, 16, 819, 33, 7036, 4672, 364, 3635, 316, 365, 18, 25731, 12, 3353, 16, 4555, 16, 3258, 16, 819, 33, 2472, 4672, 309, 3635, 18, 2019, 480...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 8255, 12, 2890, 16, 4422, 16, 4555, 16, 3258, 16, 819, 33, 7036, 4672, 364, 3635, 316, 365, 18, 25731, 12, 3353, 16, 4555, 16, 3258, 16, 819, 33, 2472, 4672, 309, 3635, 18, 2019, 480...
psout = ps.stdout.read() psout.strip()
psout = ps.communicate()[0]
def main(): # Initialize the update url to just be the ip address of this machine. ip = getmyip() updateurl = 'http://' + ip + ':12345/' if len(sys.argv) == 1: sys.argv.append(updateurl) else: sys.argv[1] = updateurl # Create a temp directory to serve the updates from that we can # automatically clean up when we are done. tmpserver = tempfile.mkdtemp() try: test_updater.create_folders(tmpserver) ############################ # Run the rsync only tests # ############################ # Run the noup test (Nothing should fail, nothing should be updated) runRsyncTest('-x', tmpserver + '/noup/') # Run the wronghash test(There should be an RsyncError, and no updates) runRsyncTest('-e', tmpserver + '/wronghash/') # Run the badkeysig test (There should be no updates) runRsyncTest('-x', tmpserver + '/badkeysig/') # Run the corruptmeta test (there should be an RsyncError, and no updates) runRsyncTest('-e', tmpserver + '/corruptmeta/') # Run the updatenmmain test (only nmmain should get updated) runRsyncTest('-u', tmpserver + '/updatenmmain/', ['nmmain.py', 'metainfo']) # Run an update that should get us into a state where the softwareupdater has # a different key than what the metainfo is signed with. The next test will # ensure that our method of dealing with this situation works. runRsyncTest('-u', tmpserver + '/updater/', ['softwareupdater.py', 'metainfo']) # Run an update that should successfully update from the strange state from # the previous test. runRsyncTest('-u', tmpserver + '/updater_new/', ['nmmain.py', 'metainfo']) ##################################### # Finished running rsync only tests # ##################################### # Copy back everything from noup so the restart tests start with a # clean slate. for originalfile in glob.glob(tmpserver + '/noup/*'): shutil.copy(originalfile, os.path.basename(originalfile)) ################################## # Software updater restart tests # ################################## # Start the web server for the first update webserver = run_webserver(tmpserver + '/updater/') # Keep track of whether ps is there (it isn't on Windows) no_ps = False if nonportable.ostype == 'Windows': # If we are running on windows, disable the ps calls. no_ps = True # ps works different on a mac, where we need to use 'ps -aww' instead of # 'ps -ef'. if nonportable.ostype == 'Darwin': pscommand = 'ps -aww' else: pscommand = 'ps -ef' updateprocess = subprocess.Popen(['python', 'softwareupdater.py']) if not no_ps: # Only do the ps check if ps is available ps = subprocess.Popen(pscommand + ' | grep "softwareupdater.py" | grep -v grep', shell=True, stdout=subprocess.PIPE) psout = ps.stdout.read() print 'Initial ps out:' print psout if psout == '': print 'Failure to start initially' # Wait for 2 minutes for the update to happen and the # process to die. for junk in range(60): if updateprocess.poll() != None: break time.sleep(2) ret = updateprocess.returncode if ret != 10: if ret == None: raise Exception("First software updater never died.") elif ret == 55: raise Exception("Software updater failed to get the process lock.") elif ret == 1: raise Exception("Softwareupdater failed with an uncaught exception. \n\ See end of softwareupdater.old in v2 for details.") else: raise Exception("Unknown return code from the software updater (" + str(ret) + ")") else: print "Old softwareupdater returned correctly" if not no_ps: # Only do the ps check if ps is available ps = subprocess.Popen(pscommand + ' | grep "softwareupdater.py" | grep -v grep', shell=True, stdout=subprocess.PIPE) psout = ps.stdout.read() psout.strip() print 'After ps out:' print psout if psout == '': print 'New updater failed to start!' else: print 'softwareupdater restart success!' # We need to kill the webserver serving from /updater, and start one serving # from updater_new kill_webserver(webserver.pid, updateurl) webserver = run_webserver(tmpserver + '/updater_new/') print "Waiting 2 minutes for the second update to happen" # Wait 2 minutes for the second update to happen. # Is there a way to get a handle for the new softwareupdater? time.sleep(120) # If nmmain's version has been updated, the second update was a success! nmmainfile = file('nmmain.py', 'r') nmmaindata = nmmainfile.read() nmmainfile.close() if 'version = "1234"' in nmmaindata: print 'Second update a success!' else: print 'Second update failed to happen within 2 minutes' # Kill the webserver again now that we are all done with it. kill_webserver(webserver.pid, updateurl) ###################################### # End Software updater restart tests # ###################################### except: traceback.print_exc() finally: # Clean up the temporary server directory. # This throws an exception on Windows if a file is still in use (for example, # if the webserver started in a test couldn't be shut down). shutil.rmtree(tmpserver)
1da3ab38f9ef6428f3669f460f649f1680e24e21 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7263/1da3ab38f9ef6428f3669f460f649f1680e24e21/test_updater_local.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 468, 9190, 326, 1089, 880, 358, 2537, 506, 326, 2359, 1758, 434, 333, 5228, 18, 2359, 273, 336, 4811, 625, 1435, 1089, 718, 273, 296, 2505, 8931, 397, 2359, 397, 4290, 2138,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 2774, 13332, 468, 9190, 326, 1089, 880, 358, 2537, 506, 326, 2359, 1758, 434, 333, 5228, 18, 2359, 273, 336, 4811, 625, 1435, 1089, 718, 273, 296, 2505, 8931, 397, 2359, 397, 4290, 2138,...
elif self.input_dir and os.path.isfile(self.input_dir):
""" if self.input_dir and os.path.isfile(self.input_dir):
def run(self): """ 2008-04-25 return None if QC_method_id==0 2008-04-20 for plone to call it just to get row_id2NA_mismatch_rate """ #database connection and etc db = Stock_250kDB.Stock_250kDB(drivername=self.drivername, username=self.user, password=self.passwd, hostname=self.hostname, database=self.dbname) db.setup() session = db.session session.begin() #transaction = session.create_transaction() self.cmp_data_filename = self.findOutCmpDataFilename(self.cmp_data_filename, self.QC_method_id, Stock_250kDB.QCMethod) import MySQLdb conn = MySQLdb.connect(db=self.dbname, host=self.hostname, user = self.user, passwd = self.passwd) curs = conn.cursor() self.curs = curs if self.debug: import pdb pdb.set_trace() readme = formReadmeObj(sys.argv, self.ad, README) session.save(readme) QC_method_id2snps_table = self.QC_method_id2snps_table if self.QC_method_id==0: self.cal_independent_NA_rate(db, self.min_probability, readme) row_id2NA_mismatch_rate = None else: #from variation.src.FilterStrainSNPMatrix import FilterStrainSNPMatrix header, strain_acc_list, category_list, data_matrix = read_data(self.cmp_data_filename) strain_acc_list = map(int, strain_acc_list) #it's ecotypeid, cast it to integer to be compatible to the later ecotype_id_ls from db snpData2 = SNPData(header=header, strain_acc_list=strain_acc_list, \ data_matrix=data_matrix, snps_table=QC_method_id2snps_table.get(self.QC_method_id)) #category_list is not used. if self.input_dir and os.path.isdir(self.input_dir): #04/22/08 Watch: call_info_id2fname here is fake, it's actually keyed by (array_id, ecotypeid) #no submission to db call_info_id2fname = self.get_array_id2fname(curs, self.input_dir) elif self.input_dir and os.path.isfile(self.input_dir): #it's file call_info_id2fname = None else: if self.run_type==2: #no filtering on call_info entries that have been QCed. filter_calls_QCed=0 elif self.run_type==1: filter_calls_QCed = 1 self.max_call_info_mismatch_rate = 1 #don't use this when doing accession-wise QC else: sys.stderr.write("run_type=%s is not supported.\n"%self.run_type) sys.exit(5) call_info_id2fname, call_info_ls_to_return = self.get_call_info_id2fname(db, self.QC_method_id, self.call_method_id, \ filter_calls_QCed, self.max_call_info_mismatch_rate, self.debug,\ min_no_of_non_NA_pairs=self.min_no_of_non_NA_pairs) if self.run_type==2: snps_name2snps_id = self.get_snps_name2snps_id(db) else: snps_name2snps_id = None if call_info_id2fname: if self.one_by_one and self.run_type==1: #one_by_one only for QC by accession row_id2NA_mismatch_rate = {} row_id12row_id2 = {} counter = 0 for call_info_id, value in call_info_id2fname.iteritems(): counter += 1 print "No", counter tmp_dict = {} tmp_dict[call_info_id] = value pdata = self.read_call_matrix(tmp_dict, self.min_probability) passingdata = self.qcDataMatrixVSsnpData(pdata, snps_name2snps_id, snpData2, curs, session, readme) row_id2NA_mismatch_rate.update(passingdata.row_id2NA_mismatch_rate) row_id12row_id2.update(passingdata.row_id12row_id2) del pdata if self.debug and counter==10: break else: pdata = self.read_call_matrix(call_info_id2fname, self.min_probability) passingdata = self.qcDataMatrixVSsnpData(pdata, snps_name2snps_id, snpData2, curs, session, readme) row_id2NA_mismatch_rate = passingdata.row_id2NA_mismatch_rate row_id12row_id2 = passingdata.row_id12row_id2 del pdata else: #input file is SNP by strain format. double header (1st two lines) header, snps_name_ls, category_list, data_matrix = FilterStrainSNPMatrix.read_data(self.input_dir, double_header=1) pdata = PassingData() pdata.ecotype_id_ls = header[0][2:] pdata.call_info_id_ls = header[1][2:] data_matrix = numpy.array(data_matrix) pdata.data_matrix = data_matrix.transpose() pdata.header = ['', ''] + snps_name_ls #fake a header for SNPData passingdata = self.qcDataMatrixVSsnpData(pdata, snps_name2snps_id, snpData2, curs, session, readme) row_id2NA_mismatch_rate = passingdata.row_id2NA_mismatch_rate row_id12row_id2 = passingdata.row_id12row_id2 del pdata if self.output_fname and self.run_type==1 and row_id2NA_mismatch_rate: self.output_row_id2NA_mismatch_rate(row_id2NA_mismatch_rate, self.output_fname) if self.run_type==1 and self.commit and not self.input_dir and row_id2NA_mismatch_rate: #if self.input_dir is given, no db submission. call_info_id2fname here is fake, it's actually keyed by (array_id, ecotypeid) #row_id2NA_mismatch_rate might be None if it's method 0. self.submit_to_call_QC(session, row_id2NA_mismatch_rate, self.QC_method_id, self.user, self.min_probability, \ row_id12row_id2, self.call_method_id, readme) if self.commit: curs.execute("commit") session.commit() else: session.rollback() self.row_id2NA_mismatch_rate = row_id2NA_mismatch_rate #for plone to get the data structure
e8a8bb38f1f872fe801dbb73dd4dc7dbd34bb539 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9645/e8a8bb38f1f872fe801dbb73dd4dc7dbd34bb539/QC_250k.py
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 4672, 3536, 4044, 28, 17, 3028, 17, 2947, 327, 599, 309, 2238, 39, 67, 2039, 67, 350, 631, 20, 4044, 28, 17, 3028, 17, 3462, 364, 886, 476, 358, 745, 518, 2537, 358, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ 1, 8585, 326, 22398, 316, 326, 981, 30, 1652, 1086, 12, 2890, 4672, 3536, 4044, 28, 17, 3028, 17, 2947, 327, 599, 309, 2238, 39, 67, 2039, 67, 350, 631, 20, 4044, 28, 17, 3028, 17, 3462, 364, 886, 476, 358, 745, 518, 2537, 358, ...