bugged
stringlengths
4
228k
fixed
stringlengths
0
96.3M
__index_level_0__
int64
0
481k
def getSentence(self): return self.question + self.nugget_list[random.randint(0, len(self.nugget_list))]
def getSentence(self): return self.question + self.nugget_list[random.randint(0, len(self.nugget_list))]
475,100
def __init__(self, archive_num): self.question = "Did you know, " #archive_max checked 13-8-2009 self.archive_max = 246 self.archive_num = archive_num if self.archive_num > self.archive_max: self.archive_num = random.randint(1, self.archive_max) self.url = "http://en.wikipedia.org/wiki/Wikipedia:Recent_additions_" + st...
def __init__(self, archive_num): self.question = "Did you know, " #archive_max checked 13-8-2009 self.archive_max = 255 self.archive_num = archive_num if self.archive_num > self.archive_max: self.archive_num = random.randint(1, self.archive_max) self.url = "http://en.wikipedia.org/wiki/Wikipedia:Recent_additions_" + st...
475,101
def buildProtocol(self, address): if re.match("[^a-z]+", address.host): log.error("Kludge fix for twisted.words weirdness") fqdn = socket.getfqdn(address.host) address = (fqdn, address.port) else: address = (address.host, address.port)
def buildProtocol(self, address): if re.match("[^a-z]+", address.host): log.error("Kludge fix for twisted.words weirdness") fqdn = socket.getfqdn(address.host) address = (fqdn, address.port) else: address = (address.host, address.port)
475,102
def cb(tries): # Ack timeout callback
def cb(tries): # Ack timeout callback
475,103
def broadcastHeader(self, kind, src_ipp, hops=64, flags=0): # Build the header used for all broadcast packets packet = [kind] packet.append(self.main.osm.me.ipp) packet.append(struct.pack('!BB', hops, flags)) packet.append(src_ipp) return packet
def broadcastHeader(self, kind, src_ipp, hops=32, flags=0): # Build the header used for all broadcast packets packet = [kind] packet.append(self.main.osm.me.ipp) packet.append(struct.pack('!BB', hops, flags)) packet.append(src_ipp) return packet
475,104
def __init__(self, body = None): SipGenericHF.__init__(self, body) self.parsed = True if body == None: self.body = md5(str((random() * 1000000000L) + time())).hexdigest() + '@' + SipConf.my_address
def __init__(self, body = None): SipGenericHF.__init__(self, body) self.parsed = True if body == None: self.body = md5(str((random() * 1000000000L) + time())).hexdigest() + '@' + SipConf.my_address
475,105
def genCallId(self): self.body = md5(str((random() * 1000000000L) + time())).hexdigest() + '@' + SipConf.my_address
def genCallId(self): self.body = md5(str((random() * 1000000000L) + time())).hexdigest() + '@' + SipConf.my_address
475,106
def recvResponse(self, resp): body = resp.getBody() code, reason = resp.getSCode() scode = (code, reason, body) if self.ua.no_reply_timer != None: self.ua.no_reply_timer.cancel() self.ua.no_reply_timer = None if code == 100 and self.ua.no_progress_time != None: self.ua.no_progress_timer = TimeoutAbs(self.ua.no_progress...
def recvResponse(self, resp): body = resp.getBody() code, reason = resp.getSCode() scode = (code, reason, body) if self.ua.no_reply_timer != None: self.ua.no_reply_timer.cancel() self.ua.no_reply_timer = None if code == 100 and self.ua.no_progress_time != None: self.ua.no_progress_timer = TimeoutAbs(self.ua.no_progress...
475,107
def recvResponse(self, resp): body = resp.getBody() code, reason = resp.getSCode() scode = (code, reason, body) if code < 200: self.ua.equeue.append(CCEventRing(scode, rtime = resp.rtime, origin = self.ua.origin)) return None if code >= 200 and code < 300: event = CCEventConnect(scode, rtime = resp.rtime, origin = self...
def recvResponse(self, resp): body = resp.getBody() code, reason = resp.getSCode() scode = (code, reason, body) if code < 200: self.ua.equeue.append(CCEventRing(scode, rtime = resp.rtime, origin = self.ua.origin)) return None if code >= 200 and code < 300: event = CCEventConnect(scode, rtime = resp.rtime, origin = self...
475,108
def mmrrc(): mmrrcfp = reportlib.init(sys.argv[0], outputdir = os.environ['QCOUTPUTDIR'], fileExt = '.mmrrc.rpt') title = 'MMRRC Strains w/ Genotype Associations where the Markers/Alleles of the Strain record\n' + \ 'do not exactly match the Markers/Alleles of the Genotype record.' mmncfp.write(title + '\n\n') mmncf...
def mmrrc(): mmrrcfp = reportlib.init(sys.argv[0], outputdir = os.environ['QCOUTPUTDIR'], fileExt = '.mmrrc.rpt') title = 'MMRRC Strains w/ Genotype Associations where the Markers/Alleles of the Strain record\n' + \ 'do not exactly match the Markers/Alleles of the Genotype record.' mmncfp.write(title + '\n\n') mmncf...
475,109
def to_point(self): '''Convert from PolarPoint to (cartesian) Point object'''
def to_point(self): '''Convert from PolarPoint to (cartesian) Point object'''
475,110
def connect_action(self, event):
def connect_action(self, event):
475,111
def connect_action(self, event):
def connect_action(self, event):
475,112
def connect_action(self, event):
def connect_action(self, event):
475,113
def connect_action(self, event):
def connect_action(self, event):
475,114
def is_point(self, line):
def is_point(self, line):
475,115
def _points(self): points = [] for row in rows: fs = row.split(',') if fs[0] == 'ST': x = fs[-3] # FIXME NEZ coord order shouldn't be hardcoded y = fs[-2] z = fs[-1] bp = BasePoint(x=x, y=y, z=z, ih=0) if fs[0] == 'SS': angle = fs[4] z_angle = fs[5] dist = fs[3] th = fs[2] p = PolarPoint(dist=dist, angle=angle, z_ang...
def _points(self): points = [] for row in rows: fs = row.split(',') if fs[0] == 'ST': x = fs[-3] # FIXME NEZ coord order shouldn't be hardcoded y = fs[-2] z = fs[-1] bp = BasePoint(x=x, y=y, z=z, ih=0) if fs[0] == 'SS': angle = fs[4] z_angle = fs[5] dist = fs[3] th = fs[2] p = PolarPoint(dist=dist, angle=angle, z_ang...
475,116
def url(dialog, link, data=None): pass
def url(dialog, link, data=None): pass
475,117
def _check_for_Flash(self): try: tmp = "/tmp" ## look for filenames that begin with 'Flash' for file in os.listdir(tmp): try: if file.startswith("Flash"): filepath = os.path.join(tmp, file) duration = utils.getFLVLength(filepath)
def _check_for_Flash(self): try: tmp = "/tmp" ## look for filenames that begin with 'Flash' for file in os.listdir(tmp): try: if file.startswith("Flash"): filepath = os.path.join(tmp, file) duration = utils.getFLVLength(filepath)
475,118
def _check_for_Flash(self): try: tmp = "/tmp" ## look for filenames that begin with 'Flash' for file in os.listdir(tmp): try: if file.startswith("Flash"): filepath = os.path.join(tmp, file) duration = utils.getFLVLength(filepath)
def _check_for_Flash(self): try: tmp = "/tmp" ## look for filenames that begin with 'Flash' for file in os.listdir(tmp): try: if file.startswith("Flash"): filepath = os.path.join(tmp, file) duration = utils.getFLVLength(filepath)
475,119
def __init__(self): self.Core = core.Caffeine()
def __init__(self): self.Core = core.Caffeine()
475,120
def on_about_button_clicked (self, button, data=None):
def on_about_button_clicked (self, button, data=None):
475,121
def on_time_submenuitem_activate(self, menuitem, time):
def on_time_submenuitem_activate(self, menuitem, time):
475,122
def _run_dialog(self): response = self.about_dialog.run() self.about_dialog.destroy() return False
def on_about_menuitem_activate(self, menuitem, data=None): if appindicator_avail: gtk.gdk.threads_enter() self.about_dialog.set_position (gtk.WIN_POS_CENTER_ALWAYS) response = self.about_dialog.run() self.about_dialog.destroy() return False
475,123
def _run_dialog(self): response = self.about_dialog.run() self.about_dialog.destroy() return False
def _run_dialog(self): response = self.about_dialog.run() self.about_dialog.destroy() return False
475,124
def quit(self): ### Do anything that needs to be done before quitting. logging.info("Caffeine is preparing to quit")
def quit(self): ### Do anything that needs to be done before quitting. logging.info("Caffeine is preparing to quit")
475,125
def os_chown(self, path, uid, gid): ino = self.ino_from_path(path) inode = self.inode_read(ino) inode["uid"] = uid inode["gid"] = gid self.inode_write(ino, inode)
def os_chown(self, path, uid, gid): ino = self.ino_from_path(path) if uid == -1 and gid == -1: return inode = self.inode_read(ino) if uid != -1: inode["uid"] = uid if gid != -1: inode["gid"] = gid if inode["mode"] & (S_IXUSR|S_IXGRP|S_IXOTH): inode["mode"] &= ~(S_ISUID|S_ISGID) inode.set_time_fields(change=True) self.i...
475,126
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]...
def _rmdir(self, pino, name): offset, dirent = self._scandir(pino, name) ino = dirent["ino"] inode = self.inode_read(ino) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent[...
475,127
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]...
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"], inode["nlink"]) # free all the clusters of the d...
475,128
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]...
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", ino, inode["n...
475,129
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]...
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]...
475,130
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]...
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]...
475,131
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]...
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]...
475,132
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values)
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values)
475,133
def __init__(self, errno): self.errno = errno
def __init__(self, errno): self.errno = errno
475,134
def create_node(self, path, type): #pdb.set_trace() node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(paren...
defcreate_node(self,path,type):#pdb.set_trace()node_dirname,node_basename=os.path.split(path)parent_dirname,parent_basename=os.path.split(node_dirname)parent_dirent=self.get_dir_entry(node_dirname)parent_inode_struct=self.get_inode_struct(parent_dirent["inode"])fordirentinself.read_directory(parent_dirent["inode"]):ifd...
475,135
def create_node(self, path, type): #pdb.set_trace() node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(paren...
def create_node(self, path, type): #pdb.set_trace() node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(paren...
475,136
def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.write("Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim.") with stattab_table_tag(doc.open_tag)...
def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.add_tag("p", data="Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim. Shown first are those that...
475,137
def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.write("Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim.") with stattab_table_tag(doc.open_tag)...
def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.write("Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim.") with stattab_table_tag(doc.open_tag)...
475,138
def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.write("Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim.") with stattab_table_tag(doc.open_tag)...
def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.write("Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim.") with stattab_table_tag(doc.open_tag)...
475,139
def curses_color(): import curses class TermInfo(): # this. is. ANSIIIII!!! _ANSI_COLORS = """BLACK RED GREEN YELLOW BLUE MAGENTA CYAN WHITE""".split() _STRING_CAPS = """NORMAL=sgr0""".split() def __init__(self, stream=sys.stdout): # isatty might be needed here? curses.setupterm(None, stream.fileno()) for prefix, capna...
def curses_color(): import curses class TermInfo(): # this. is. ANSIIIII!!! _ANSI_COLORS = """BLACK RED GREEN YELLOW BLUE MAGENTA CYAN WHITE""".split() _STRING_CAPS = """NORMAL=sgr0""".split() def __init__(self, stream=sys.stdout): # isatty might be needed here? curses.setupterm(None, stream.fileno()) for prefix, capna...
475,140
def select_color(): try: import curses except ImportError: import ctypes try: from ctypes import wintypes finally: pass return windows_color() else: return curses_color() assert False
def select_color(): for a in [curses_color, windows_color]: b = a() if b != None: return b else: return curses_color() assert False
475,141
def select_color(): try: import curses except ImportError: import ctypes try: from ctypes import wintypes finally: pass return windows_color() else: return curses_color() assert False
def select_color(): try: import curses except ImportError: import ctypes try: from ctypes import wintypes finally: pass return windows_color() else: assert False
475,142
def __get__(self, instance, owner): #assert instance == None, instance return property.__get__(self, owner)
def__get__(self,instance,owner):#assertinstance==None,instancereturnproperty.__get__(self,owner)
475,143
def ClfsStructField(*args): a = list(args) if len(a) < 3: a.append(None) else: assert a[2] != None return collections.namedtuple( "ClfsStructField", ("name", "format", "initval") )(*a)
def ClfsStructField(*args): a = list(args) if len(a) < 3: a.append(None) else: assert a[2] != None, "None is reserved for indicating uninitialized fields" return namedtuple( "ClfsStructField", ("name", "format", "initval") )(*a)
475,144
def all_fields(): for base in bases: if hasattr(base, "fields"): for field in base.fields: yield field for a in attrs["_fields_"]: yield ClfsStructField(*a)
def all_fields(): for base in bases: if hasattr(base, "fields"): for field in base.fields: yield field for a in attrs["_fields_"]: yield ClfsStructField(*a)
475,145
def from_fileobj(class_, fileobj): return class_.unpack(fileobj.read(class_.size))
def from_fileobj(class_, fileobj): return class_.unpack(fileobj.read(class_.size))
475,146
def unpack(class_, buffer): instance = class_() offset = 0 for field in instance.fields: #assert field.name in instance.__values unpacked = struct.unpack_from(field.format, buffer, offset) assert len(unpacked) == 1 instance[field.name] = unpacked[0] #instance.__values[field.name] = unpacked[0] offset += struct.calcsize...
def unpack(class_, buffer): instance = class_() offset = 0 for field in instance.fields: #assert field.name in instance.__values unpacked = struct.unpack_from(field.format, buffer, offset) if len(unpacked): instance[field.name], = unpacked #instance.__values[field.name] = unpacked[0] offset += struct.calcsize(field.for...
475,147
def pack(self): buffer = "" for field in self.fields: assert field.name in self.__values, "Field %r is uninitialized" % (field.name,) value = self.__values[field.name] try: buffer += struct.pack(field.format, value) except struct.error as exc: raise struct.error("Error packing %r into %s.%s, %s" % (value, self.__class_...
def pack(self): buffer = "" for field in self.fields: try: buffer += struct.pack(field.format, value) except struct.error as exc: raise struct.error("Error packing %r into %s.%s, %s" % (value, self.__class__.__name__, field.name, exc.message)) assert len(buffer) == self.size return buffer
475,148
def pack(self): buffer = "" for field in self.fields: assert field.name in self.__values, "Field %r is uninitialized" % (field.name,) value = self.__values[field.name] try: buffer += struct.pack(field.format, value) except struct.error as exc: raise struct.error("Error packing %r into %s.%s, %s" % (value, self.__class_...
def pack(self): buffer = "" for field in self.fields: assert field.name in self.__values, "Field %r is uninitialized" % (field.name,) value = self.__values[field.name] try: value = (self.__values[field.name],) except KeyError: value = () try: buffer += struct.pack(field.format, *value) except struct.error as exc: raise...
475,149
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values)
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values)
475,150
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values)
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values)
475,151
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values)
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values)
475,152
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values)
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values)
475,153
def __init__(self, path=None, fileobj=None): assert bool(path) ^ bool(fileobj) if path: self.f = open(path, "r+b") else: self.f = fileobj self.f.seek(0) br = BootRecord.from_fileobj(self.f) assert br["ident"].rstrip("\0") == "clfs", repr(br["ident"]) assert br["version"] == 1 self.cluster_size = br["clrsize"] self.mast...
def __init__(self, path): self.f = open(path, "r+b") br = BootRecord.from_fileobj(self.f) assert br["ident"].rstrip("\0") == "clfs", repr(br["ident"]) assert br["version"] == 1 self.cluster_size = br["clrsize"] self.master_region_cluster_count = br["mstrclrs"] self.allocation_table_cluster_count = br["atabclrs"] self.d...
475,154
def __init__(self, path=None, fileobj=None): assert bool(path) ^ bool(fileobj) if path: self.f = open(path, "r+b") else: self.f = fileobj self.f.seek(0) br = BootRecord.from_fileobj(self.f) assert br["ident"].rstrip("\0") == "clfs", repr(br["ident"]) assert br["version"] == 1 self.cluster_size = br["clrsize"] self.mast...
def __init__(self, path=None, fileobj=None): assert bool(path) ^ bool(fileobj) if path: self.f = open(path, "r+b") else: self.f = fileobj self.f.seek(0) br = BootRecord.from_fileobj(self.f) assert br["ident"].rstrip("\0") == "clfs", repr(br["ident"]) assert br["version"] == 1 self.cluster_size = br["clrsize"] self.mast...
475,155
def read_directory(self, inode): inode_struct = self.get_inode_struct(inode) if inode_struct["type"] != TYPE_DIRECTORY: raise ClfsError(ENOTDIR) offset = 0 while offset < inode_struct["size"]: dirent = DirEntry.unpack(self.read_inode_data( inode, offset, DirEntry.size)) if dirent["name"].rstrip("\0"): yield dirent offs...
def read_directory(self, inode): inode_struct = self.get_inode_struct(inode) if not S_ISDIR(inode_struct["mode"]): raise ClfsError(ENOTDIR) offset = 0 while offset < inode_struct["size"]: dirent = DirEntry.unpack(self.read_inode_data( inode, offset, DirEntry.size)) if dirent["name"].rstrip("\0"): yield dirent offset +=...
475,156
def get_dir_entry(self, path): for name in path.split("/"): if not name: cur_dirent = self.get_root_dir_entry() else: # pdb.set_trace() for dirent in self.read_directory(cur_dirent["inode"]): if dirent["name"].rstrip("\0") == name: cur_dirent = dirent break else: raise ClfsError(ENOENT) return cur_dirent
def get_dir_entry(self, path): for name in path.split("/"): if not name: cur_dirent = self.get_root_dir_entry() else: # pdb.set_trace() for dirent in self.read_directory(cur_dirent["ino"]): if dirent["name"].rstrip("\0") == name: cur_dirent = dirent break else: raise ClfsError(ENOENT) return cur_dirent
475,157
def get_dir_entry(self, path): for name in path.split("/"): if not name: cur_dirent = self.get_root_dir_entry() else: # pdb.set_trace() for dirent in self.read_directory(cur_dirent["inode"]): if dirent["name"].rstrip("\0") == name: cur_dirent = dirent break else: raise ClfsError(ENOENT) return cur_dirent
def get_dir_entry(self, path): for name in path.split("/"): if not name: cur_dirent = self.get_root_dir_entry() else: # pdb.set_trace() for dirent in self.read_directory(cur_dirent["inode"]): if dirent["name"].rstrip("\0") == name: cur_dirent = dirent break else: raise ClfsError(ENOENT) return cur_dirent
475,158
#def write(self, path, buf, offset):
#def write(self, path, buf, offset):
475,159
def write_inode_data(self, inode, offset, buffer): inode_struct = self.get_inode_struct(inode) data_offset = inode_struct.size write_size, new_size = self.write_to_chain( inode, inode_struct["size"] + data_offset, offset + data_offset, buffer) assert write_size == len(buffer), write_size expected_size = data_offset + m...
def write_inode_data(self, ino, offset, buffer): inode_struct = self.get_inode_struct(inode) data_offset = inode_struct.size write_size, new_size = self.write_to_chain( ino, inode_struct["size"] + data_offset, offset + data_offset, buffer) assert write_size == len(buffer), write_size expected_size = data_offset + max(o...
475,160
def write_inode_data(self, inode, offset, buffer): inode_struct = self.get_inode_struct(inode) data_offset = inode_struct.size write_size, new_size = self.write_to_chain( inode, inode_struct["size"] + data_offset, offset + data_offset, buffer) assert write_size == len(buffer), write_size expected_size = data_offset + m...
def write_inode_data(self, inode, offset, buffer): inode_struct = self.get_inode_struct(inode) data_offset = inode_struct.size write_size, new_size = self.write_to_chain( inode, inode_struct["size"] + data_offset, offset + data_offset, buffer) assert write_size == len(buffer), write_size expected_size = data_offset + m...
475,161
def create_node(self, path, type): node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(parent_dirent["inode"]...
def create_node(self, path, mode): """Create an allocate a new inode, update relevant structures elsewhere""" node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(pa...
475,162
def create_node(self, path, type): node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(parent_dirent["inode"]...
def create_node(self, path, type): node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) create_rootdir = bool( (not node_basename) and (node_dirname == parent_dirname == "/")) if create_rootdir: assert S_ISDIR(mode) new_inode = Inode(size=0, uid=0, gid=0, rde...
475,163
def create_node(self, path, type): node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(parent_dirent["inode"]...
def create_node(self, path, type): node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(parent_dirent["inode"]...
475,164
def add_share(self, button): namelbl = gtk.Label("Share name:")
def add_share(self, button): namelbl = gtk.Label("Share name:")
475,165
def browse_peer_by_url(self, url): """Open the given peer URL with the most natural file manager for the current platform that we can find""" import os, subprocess if os.name == "nt": # this is how it's done programmatically? except that it won't invoke # the default file manager (explorer) on winders #ShellExecute(Non...
def browse_peer_by_url(self, url): """Open the given peer URL with the most natural file manager for the current platform that we can find""" import os, subprocess if os.name == "nt": # this is how it's done programmatically? except that it won't invoke # the default file manager (explorer) on winders #ShellExecute(Non...
475,166
def browse_peer_by_url(self, url): """Open the given peer URL with the most natural file manager for the current platform that we can find""" import os, subprocess if os.name == "nt": # this is how it's done programmatically? except that it won't invoke # the default file manager (explorer) on winders #ShellExecute(Non...
def browse_peer_by_url(self, url): """Open the given peer URL with the most natural file manager for the current platform that we can find""" import os, subprocess if os.name == "nt": # this is how it's done programmatically? except that it won't invoke # the default file manager (explorer) on winders #ShellExecute(Non...
475,167
def read_from_chain(self, first_cluster, chain_size, read_offset, read_size): if chain_size <= 0: return "" #assert read_offset + read_size <= chain_size, (read_offset, read_size, chain_size) if read_offset > self.cluster_size: return self.read_from_chain( self.next_cluster(first_cluster), chain_size - self.cluster_siz...
def read_from_chain(self, first_cluster, chain_size, read_offset, read_size): if chain_size <= 0: return "" #assert read_offset + read_size <= chain_size, (read_offset, read_size, chain_size) if read_offset > self.cluster_size: return self.read_from_chain( self.next_cluster(first_cluster), chain_size - self.cluster_siz...
475,168
def trans_include(repo_uri, fargs, transaction=None): basedirs = [] timestamp_files = [] error_occurred = False opts, pargs = getopt.getopt(fargs, "d:T:") for opt, arg in opts: if opt == "-d": basedirs.append(arg) elif opt == "-T": timestamp_files.append(arg) if transaction == None: try: trans_id = os.environ["PKG_TR...
def trans_include(repo_uri, fargs, transaction=None): basedirs = [] timestamp_files = [] error_occurred = False opts, pargs = getopt.getopt(fargs, "d:T:") for opt, arg in opts: if opt == "-d": basedirs.append(arg) elif opt == "-T": timestamp_files.append(arg) if transaction == None: try: trans_id = os.environ["PKG_TR...
475,169
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header."""
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header."""
475,170
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header."""
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header."""
475,171
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header."""
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header."""
475,172
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header."""
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header."""
475,173
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header."""
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header."""
475,174
def run(self): # nuke everything print("deleting " + dist_dir) shutil.rmtree(dist_dir, True) print("deleting " + build_dir) shutil.rmtree(build_dir, True) print("deleting " + root_dir) shutil.rmtree(root_dir, True) print("deleting " + pkgs_dir) shutil.rmtree(pkgs_dir, True) print("deleting " + extern_dir) shutil.rmtree...
def run(self): # nuke everything print("deleting " + dist_dir) shutil.rmtree(dist_dir, True) print("deleting " + build_dir) shutil.rmtree(build_dir, True) print("deleting " + root_dir) shutil.rmtree(root_dir, True) print("deleting " + pkgs_dir) shutil.rmtree(pkgs_dir, True) print("deleting " + extern_dir) shutil.rmtree...
475,175
def initialize_options(self): self.only = "" self.baselinefile = "" self.verbosemode = 0 self.parseable = 0 self.genbaseline = 0 self.timing = 0 self.coverage = 0 self.stoponerr = 0 self.debugoutput = 0 self.show_on_expected_fail = 0 self.startattest = "" self.archivedir = ""
def initialize_options(self): self.only = "" self.baselinefile = "" self.verbosemode = 0 self.parseable = 0 self.genbaseline = 0 self.timing = 0 self.coverage = 0 self.stoponerr = 0 self.debugoutput = 0 self.showonexpectedfail = 0 self.startattest = "" self.archivedir = ""
475,176
def main_func(): global cache_dir, download_start, xport, xport_cfg all_timestamps = False all_versions = False keep_compressed = False list_newest = False recursive = False src_uri = None target = None incoming_dir = None src_pub = None targ_pub = None temp_root = misc.config_temp_root() gettext.install("pkg", "/usr...
def main_func(): global cache_dir, download_start, xport, xport_cfg all_timestamps = False all_versions = False keep_compressed = False list_newest = False recursive = False src_uri = None target = None incoming_dir = None src_pub = None targ_pub = None temp_root = misc.config_temp_root() gettext.install("pkg", "/usr...
475,177
def main_func(): global cache_dir, download_start, xport, xport_cfg all_timestamps = False all_versions = False keep_compressed = False list_newest = False recursive = False src_uri = None target = None incoming_dir = None src_pub = None targ_pub = None temp_root = misc.config_temp_root() gettext.install("pkg", "/usr...
def main_func(): global cache_dir, download_start, xport, xport_cfg all_timestamps = False all_versions = False keep_compressed = False list_newest = False recursive = False src_uri = None target = None incoming_dir = None src_pub = None targ_pub = None temp_root = misc.config_temp_root() gettext.install("pkg", "/usr...
475,178
def get_basename(pfmri): open_time = pfmri.get_timestamp() return "%d_%s" % \ (calendar.timegm(open_time.utctimetuple()), urllib.quote(str(pfmri), ""))
def get_basename(pfmri): open_time = pfmri.get_timestamp() return "%d_%s" % \ (calendar.timegm(open_time.utctimetuple()), urllib.quote(str(pfmri), ""))
475,179
def get_publisherinfo(self, pub, ccancel=None): """Given a publisher pub, return the publisher/0 information in a StringIO object."""
def get_publisherinfo(self, pub, ccancel=None): """Given a publisher pub, return the publisher/0 information in a StringIO object."""
475,180
def main_func(): global file_repo global def_branch global def_repo global def_vers global extra_entire_contents global just_these_pkgs global not_these_pkgs global nopublish global publish_all global print_pkg_names global reference_uris global show_debug global wos_path global not_these_consolidations global curpkg ...
def main_func(): global file_repo global def_branch global def_repo global def_vers global extra_entire_contents global just_these_pkgs global not_these_pkgs global nopublish global publish_all global print_pkg_names global reference_uris global show_debug global wos_path global not_these_consolidations global curpkg ...
475,181
def get_smf_packages(server_url, manifest_locations, filter): """ Performs a search against server_url looking for packages which contain SMF manifests, returning a list of those pfmris """ dir = os.getcwd() tracker = pkg.client.progress.QuietProgressTracker() image_dir = tempfile.mkdtemp("", "pkg_importer_smfsearch."...
def get_smf_packages(server_url, manifest_locations, filter): """ Performs a search against server_url looking for packages which contain SMF manifests, returning a list of those pfmris """ dir = os.getcwd() tracker = pkg.client.progress.QuietProgressTracker() image_dir = tempfile.mkdtemp("", "pkg_importer_smfsearch."...
475,182
def get_smf_packages(server_url, manifest_locations, filter): """ Performs a search against server_url looking for packages which contain SMF manifests, returning a list of those pfmris """ dir = os.getcwd() tracker = pkg.client.progress.QuietProgressTracker() image_dir = tempfile.mkdtemp("", "pkg_importer_smfsearch."...
def get_smf_packages(server_url, manifest_locations, filter): """ Performs a search against server_url looking for packages which contain SMF manifests, returning a list of those pfmris """ dir = os.getcwd() tracker = pkg.client.progress.QuietProgressTracker() image_dir = tempfile.mkdtemp("", "pkg_importer_smfsearch."...
475,183
def set_function_fields(self, cursor, user, id, name, value, arg, context=None): request_obj = self.pool.get('res.request') req_ref_obj = self.pool.get('res.request.reference')
def set_function_fields(self, cursor, user, id, name, value, arg, context=None): request_obj = self.pool.get('res.request') req_ref_obj = self.pool.get('res.request.reference')
475,184
def add_minutes(self, cursor, user, company, date, minutes, context=None): minutes = int(round(minutes)) minutes = date.minute + minutes
def add_minutes(self, cursor, user, company, date, minutes, context=None): minutes = int(round(minutes)) minutes = date.minute + minutes
475,185
def add_hours(self, cursor, user, company, date, hours, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day
def add_hours(self, cursor, user, company, date, hours, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day
475,186
def add_days(self, cursor, user, company, date, days, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day
def add_days(self, cursor, user, company, date, days, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day
475,187
def add_weeks(self, cursor, user, company, date, weeks, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day
def add_weeks(self, cursor, user, company, date, weeks, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day
475,188
def get_function_fields(self, ids, names): ''' Function to compute function fields
def get_function_fields(self, ids, names): ''' Function to compute function fields
475,189
def test0005views(self): ''' Test views. ''' self.assertRaises(Exception, test_view('project_plan'))
def test0005views(self): ''' Test views. ''' self.assertRaises(Exception, test_view('project_plan'))
475,190
def set_function_fields(self, cursor, user, ids, name, value, context=None): request_obj = self.pool.get('res.request') req_ref_obj = self.pool.get('res.request.reference')
def set_function_fields(self, cursor, user, ids, name, value, context=None): request_obj = self.pool.get('res.request') req_ref_obj = self.pool.get('res.request.reference')
475,191
def log_job_closed(self): log.info("Job finished: Ticket:%s Task:%s User:%s Start:%f End:%f Time:%f Price:%f State:%s" % (self.ticket(), self.task(), self.getUser(), self.jobStart(), self.jobEnd(), self.getTime(), self.getCost(), self.getState()))
def log_job_closed(self): log.info("Job finished: Ticket:%s Task:%s User:%s Start:%f End:%f Time:%f Price:%f State:%s" % (self.ticket(), self.task(), self.getUser(), self.jobStart(), self.jobEnd(), self.getTime(), self.getCost(), self.getState()))
475,192
def __init__(self, ticket_id, id, mgr): """Initialize a new task.""" TaskFSM.__init__(self) self.mtx = self.fsm.getLock()
def __init__(self, ticket_id, id, mgr): """Initialize a new task.""" TaskFSM.__init__(self) self.mtx = self.fsm.getLock()
475,193
def cleanUp(self): """cleans up the task, i.e. removes the task's spool directory""" self.log.info("removing spool directory") from xbe.util import removeDirCompletely removeDirCompletely(self.__spool)
def cleanUp(self): """cleans up the task, i.e. removes the task's spool directory""" self.log.info("removing spool directory") from xbe.util import removeDirCompletely removeDirCompletely(self.__spool)
475,194
def getState(self): return self.__fsm.getState().getName()
def getState(self): return self.__fsm.getState().getName()
475,195
def do_Event(self, event, reqCtxt): log.debug("JOB '%s' run in state '%s' event '%s'" % (self.ticket(), self.__fsm.getState().getName(), event)) if hasattr(self.__fsm, event): log.debug("Run event '%s'" % event) getattr(self.__fsm, event)(self, reqCtxt) else: log.debug("Event '%s' not found." % event) raise CommandFai...
def do_Event(self, event, reqCtxt): log.debug("JOB '%s' run in state '%s' event '%s'" % (self.ticket(), self.__fsm.getState().getName(), event)) if hasattr(self.__fsm, event): log.debug("Run event '%s'" % event) getattr(self.__fsm, event)(self, reqCtxt) else: log.error("Event '%s' not found." % event) raise CommandFai...
475,196
def do_EventByMap(self, eventkey, reqCtxt): eventMap = { "Pending:Reserved" : 1, "Pending:Confirmed" : "confirm", "Running:Stage-In" : "runJob_StageIn", "Running:Instance-Starting" : "", "Running:Executing" : "runJob_Execute", "Running:Stage-Out" : "runJob_StageOut", "Running:Instance-Stopping" : "", "...
def do_EventByMap(self, eventkey, reqCtxt): eventMap = { "Pending:Reserved" : 1, "Pending:Confirmed" : "confirm", "Running:Stage-In" : "runJob_StageIn", "Running:Instance-Starting" : "", "Running:Executing" : "runJob_Execute", "Running:Stage-Out" : "runJob_StageOut", "Running:Instance-Stopping" : "", "...
475,197
def do_EventByMap(self, eventkey, reqCtxt): eventMap = { "Pending:Reserved" : 1, "Pending:Confirmed" : "confirm", "Running:Stage-In" : "runJob_StageIn", "Running:Instance-Starting" : "", "Running:Executing" : "runJob_Execute", "Running:Stage-Out" : "runJob_StageOut", "Running:Instance-Stopping" : "", "...
def do_EventByMap(self, eventkey, reqCtxt): eventMap = { "Pending:Reserved" : 1, "Pending:Confirmed" : "confirm", "Running:Stage-In" : "runJob_StageIn", "Running:Instance-Starting" : "", "Running:Executing" : "runJob_Execute", "Running:Stage-Out" : "runJob_StageOut", "Running:Instance-Stopping" : "", "...
475,198
def log_job_closed(self): log.info("Job finished: Ticket:%s Task:%s User:%s Start:%f End:%f Time:%f Price:%f State:%s" % (self.ticket(), self.task(), self.getUser(), self.getStart(), self.getEnd(), self.getTime(), self.getCost(), self.getState()))
def log_job_closed(self): log.info("Job finished: Ticket:%s Task:%s User:%s Start:%f End:%f Time:%f Price:%f State:%s" % (self.ticket(), self.task(), self.getUser(), self.getStart(), self.getEnd(), self.getTime(), self.getCost(), self.getState()))
475,199