bugged stringlengths 4 228k | fixed stringlengths 0 96.3M | __index_level_0__ int64 0 481k |
|---|---|---|
def getSentence(self): return self.question + self.nugget_list[random.randint(0, len(self.nugget_list))] | def getSentence(self): return self.question + self.nugget_list[random.randint(0, len(self.nugget_list))] | 475,100 |
def __init__(self, archive_num): self.question = "Did you know, " #archive_max checked 13-8-2009 self.archive_max = 246 self.archive_num = archive_num if self.archive_num > self.archive_max: self.archive_num = random.randint(1, self.archive_max) self.url = "http://en.wikipedia.org/wiki/Wikipedia:Recent_additions_" + st... | def __init__(self, archive_num): self.question = "Did you know, " #archive_max checked 13-8-2009 self.archive_max = 255 self.archive_num = archive_num if self.archive_num > self.archive_max: self.archive_num = random.randint(1, self.archive_max) self.url = "http://en.wikipedia.org/wiki/Wikipedia:Recent_additions_" + st... | 475,101 |
def buildProtocol(self, address): if re.match("[^a-z]+", address.host): log.error("Kludge fix for twisted.words weirdness") fqdn = socket.getfqdn(address.host) address = (fqdn, address.port) else: address = (address.host, address.port) | def buildProtocol(self, address): if re.match("[^a-z]+", address.host): log.error("Kludge fix for twisted.words weirdness") fqdn = socket.getfqdn(address.host) address = (fqdn, address.port) else: address = (address.host, address.port) | 475,102 |
def cb(tries): # Ack timeout callback | def cb(tries): # Ack timeout callback | 475,103 |
def broadcastHeader(self, kind, src_ipp, hops=64, flags=0): # Build the header used for all broadcast packets packet = [kind] packet.append(self.main.osm.me.ipp) packet.append(struct.pack('!BB', hops, flags)) packet.append(src_ipp) return packet | def broadcastHeader(self, kind, src_ipp, hops=32, flags=0): # Build the header used for all broadcast packets packet = [kind] packet.append(self.main.osm.me.ipp) packet.append(struct.pack('!BB', hops, flags)) packet.append(src_ipp) return packet | 475,104 |
def __init__(self, body = None): SipGenericHF.__init__(self, body) self.parsed = True if body == None: self.body = md5(str((random() * 1000000000L) + time())).hexdigest() + '@' + SipConf.my_address | def __init__(self, body = None): SipGenericHF.__init__(self, body) self.parsed = True if body == None: self.body = md5(str((random() * 1000000000L) + time())).hexdigest() + '@' + SipConf.my_address | 475,105 |
def genCallId(self): self.body = md5(str((random() * 1000000000L) + time())).hexdigest() + '@' + SipConf.my_address | def genCallId(self): self.body = md5(str((random() * 1000000000L) + time())).hexdigest() + '@' + SipConf.my_address | 475,106 |
def recvResponse(self, resp): body = resp.getBody() code, reason = resp.getSCode() scode = (code, reason, body) if self.ua.no_reply_timer != None: self.ua.no_reply_timer.cancel() self.ua.no_reply_timer = None if code == 100 and self.ua.no_progress_time != None: self.ua.no_progress_timer = TimeoutAbs(self.ua.no_progress... | def recvResponse(self, resp): body = resp.getBody() code, reason = resp.getSCode() scode = (code, reason, body) if self.ua.no_reply_timer != None: self.ua.no_reply_timer.cancel() self.ua.no_reply_timer = None if code == 100 and self.ua.no_progress_time != None: self.ua.no_progress_timer = TimeoutAbs(self.ua.no_progress... | 475,107 |
def recvResponse(self, resp): body = resp.getBody() code, reason = resp.getSCode() scode = (code, reason, body) if code < 200: self.ua.equeue.append(CCEventRing(scode, rtime = resp.rtime, origin = self.ua.origin)) return None if code >= 200 and code < 300: event = CCEventConnect(scode, rtime = resp.rtime, origin = self... | def recvResponse(self, resp): body = resp.getBody() code, reason = resp.getSCode() scode = (code, reason, body) if code < 200: self.ua.equeue.append(CCEventRing(scode, rtime = resp.rtime, origin = self.ua.origin)) return None if code >= 200 and code < 300: event = CCEventConnect(scode, rtime = resp.rtime, origin = self... | 475,108 |
def mmrrc(): mmrrcfp = reportlib.init(sys.argv[0], outputdir = os.environ['QCOUTPUTDIR'], fileExt = '.mmrrc.rpt') title = 'MMRRC Strains w/ Genotype Associations where the Markers/Alleles of the Strain record\n' + \ 'do not exactly match the Markers/Alleles of the Genotype record.' mmncfp.write(title + '\n\n') mmncf... | def mmrrc(): mmrrcfp = reportlib.init(sys.argv[0], outputdir = os.environ['QCOUTPUTDIR'], fileExt = '.mmrrc.rpt') title = 'MMRRC Strains w/ Genotype Associations where the Markers/Alleles of the Strain record\n' + \ 'do not exactly match the Markers/Alleles of the Genotype record.' mmncfp.write(title + '\n\n') mmncf... | 475,109 |
def to_point(self): '''Convert from PolarPoint to (cartesian) Point object''' | def to_point(self): '''Convert from PolarPoint to (cartesian) Point object''' | 475,110 |
def connect_action(self, event): | def connect_action(self, event): | 475,111 |
def connect_action(self, event): | def connect_action(self, event): | 475,112 |
def connect_action(self, event): | def connect_action(self, event): | 475,113 |
def connect_action(self, event): | def connect_action(self, event): | 475,114 |
def is_point(self, line): | def is_point(self, line): | 475,115 |
def _points(self): points = [] for row in rows: fs = row.split(',') if fs[0] == 'ST': x = fs[-3] # FIXME NEZ coord order shouldn't be hardcoded y = fs[-2] z = fs[-1] bp = BasePoint(x=x, y=y, z=z, ih=0) if fs[0] == 'SS': angle = fs[4] z_angle = fs[5] dist = fs[3] th = fs[2] p = PolarPoint(dist=dist, angle=angle, z_ang... | def _points(self): points = [] for row in rows: fs = row.split(',') if fs[0] == 'ST': x = fs[-3] # FIXME NEZ coord order shouldn't be hardcoded y = fs[-2] z = fs[-1] bp = BasePoint(x=x, y=y, z=z, ih=0) if fs[0] == 'SS': angle = fs[4] z_angle = fs[5] dist = fs[3] th = fs[2] p = PolarPoint(dist=dist, angle=angle, z_ang... | 475,116 |
def url(dialog, link, data=None): pass | def url(dialog, link, data=None): pass | 475,117 |
def _check_for_Flash(self): try: tmp = "/tmp" ## look for filenames that begin with 'Flash' for file in os.listdir(tmp): try: if file.startswith("Flash"): filepath = os.path.join(tmp, file) duration = utils.getFLVLength(filepath) | def _check_for_Flash(self): try: tmp = "/tmp" ## look for filenames that begin with 'Flash' for file in os.listdir(tmp): try: if file.startswith("Flash"): filepath = os.path.join(tmp, file) duration = utils.getFLVLength(filepath) | 475,118 |
def _check_for_Flash(self): try: tmp = "/tmp" ## look for filenames that begin with 'Flash' for file in os.listdir(tmp): try: if file.startswith("Flash"): filepath = os.path.join(tmp, file) duration = utils.getFLVLength(filepath) | def _check_for_Flash(self): try: tmp = "/tmp" ## look for filenames that begin with 'Flash' for file in os.listdir(tmp): try: if file.startswith("Flash"): filepath = os.path.join(tmp, file) duration = utils.getFLVLength(filepath) | 475,119 |
def __init__(self): self.Core = core.Caffeine() | def __init__(self): self.Core = core.Caffeine() | 475,120 |
def on_about_button_clicked (self, button, data=None): | def on_about_button_clicked (self, button, data=None): | 475,121 |
def on_time_submenuitem_activate(self, menuitem, time): | def on_time_submenuitem_activate(self, menuitem, time): | 475,122 |
def _run_dialog(self): response = self.about_dialog.run() self.about_dialog.destroy() return False | def on_about_menuitem_activate(self, menuitem, data=None): if appindicator_avail: gtk.gdk.threads_enter() self.about_dialog.set_position (gtk.WIN_POS_CENTER_ALWAYS) response = self.about_dialog.run() self.about_dialog.destroy() return False | 475,123 |
def _run_dialog(self): response = self.about_dialog.run() self.about_dialog.destroy() return False | def _run_dialog(self): response = self.about_dialog.run() self.about_dialog.destroy() return False | 475,124 |
def quit(self): ### Do anything that needs to be done before quitting. logging.info("Caffeine is preparing to quit") | def quit(self): ### Do anything that needs to be done before quitting. logging.info("Caffeine is preparing to quit") | 475,125 |
def os_chown(self, path, uid, gid): ino = self.ino_from_path(path) inode = self.inode_read(ino) inode["uid"] = uid inode["gid"] = gid self.inode_write(ino, inode) | def os_chown(self, path, uid, gid): ino = self.ino_from_path(path) if uid == -1 and gid == -1: return inode = self.inode_read(ino) if uid != -1: inode["uid"] = uid if gid != -1: inode["gid"] = gid if inode["mode"] & (S_IXUSR|S_IXGRP|S_IXOTH): inode["mode"] &= ~(S_ISUID|S_ISGID) inode.set_time_fields(change=True) self.i... | 475,126 |
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]... | def _rmdir(self, pino, name): offset, dirent = self._scandir(pino, name) ino = dirent["ino"] inode = self.inode_read(ino) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent[... | 475,127 |
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]... | def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"], inode["nlink"]) # free all the clusters of the d... | 475,128 |
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]... | def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", ino, inode["n... | 475,129 |
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]... | def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]... | 475,130 |
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]... | def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]... | 475,131 |
def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]... | def _rmdir(self, dirino, name): offset, dirent = self._scandir(dirino, name) inode = self.inode_read(dirent["ino"]) if not S_ISDIR(inode["mode"]): raise ClfsError(ENOTDIR) if inode["nlink"] > 2: raise ClfsError(ENOTEMPTY) if inode["nlink"] < 2: logger.error( "Directory inode %i has invalid link count %i", dirent["ino"]... | 475,132 |
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values) | def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values) | 475,133 |
def __init__(self, errno): self.errno = errno | def __init__(self, errno): self.errno = errno | 475,134 |
def create_node(self, path, type): #pdb.set_trace() node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(paren... | defcreate_node(self,path,type):#pdb.set_trace()node_dirname,node_basename=os.path.split(path)parent_dirname,parent_basename=os.path.split(node_dirname)parent_dirent=self.get_dir_entry(node_dirname)parent_inode_struct=self.get_inode_struct(parent_dirent["inode"])fordirentinself.read_directory(parent_dirent["inode"]):ifd... | 475,135 |
def create_node(self, path, type): #pdb.set_trace() node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(paren... | def create_node(self, path, type): #pdb.set_trace() node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(paren... | 475,136 |
def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.write("Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim.") with stattab_table_tag(doc.open_tag)... | def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.add_tag("p", data="Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim. Shown first are those that... | 475,137 |
def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.write("Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim.") with stattab_table_tag(doc.open_tag)... | def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.write("Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim.") with stattab_table_tag(doc.open_tag)... | 475,138 |
def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.write("Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim.") with stattab_table_tag(doc.open_tag)... | def pz_locked(doc, pageContext): world = pageContext.get_selected_world() curtime = int(time.time()) limits = (0, 30) #limits = (0, 200) doc.write("Players sorted by descending protection zone lock time remaining. Also shown is their level, vocation, guild, and most recent victim.") with stattab_table_tag(doc.open_tag)... | 475,139 |
def curses_color(): import curses class TermInfo(): # this. is. ANSIIIII!!! _ANSI_COLORS = """BLACK RED GREEN YELLOW BLUE MAGENTA CYAN WHITE""".split() _STRING_CAPS = """NORMAL=sgr0""".split() def __init__(self, stream=sys.stdout): # isatty might be needed here? curses.setupterm(None, stream.fileno()) for prefix, capna... | def curses_color(): import curses class TermInfo(): # this. is. ANSIIIII!!! _ANSI_COLORS = """BLACK RED GREEN YELLOW BLUE MAGENTA CYAN WHITE""".split() _STRING_CAPS = """NORMAL=sgr0""".split() def __init__(self, stream=sys.stdout): # isatty might be needed here? curses.setupterm(None, stream.fileno()) for prefix, capna... | 475,140 |
def select_color(): try: import curses except ImportError: import ctypes try: from ctypes import wintypes finally: pass return windows_color() else: return curses_color() assert False | def select_color(): for a in [curses_color, windows_color]: b = a() if b != None: return b else: return curses_color() assert False | 475,141 |
def select_color(): try: import curses except ImportError: import ctypes try: from ctypes import wintypes finally: pass return windows_color() else: return curses_color() assert False | def select_color(): try: import curses except ImportError: import ctypes try: from ctypes import wintypes finally: pass return windows_color() else: assert False | 475,142 |
def __get__(self, instance, owner): #assert instance == None, instance return property.__get__(self, owner) | def__get__(self,instance,owner):#assertinstance==None,instancereturnproperty.__get__(self,owner) | 475,143 |
def ClfsStructField(*args): a = list(args) if len(a) < 3: a.append(None) else: assert a[2] != None return collections.namedtuple( "ClfsStructField", ("name", "format", "initval") )(*a) | def ClfsStructField(*args): a = list(args) if len(a) < 3: a.append(None) else: assert a[2] != None, "None is reserved for indicating uninitialized fields" return namedtuple( "ClfsStructField", ("name", "format", "initval") )(*a) | 475,144 |
def all_fields(): for base in bases: if hasattr(base, "fields"): for field in base.fields: yield field for a in attrs["_fields_"]: yield ClfsStructField(*a) | def all_fields(): for base in bases: if hasattr(base, "fields"): for field in base.fields: yield field for a in attrs["_fields_"]: yield ClfsStructField(*a) | 475,145 |
def from_fileobj(class_, fileobj): return class_.unpack(fileobj.read(class_.size)) | def from_fileobj(class_, fileobj): return class_.unpack(fileobj.read(class_.size)) | 475,146 |
def unpack(class_, buffer): instance = class_() offset = 0 for field in instance.fields: #assert field.name in instance.__values unpacked = struct.unpack_from(field.format, buffer, offset) assert len(unpacked) == 1 instance[field.name] = unpacked[0] #instance.__values[field.name] = unpacked[0] offset += struct.calcsize... | def unpack(class_, buffer): instance = class_() offset = 0 for field in instance.fields: #assert field.name in instance.__values unpacked = struct.unpack_from(field.format, buffer, offset) if len(unpacked): instance[field.name], = unpacked #instance.__values[field.name] = unpacked[0] offset += struct.calcsize(field.for... | 475,147 |
def pack(self): buffer = "" for field in self.fields: assert field.name in self.__values, "Field %r is uninitialized" % (field.name,) value = self.__values[field.name] try: buffer += struct.pack(field.format, value) except struct.error as exc: raise struct.error("Error packing %r into %s.%s, %s" % (value, self.__class_... | def pack(self): buffer = "" for field in self.fields: try: buffer += struct.pack(field.format, value) except struct.error as exc: raise struct.error("Error packing %r into %s.%s, %s" % (value, self.__class__.__name__, field.name, exc.message)) assert len(buffer) == self.size return buffer | 475,148 |
def pack(self): buffer = "" for field in self.fields: assert field.name in self.__values, "Field %r is uninitialized" % (field.name,) value = self.__values[field.name] try: buffer += struct.pack(field.format, value) except struct.error as exc: raise struct.error("Error packing %r into %s.%s, %s" % (value, self.__class_... | def pack(self): buffer = "" for field in self.fields: assert field.name in self.__values, "Field %r is uninitialized" % (field.name,) value = self.__values[field.name] try: value = (self.__values[field.name],) except KeyError: value = () try: buffer += struct.pack(field.format, *value) except struct.error as exc: raise... | 475,149 |
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values) | def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values) | 475,150 |
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values) | def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values) | 475,151 |
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values) | def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values) | 475,152 |
def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values) | def __repr__(self): return "<{0} {1}>".format(self.__class__.__name__, self.__values) | 475,153 |
def __init__(self, path=None, fileobj=None): assert bool(path) ^ bool(fileobj) if path: self.f = open(path, "r+b") else: self.f = fileobj self.f.seek(0) br = BootRecord.from_fileobj(self.f) assert br["ident"].rstrip("\0") == "clfs", repr(br["ident"]) assert br["version"] == 1 self.cluster_size = br["clrsize"] self.mast... | def __init__(self, path): self.f = open(path, "r+b") br = BootRecord.from_fileobj(self.f) assert br["ident"].rstrip("\0") == "clfs", repr(br["ident"]) assert br["version"] == 1 self.cluster_size = br["clrsize"] self.master_region_cluster_count = br["mstrclrs"] self.allocation_table_cluster_count = br["atabclrs"] self.d... | 475,154 |
def __init__(self, path=None, fileobj=None): assert bool(path) ^ bool(fileobj) if path: self.f = open(path, "r+b") else: self.f = fileobj self.f.seek(0) br = BootRecord.from_fileobj(self.f) assert br["ident"].rstrip("\0") == "clfs", repr(br["ident"]) assert br["version"] == 1 self.cluster_size = br["clrsize"] self.mast... | def __init__(self, path=None, fileobj=None): assert bool(path) ^ bool(fileobj) if path: self.f = open(path, "r+b") else: self.f = fileobj self.f.seek(0) br = BootRecord.from_fileobj(self.f) assert br["ident"].rstrip("\0") == "clfs", repr(br["ident"]) assert br["version"] == 1 self.cluster_size = br["clrsize"] self.mast... | 475,155 |
def read_directory(self, inode): inode_struct = self.get_inode_struct(inode) if inode_struct["type"] != TYPE_DIRECTORY: raise ClfsError(ENOTDIR) offset = 0 while offset < inode_struct["size"]: dirent = DirEntry.unpack(self.read_inode_data( inode, offset, DirEntry.size)) if dirent["name"].rstrip("\0"): yield dirent offs... | def read_directory(self, inode): inode_struct = self.get_inode_struct(inode) if not S_ISDIR(inode_struct["mode"]): raise ClfsError(ENOTDIR) offset = 0 while offset < inode_struct["size"]: dirent = DirEntry.unpack(self.read_inode_data( inode, offset, DirEntry.size)) if dirent["name"].rstrip("\0"): yield dirent offset +=... | 475,156 |
def get_dir_entry(self, path): for name in path.split("/"): if not name: cur_dirent = self.get_root_dir_entry() else: # pdb.set_trace() for dirent in self.read_directory(cur_dirent["inode"]): if dirent["name"].rstrip("\0") == name: cur_dirent = dirent break else: raise ClfsError(ENOENT) return cur_dirent | def get_dir_entry(self, path): for name in path.split("/"): if not name: cur_dirent = self.get_root_dir_entry() else: # pdb.set_trace() for dirent in self.read_directory(cur_dirent["ino"]): if dirent["name"].rstrip("\0") == name: cur_dirent = dirent break else: raise ClfsError(ENOENT) return cur_dirent | 475,157 |
def get_dir_entry(self, path): for name in path.split("/"): if not name: cur_dirent = self.get_root_dir_entry() else: # pdb.set_trace() for dirent in self.read_directory(cur_dirent["inode"]): if dirent["name"].rstrip("\0") == name: cur_dirent = dirent break else: raise ClfsError(ENOENT) return cur_dirent | def get_dir_entry(self, path): for name in path.split("/"): if not name: cur_dirent = self.get_root_dir_entry() else: # pdb.set_trace() for dirent in self.read_directory(cur_dirent["inode"]): if dirent["name"].rstrip("\0") == name: cur_dirent = dirent break else: raise ClfsError(ENOENT) return cur_dirent | 475,158 |
#def write(self, path, buf, offset): | #def write(self, path, buf, offset): | 475,159 |
def write_inode_data(self, inode, offset, buffer): inode_struct = self.get_inode_struct(inode) data_offset = inode_struct.size write_size, new_size = self.write_to_chain( inode, inode_struct["size"] + data_offset, offset + data_offset, buffer) assert write_size == len(buffer), write_size expected_size = data_offset + m... | def write_inode_data(self, ino, offset, buffer): inode_struct = self.get_inode_struct(inode) data_offset = inode_struct.size write_size, new_size = self.write_to_chain( ino, inode_struct["size"] + data_offset, offset + data_offset, buffer) assert write_size == len(buffer), write_size expected_size = data_offset + max(o... | 475,160 |
def write_inode_data(self, inode, offset, buffer): inode_struct = self.get_inode_struct(inode) data_offset = inode_struct.size write_size, new_size = self.write_to_chain( inode, inode_struct["size"] + data_offset, offset + data_offset, buffer) assert write_size == len(buffer), write_size expected_size = data_offset + m... | def write_inode_data(self, inode, offset, buffer): inode_struct = self.get_inode_struct(inode) data_offset = inode_struct.size write_size, new_size = self.write_to_chain( inode, inode_struct["size"] + data_offset, offset + data_offset, buffer) assert write_size == len(buffer), write_size expected_size = data_offset + m... | 475,161 |
def create_node(self, path, type): node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(parent_dirent["inode"]... | def create_node(self, path, mode): """Create an allocate a new inode, update relevant structures elsewhere""" node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(pa... | 475,162 |
def create_node(self, path, type): node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(parent_dirent["inode"]... | def create_node(self, path, type): node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) create_rootdir = bool( (not node_basename) and (node_dirname == parent_dirname == "/")) if create_rootdir: assert S_ISDIR(mode) new_inode = Inode(size=0, uid=0, gid=0, rde... | 475,163 |
def create_node(self, path, type): node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(parent_dirent["inode"]... | def create_node(self, path, type): node_dirname, node_basename = os.path.split(path) parent_dirname, parent_basename = os.path.split(node_dirname) parent_dirent = self.get_dir_entry(node_dirname) parent_inode_struct = self.get_inode_struct(parent_dirent["inode"]) for dirent in self.read_directory(parent_dirent["inode"]... | 475,164 |
def add_share(self, button): namelbl = gtk.Label("Share name:") | def add_share(self, button): namelbl = gtk.Label("Share name:") | 475,165 |
def browse_peer_by_url(self, url): """Open the given peer URL with the most natural file manager for the current platform that we can find""" import os, subprocess if os.name == "nt": # this is how it's done programmatically? except that it won't invoke # the default file manager (explorer) on winders #ShellExecute(Non... | def browse_peer_by_url(self, url): """Open the given peer URL with the most natural file manager for the current platform that we can find""" import os, subprocess if os.name == "nt": # this is how it's done programmatically? except that it won't invoke # the default file manager (explorer) on winders #ShellExecute(Non... | 475,166 |
def browse_peer_by_url(self, url): """Open the given peer URL with the most natural file manager for the current platform that we can find""" import os, subprocess if os.name == "nt": # this is how it's done programmatically? except that it won't invoke # the default file manager (explorer) on winders #ShellExecute(Non... | def browse_peer_by_url(self, url): """Open the given peer URL with the most natural file manager for the current platform that we can find""" import os, subprocess if os.name == "nt": # this is how it's done programmatically? except that it won't invoke # the default file manager (explorer) on winders #ShellExecute(Non... | 475,167 |
def read_from_chain(self, first_cluster, chain_size, read_offset, read_size): if chain_size <= 0: return "" #assert read_offset + read_size <= chain_size, (read_offset, read_size, chain_size) if read_offset > self.cluster_size: return self.read_from_chain( self.next_cluster(first_cluster), chain_size - self.cluster_siz... | def read_from_chain(self, first_cluster, chain_size, read_offset, read_size): if chain_size <= 0: return "" #assert read_offset + read_size <= chain_size, (read_offset, read_size, chain_size) if read_offset > self.cluster_size: return self.read_from_chain( self.next_cluster(first_cluster), chain_size - self.cluster_siz... | 475,168 |
def trans_include(repo_uri, fargs, transaction=None): basedirs = [] timestamp_files = [] error_occurred = False opts, pargs = getopt.getopt(fargs, "d:T:") for opt, arg in opts: if opt == "-d": basedirs.append(arg) elif opt == "-T": timestamp_files.append(arg) if transaction == None: try: trans_id = os.environ["PKG_TR... | def trans_include(repo_uri, fargs, transaction=None): basedirs = [] timestamp_files = [] error_occurred = False opts, pargs = getopt.getopt(fargs, "d:T:") for opt, arg in opts: if opt == "-d": basedirs.append(arg) elif opt == "-T": timestamp_files.append(arg) if transaction == None: try: trans_id = os.environ["PKG_TR... | 475,169 |
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header.""" | def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header.""" | 475,170 |
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header.""" | def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header.""" | 475,171 |
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header.""" | def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header.""" | 475,172 |
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header.""" | def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header.""" | 475,173 |
def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header.""" | def test_bug_12271_14088(self): """Check that consecutive duplicate lines are removed and that having a single option to -o still prints the header.""" | 475,174 |
def run(self): # nuke everything print("deleting " + dist_dir) shutil.rmtree(dist_dir, True) print("deleting " + build_dir) shutil.rmtree(build_dir, True) print("deleting " + root_dir) shutil.rmtree(root_dir, True) print("deleting " + pkgs_dir) shutil.rmtree(pkgs_dir, True) print("deleting " + extern_dir) shutil.rmtree... | def run(self): # nuke everything print("deleting " + dist_dir) shutil.rmtree(dist_dir, True) print("deleting " + build_dir) shutil.rmtree(build_dir, True) print("deleting " + root_dir) shutil.rmtree(root_dir, True) print("deleting " + pkgs_dir) shutil.rmtree(pkgs_dir, True) print("deleting " + extern_dir) shutil.rmtree... | 475,175 |
def initialize_options(self): self.only = "" self.baselinefile = "" self.verbosemode = 0 self.parseable = 0 self.genbaseline = 0 self.timing = 0 self.coverage = 0 self.stoponerr = 0 self.debugoutput = 0 self.show_on_expected_fail = 0 self.startattest = "" self.archivedir = "" | def initialize_options(self): self.only = "" self.baselinefile = "" self.verbosemode = 0 self.parseable = 0 self.genbaseline = 0 self.timing = 0 self.coverage = 0 self.stoponerr = 0 self.debugoutput = 0 self.showonexpectedfail = 0 self.startattest = "" self.archivedir = "" | 475,176 |
def main_func(): global cache_dir, download_start, xport, xport_cfg all_timestamps = False all_versions = False keep_compressed = False list_newest = False recursive = False src_uri = None target = None incoming_dir = None src_pub = None targ_pub = None temp_root = misc.config_temp_root() gettext.install("pkg", "/usr... | def main_func(): global cache_dir, download_start, xport, xport_cfg all_timestamps = False all_versions = False keep_compressed = False list_newest = False recursive = False src_uri = None target = None incoming_dir = None src_pub = None targ_pub = None temp_root = misc.config_temp_root() gettext.install("pkg", "/usr... | 475,177 |
def main_func(): global cache_dir, download_start, xport, xport_cfg all_timestamps = False all_versions = False keep_compressed = False list_newest = False recursive = False src_uri = None target = None incoming_dir = None src_pub = None targ_pub = None temp_root = misc.config_temp_root() gettext.install("pkg", "/usr... | def main_func(): global cache_dir, download_start, xport, xport_cfg all_timestamps = False all_versions = False keep_compressed = False list_newest = False recursive = False src_uri = None target = None incoming_dir = None src_pub = None targ_pub = None temp_root = misc.config_temp_root() gettext.install("pkg", "/usr... | 475,178 |
def get_basename(pfmri): open_time = pfmri.get_timestamp() return "%d_%s" % \ (calendar.timegm(open_time.utctimetuple()), urllib.quote(str(pfmri), "")) | def get_basename(pfmri): open_time = pfmri.get_timestamp() return "%d_%s" % \ (calendar.timegm(open_time.utctimetuple()), urllib.quote(str(pfmri), "")) | 475,179 |
def get_publisherinfo(self, pub, ccancel=None): """Given a publisher pub, return the publisher/0 information in a StringIO object.""" | def get_publisherinfo(self, pub, ccancel=None): """Given a publisher pub, return the publisher/0 information in a StringIO object.""" | 475,180 |
def main_func(): global file_repo global def_branch global def_repo global def_vers global extra_entire_contents global just_these_pkgs global not_these_pkgs global nopublish global publish_all global print_pkg_names global reference_uris global show_debug global wos_path global not_these_consolidations global curpkg ... | def main_func(): global file_repo global def_branch global def_repo global def_vers global extra_entire_contents global just_these_pkgs global not_these_pkgs global nopublish global publish_all global print_pkg_names global reference_uris global show_debug global wos_path global not_these_consolidations global curpkg ... | 475,181 |
def get_smf_packages(server_url, manifest_locations, filter): """ Performs a search against server_url looking for packages which contain SMF manifests, returning a list of those pfmris """ dir = os.getcwd() tracker = pkg.client.progress.QuietProgressTracker() image_dir = tempfile.mkdtemp("", "pkg_importer_smfsearch."... | def get_smf_packages(server_url, manifest_locations, filter): """ Performs a search against server_url looking for packages which contain SMF manifests, returning a list of those pfmris """ dir = os.getcwd() tracker = pkg.client.progress.QuietProgressTracker() image_dir = tempfile.mkdtemp("", "pkg_importer_smfsearch."... | 475,182 |
def get_smf_packages(server_url, manifest_locations, filter): """ Performs a search against server_url looking for packages which contain SMF manifests, returning a list of those pfmris """ dir = os.getcwd() tracker = pkg.client.progress.QuietProgressTracker() image_dir = tempfile.mkdtemp("", "pkg_importer_smfsearch."... | def get_smf_packages(server_url, manifest_locations, filter): """ Performs a search against server_url looking for packages which contain SMF manifests, returning a list of those pfmris """ dir = os.getcwd() tracker = pkg.client.progress.QuietProgressTracker() image_dir = tempfile.mkdtemp("", "pkg_importer_smfsearch."... | 475,183 |
def set_function_fields(self, cursor, user, id, name, value, arg, context=None): request_obj = self.pool.get('res.request') req_ref_obj = self.pool.get('res.request.reference') | def set_function_fields(self, cursor, user, id, name, value, arg, context=None): request_obj = self.pool.get('res.request') req_ref_obj = self.pool.get('res.request.reference') | 475,184 |
def add_minutes(self, cursor, user, company, date, minutes, context=None): minutes = int(round(minutes)) minutes = date.minute + minutes | def add_minutes(self, cursor, user, company, date, minutes, context=None): minutes = int(round(minutes)) minutes = date.minute + minutes | 475,185 |
def add_hours(self, cursor, user, company, date, hours, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day | def add_hours(self, cursor, user, company, date, hours, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day | 475,186 |
def add_days(self, cursor, user, company, date, days, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day | def add_days(self, cursor, user, company, date, days, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day | 475,187 |
def add_weeks(self, cursor, user, company, date, weeks, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day | def add_weeks(self, cursor, user, company, date, weeks, context=None): day_per_week = company.hours_per_work_week / company.hours_per_work_day | 475,188 |
def get_function_fields(self, ids, names): ''' Function to compute function fields | def get_function_fields(self, ids, names): ''' Function to compute function fields | 475,189 |
def test0005views(self): ''' Test views. ''' self.assertRaises(Exception, test_view('project_plan')) | def test0005views(self): ''' Test views. ''' self.assertRaises(Exception, test_view('project_plan')) | 475,190 |
def set_function_fields(self, cursor, user, ids, name, value, context=None): request_obj = self.pool.get('res.request') req_ref_obj = self.pool.get('res.request.reference') | def set_function_fields(self, cursor, user, ids, name, value, context=None): request_obj = self.pool.get('res.request') req_ref_obj = self.pool.get('res.request.reference') | 475,191 |
def log_job_closed(self): log.info("Job finished: Ticket:%s Task:%s User:%s Start:%f End:%f Time:%f Price:%f State:%s" % (self.ticket(), self.task(), self.getUser(), self.jobStart(), self.jobEnd(), self.getTime(), self.getCost(), self.getState())) | def log_job_closed(self): log.info("Job finished: Ticket:%s Task:%s User:%s Start:%f End:%f Time:%f Price:%f State:%s" % (self.ticket(), self.task(), self.getUser(), self.jobStart(), self.jobEnd(), self.getTime(), self.getCost(), self.getState())) | 475,192 |
def __init__(self, ticket_id, id, mgr): """Initialize a new task.""" TaskFSM.__init__(self) self.mtx = self.fsm.getLock() | def __init__(self, ticket_id, id, mgr): """Initialize a new task.""" TaskFSM.__init__(self) self.mtx = self.fsm.getLock() | 475,193 |
def cleanUp(self): """cleans up the task, i.e. removes the task's spool directory""" self.log.info("removing spool directory") from xbe.util import removeDirCompletely removeDirCompletely(self.__spool) | def cleanUp(self): """cleans up the task, i.e. removes the task's spool directory""" self.log.info("removing spool directory") from xbe.util import removeDirCompletely removeDirCompletely(self.__spool) | 475,194 |
def getState(self): return self.__fsm.getState().getName() | def getState(self): return self.__fsm.getState().getName() | 475,195 |
def do_Event(self, event, reqCtxt): log.debug("JOB '%s' run in state '%s' event '%s'" % (self.ticket(), self.__fsm.getState().getName(), event)) if hasattr(self.__fsm, event): log.debug("Run event '%s'" % event) getattr(self.__fsm, event)(self, reqCtxt) else: log.debug("Event '%s' not found." % event) raise CommandFai... | def do_Event(self, event, reqCtxt): log.debug("JOB '%s' run in state '%s' event '%s'" % (self.ticket(), self.__fsm.getState().getName(), event)) if hasattr(self.__fsm, event): log.debug("Run event '%s'" % event) getattr(self.__fsm, event)(self, reqCtxt) else: log.error("Event '%s' not found." % event) raise CommandFai... | 475,196 |
def do_EventByMap(self, eventkey, reqCtxt): eventMap = { "Pending:Reserved" : 1, "Pending:Confirmed" : "confirm", "Running:Stage-In" : "runJob_StageIn", "Running:Instance-Starting" : "", "Running:Executing" : "runJob_Execute", "Running:Stage-Out" : "runJob_StageOut", "Running:Instance-Stopping" : "", "... | def do_EventByMap(self, eventkey, reqCtxt): eventMap = { "Pending:Reserved" : 1, "Pending:Confirmed" : "confirm", "Running:Stage-In" : "runJob_StageIn", "Running:Instance-Starting" : "", "Running:Executing" : "runJob_Execute", "Running:Stage-Out" : "runJob_StageOut", "Running:Instance-Stopping" : "", "... | 475,197 |
def do_EventByMap(self, eventkey, reqCtxt): eventMap = { "Pending:Reserved" : 1, "Pending:Confirmed" : "confirm", "Running:Stage-In" : "runJob_StageIn", "Running:Instance-Starting" : "", "Running:Executing" : "runJob_Execute", "Running:Stage-Out" : "runJob_StageOut", "Running:Instance-Stopping" : "", "... | def do_EventByMap(self, eventkey, reqCtxt): eventMap = { "Pending:Reserved" : 1, "Pending:Confirmed" : "confirm", "Running:Stage-In" : "runJob_StageIn", "Running:Instance-Starting" : "", "Running:Executing" : "runJob_Execute", "Running:Stage-Out" : "runJob_StageOut", "Running:Instance-Stopping" : "", "... | 475,198 |
def log_job_closed(self): log.info("Job finished: Ticket:%s Task:%s User:%s Start:%f End:%f Time:%f Price:%f State:%s" % (self.ticket(), self.task(), self.getUser(), self.getStart(), self.getEnd(), self.getTime(), self.getCost(), self.getState())) | def log_job_closed(self): log.info("Job finished: Ticket:%s Task:%s User:%s Start:%f End:%f Time:%f Price:%f State:%s" % (self.ticket(), self.task(), self.getUser(), self.getStart(), self.getEnd(), self.getTime(), self.getCost(), self.getState())) | 475,199 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.