rem stringlengths 1 322k | add stringlengths 0 2.05M | context stringlengths 4 228k | meta stringlengths 156 215 |
|---|---|---|---|
responseElements = responseText.split('\'') housenumber = responseElements[7] street = responseElements[3] postcode = responseElements[5] city = responseElements[11] | responseElements = responseText.split('\n') for element in responseElements: if element.rfind('strname1') > -1: strname1_s = element.split('=') street = strname1_s[1].lstrip().lstrip('\'').rstrip().rstrip('\'') if element.rfind('plz4') > -1: plz4_s = element.split('=') postcode = plz4_s[1].lstrip().lstrip('\'').rstrip(... | def index(self): if 'latitude' in request.params and 'longitude' in request.params: latitude = float(request.params['latitude']) longitude = float(request.params['longitude']) if 'easting' in request.params: easting = float(request.params['easting']) if 'northing' in request.params: northing = float(request.params['nor... | 3897a1d010b764abbd47a819185a92f461de6f3f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4401/3897a1d010b764abbd47a819185a92f461de6f3f/swissBuilding.py |
tsvector = 'tsvector_street_housenumber_city' | tsvector = "to_tsvector('english', coalesce(street,'') || ' ' || coalesce(housenumber,'') || ' ' || coalesce(city,''))" | def index(self, format='json'): """GET /: return all features.""" # If no filter argument is passed to the protocol index method # then the default MapFish filter is used. This default filter # is constructed based on the box, lon, lat, tolerance GET # params. # # If you need your own filter with application-specific p... | 75c116903bf18933c805a3dbd4e8965b5542ab62 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4401/75c116903bf18933c805a3dbd4e8965b5542ab62/addresses.py |
tsvector = 'tsvector_street' | tsvector = "to_tsvector('english', coalesce(street,''))" | def index(self, format='json'): """GET /: return all features.""" # If no filter argument is passed to the protocol index method # then the default MapFish filter is used. This default filter # is constructed based on the box, lon, lat, tolerance GET # params. # # If you need your own filter with application-specific p... | 75c116903bf18933c805a3dbd4e8965b5542ab62 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4401/75c116903bf18933c805a3dbd4e8965b5542ab62/addresses.py |
limit = request.params['limit'] | limit = int(request.params['limit']) | def fullTextSearch(self,request): # addresses/fullTextSearch?fields=street,city,housenumber&query=ch%20du%2028&tolerance=0.005&easting=6.62379551&northing=46.51687241&limit=20&distinct=true # Read request parameters fields = request.params['fields'] | 1eed76baae632ad892bfdc293b1423bf91b7b1cf /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4401/1eed76baae632ad892bfdc293b1423bf91b7b1cf/addresses.py |
yield UInt16(self, "left", "Text Grid Left") yield UInt16(self, "top", "Text Grid Top") yield UInt16(self, "width", "Text Grid Width") yield UInt16(self, "height", "Text Grid Height") | yield UInt16(parent, "left", "Text Grid Left") yield UInt16(parent, "top", "Text Grid Top") yield UInt16(parent, "width", "Text Grid Width") yield UInt16(parent, "height", "Text Grid Height") | def parseTextExtension(parent): yield UInt8(parent, "block_size", "Block Size") yield UInt16(self, "left", "Text Grid Left") yield UInt16(self, "top", "Text Grid Top") yield UInt16(self, "width", "Text Grid Width") yield UInt16(self, "height", "Text Grid Height") yield UInt8(parent, "cell_width", "Character Cell Width"... | 53d940dc57135bdf11782fc668764795df4a40c7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/53d940dc57135bdf11782fc668764795df4a40c7/gif.py |
def parseTextExtension(parent): yield UInt8(parent, "block_size", "Block Size") yield UInt16(self, "left", "Text Grid Left") yield UInt16(self, "top", "Text Grid Top") yield UInt16(self, "width", "Text Grid Width") yield UInt16(self, "height", "Text Grid Height") yield UInt8(parent, "cell_width", "Character Cell Width"... | 53d940dc57135bdf11782fc668764795df4a40c7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/53d940dc57135bdf11782fc668764795df4a40c7/gif.py | ||
cvt_time=lambda v:datetime(2001,1,1) + timedelta(seconds=v) | def cvt_time(v): v=timedelta(seconds=v) epoch2001 = datetime(2001,1,1) epoch1970 = datetime(1970,1,1) if (epoch2001 + v - datetime.today()).days > 5*365: return epoch1970 + v return epoch2001 + v | def createFields(self): yield Enum(Bits(self, "marker_type", 4), {0: "Simple", 1: "Int", 2: "Real", 3: "Date", 4: "Data", 5: "ASCII String", 6: "UTF-16-BE String", 8: "UID", 10: "Array", 13: "Dict",}) markertype = self['marker_type'].value if markertype == 0: # Simple (Null) yield Enum(Bits(self, "value", 4), {0: "Null... | 879aa46a777fb162d538d6b14d159fddff160ba3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/879aa46a777fb162d538d6b14d159fddff160ba3/bplist.py |
self.xml=lambda prefix:prefix + "<date>%s</date>"%(cvt_time(self['value'].value).isoformat()) | self.xml=lambda prefix:prefix + "<date>%sZ</date>"%(cvt_time(self['value'].value).isoformat()) | def createFields(self): yield Enum(Bits(self, "marker_type", 4), {0: "Simple", 1: "Int", 2: "Real", 3: "Date", 4: "Data", 5: "ASCII String", 6: "UTF-16-BE String", 8: "UID", 10: "Array", 13: "Dict",}) markertype = self['marker_type'].value if markertype == 0: # Simple (Null) yield Enum(Bits(self, "value", 4), {0: "Null... | 879aa46a777fb162d538d6b14d159fddff160ba3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/879aa46a777fb162d538d6b14d159fddff160ba3/bplist.py |
self.xml=lambda prefix:prefix + "<string>%s</string>"%(self['value'].value.encode('iso-8859-1')) | self.xml=lambda prefix:prefix + "<string>%s</string>"%(self['value'].value.replace('&','&').encode('iso-8859-1')) | def createFields(self): yield Enum(Bits(self, "marker_type", 4), {0: "Simple", 1: "Int", 2: "Real", 3: "Date", 4: "Data", 5: "ASCII String", 6: "UTF-16-BE String", 8: "UID", 10: "Array", 13: "Dict",}) markertype = self['marker_type'].value if markertype == 0: # Simple (Null) yield Enum(Bits(self, "value", 4), {0: "Null... | 879aa46a777fb162d538d6b14d159fddff160ba3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/879aa46a777fb162d538d6b14d159fddff160ba3/bplist.py |
self.xml=lambda prefix:prefix + "<string>%s</string>"%(self['value'].value.encode('utf-8')) | self.xml=lambda prefix:prefix + "<string>%s</string>"%(self['value'].value.replace('&','&').encode('utf-8')) | def createFields(self): yield Enum(Bits(self, "marker_type", 4), {0: "Simple", 1: "Int", 2: "Real", 3: "Date", 4: "Data", 5: "ASCII String", 6: "UTF-16-BE String", 8: "UID", 10: "Array", 13: "Dict",}) markertype = self['marker_type'].value if markertype == 0: # Simple (Null) yield Enum(Bits(self, "value", 4), {0: "Null... | 879aa46a777fb162d538d6b14d159fddff160ba3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/879aa46a777fb162d538d6b14d159fddff160ba3/bplist.py |
while field: | while field is not None: | def _getPath(self): if not self._parent: return '/' names = [] field = self while field: names.append(field._name) field = field._parent names[-1] = '' return '/'.join(reversed(names)) | 49e0048f74af99b500949ef7bc8882ae02f940b2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/49e0048f74af99b500949ef7bc8882ae02f940b2/field.py |
addr_text = '' | addr_text_list = [] | def update_addr_view(self): addr_text = '' for i in xrange(self.view.get_height_chars()): addr_text += self.format_addr(self.pos+i*self.view.get_width_chars())+'\n' self.view.addr_view.SetValue(addr_text) | 1bd908f590c7f16fae10a5627fd938d03251d4e2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/1bd908f590c7f16fae10a5627fd938d03251d4e2/hex_view_imp.py |
addr_text += self.format_addr(self.pos+i*self.view.get_width_chars())+'\n' self.view.addr_view.SetValue(addr_text) | addr_text_list.append( self.format_addr(self.pos+i*self.view.get_width_chars())+'\n') self.view.addr_view.SetValue(''.join(addr_text_list)) | def update_addr_view(self): addr_text = '' for i in xrange(self.view.get_height_chars()): addr_text += self.format_addr(self.pos+i*self.view.get_width_chars())+'\n' self.view.addr_view.SetValue(addr_text) | 1bd908f590c7f16fae10a5627fd938d03251d4e2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/1bd908f590c7f16fae10a5627fd938d03251d4e2/hex_view_imp.py |
self.trackCommon(track, video) | def processVideo(self, track): video = Metadata(self) try: self.trackCommon(track, video) video.compression = track["CodecID/string"].value if "Video" in track: video.width = track["Video/PixelWidth/unsigned"].value video.height = track["Video/PixelHeight/unsigned"].value except MissingField: pass self.addGroup("video[... | 897be799c541a0c433231c35853f122b1e5ac550 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/897be799c541a0c433231c35853f122b1e5ac550/video.py | |
try: self.trackCommon(track, audio) if "Audio" in track: audio.sample_rate = track["Audio/SamplingFrequency/float"].value | self.trackCommon(track, audio) if "Audio" in track: frequency = self.getDouble(track, "Audio/SamplingFrequency") if frequency is not None: audio.sample_rate = frequency if "Audio/Channels/unsigned" in track: | def processAudio(self, track): audio = Metadata(self) try: self.trackCommon(track, audio) if "Audio" in track: audio.sample_rate = track["Audio/SamplingFrequency/float"].value audio.nb_channel = track["Audio/Channels/unsigned"].value audio.compression = track["CodecID/string"].value except MissingField: pass self.addGr... | 897be799c541a0c433231c35853f122b1e5ac550 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/897be799c541a0c433231c35853f122b1e5ac550/video.py |
except MissingField: pass | def processAudio(self, track): audio = Metadata(self) try: self.trackCommon(track, audio) if "Audio" in track: audio.sample_rate = track["Audio/SamplingFrequency/float"].value audio.nb_channel = track["Audio/Channels/unsigned"].value audio.compression = track["CodecID/string"].value except MissingField: pass self.addGr... | 897be799c541a0c433231c35853f122b1e5ac550 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/897be799c541a0c433231c35853f122b1e5ac550/video.py | |
self.trackCommon(track, sub) | def processSubtitle(self, track): sub = Metadata(self) try: self.trackCommon(track, sub) sub.compression = track["CodecID/string"].value except MissingField: pass self.addGroup("subtitle[]", sub, "Subtitle") | 897be799c541a0c433231c35853f122b1e5ac550 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/897be799c541a0c433231c35853f122b1e5ac550/video.py | |
@fault_tolerant def readDuration(self, duration, timecode_scale): seconds = duration * timecode_scale self.duration = timedelta(seconds=seconds) | def processSimpleTag(self, tag): if "TagName/unicode" not in tag \ or "TagString/unicode" not in tag: return name = tag["TagName/unicode"].value if name not in self.tag_key: return key = self.tag_key[name] value = tag["TagString/unicode"].value setattr(self, key, value) | 897be799c541a0c433231c35853f122b1e5ac550 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/897be799c541a0c433231c35853f122b1e5ac550/video.py | |
timecode_scale = info["TimecodeScale/unsigned"].value * 1e-9 if "Duration/float" in info: self.readDuration(info["Duration/float"].value, timecode_scale) elif "Duration/double" in info: self.readDuration(info["Duration/double"].value, timecode_scale) | duration = self.getDouble(info, "Duration") if duration is not None: try: seconds = duration * info["TimecodeScale/unsigned"].value * 1e-9 self.duration = timedelta(seconds=seconds) except OverflowError: pass | def processInfo(self, info): if "TimecodeScale/unsigned" in info: timecode_scale = info["TimecodeScale/unsigned"].value * 1e-9 if "Duration/float" in info: self.readDuration(info["Duration/float"].value, timecode_scale) elif "Duration/double" in info: self.readDuration(info["Duration/double"].value, timecode_scale) if ... | 897be799c541a0c433231c35853f122b1e5ac550 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/897be799c541a0c433231c35853f122b1e5ac550/video.py |
def createFields(self): yield Bit(self, "negative") yield FloatExponent(self, "exponent", exponent_bits) if 64 <= mantissa_bits: yield Bit(self, "one") yield FloatMantissa(self, "mantissa", mantissa_bits-1) else: yield FloatMantissa(self, "mantissa", mantissa_bits) | d1673c036515cd8008a4391fe6ee4e7c2b114020 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/d1673c036515cd8008a4391fe6ee4e7c2b114020/float.py | ||
def createFields(self): yield Bit(self, "negative") yield FloatExponent(self, "exponent", exponent_bits) if 64 <= mantissa_bits: yield Bit(self, "one") yield FloatMantissa(self, "mantissa", mantissa_bits-1) else: yield FloatMantissa(self, "mantissa", mantissa_bits) | d1673c036515cd8008a4391fe6ee4e7c2b114020 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/d1673c036515cd8008a4391fe6ee4e7c2b114020/float.py | ||
def createFields(self): yield Bit(self, "negative") yield FloatExponent(self, "exponent", exponent_bits) if 64 <= mantissa_bits: yield Bit(self, "one") yield FloatMantissa(self, "mantissa", mantissa_bits-1) else: yield FloatMantissa(self, "mantissa", mantissa_bits) | d1673c036515cd8008a4391fe6ee4e7c2b114020 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/d1673c036515cd8008a4391fe6ee4e7c2b114020/float.py | ||
return "section[]" | return "section[]" | def createSectionName(self): try: name = str(self["name"].value.strip(".")) if name: return "section_%s" % name except HACHOIR_ERRORS, err: self.warning(unicode(err)) return "section[]" | 359a3cbacd4c08ef6f20d1d7f8532d0686a60570 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/359a3cbacd4c08ef6f20d1d7f8532d0686a60570/exe_pe.py |
def createFields(self): yield UInt32(self, "unk") yield AtomList(self, "tags") | 3b1d5b3fc1813e8ce461e86889a0bbdd5760acb3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/3b1d5b3fc1813e8ce461e86889a0bbdd5760acb3/mov.py | ||
"stbl": (AtomList, "stbl", ""), | "stbl": (AtomList, "stbl", "Sample Table"), "stco": (STCO, "stsd", "Sample Table Chunk Offset"), "stsd": (STSD, "stsd", "Sample Table Sample Description"), "stss": (STSS, "stss", "Sample Table Sync Samples"), "stsz": (STSZ, "stsz", "Sample Table Sizes"), | def createFields(self): yield UInt32(self, "unk") yield AtomList(self, "tags") | 3b1d5b3fc1813e8ce461e86889a0bbdd5760acb3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/3b1d5b3fc1813e8ce461e86889a0bbdd5760acb3/mov.py |
"file_ext": ("mka", "mkv"), | "file_ext": ("mka", "mkv", "webm"), | def createFields(self): yield RawInt(self, 'id') yield Unsigned(self, 'size') for val in self.val[1:]: if callable(val): yield val(self) else: while not self.eof: yield EBML(self, val) | 374c6400973604655ce6edcef83589dc197647c1 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/374c6400973604655ce6edcef83589dc197647c1/mkv.py |
return False return self.stream.searchBytes('\x42\x82\x88matroska', 5*8, first._size) is not None | return "First chunk size is invalid" if self[0]['DocType/string'].value not in ('matroska', 'webm'): return "Stream isn't a matroska document." return True | def validate(self): if self.stream.readBits(0, 32, self.endian) != self.EBML_SIGNATURE: return False try: first = self[0] except ParserError: return False if None < self._size < first._size: return False return self.stream.searchBytes('\x42\x82\x88matroska', 5*8, first._size) is not None | 374c6400973604655ce6edcef83589dc197647c1 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/374c6400973604655ce6edcef83589dc197647c1/mkv.py |
if hdr['DocType/string'].value != 'matroska': raise ParserError("Stream isn't a matroska document.") | def createFields(self): hdr = EBML(self, ebml) yield hdr if hdr['DocType/string'].value != 'matroska': raise ParserError("Stream isn't a matroska document.") | 374c6400973604655ce6edcef83589dc197647c1 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/374c6400973604655ce6edcef83589dc197647c1/mkv.py | |
yield UInt32(parent, "gamma", "Gamma (x10,000)") | yield UInt32(parent, "gamma", "Gamma (x100,000)") | def gammaParse(parent): yield UInt32(parent, "gamma", "Gamma (x10,000)") | 544aa5374be2183164d12f6a24c3378b35aa7832 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/544aa5374be2183164d12f6a24c3378b35aa7832/png.py |
return float(parent["gamma"].value) / 10000 | return float(parent["gamma"].value) / 100000 | def gammaValue(parent): return float(parent["gamma"].value) / 10000 | 544aa5374be2183164d12f6a24c3378b35aa7832 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/544aa5374be2183164d12f6a24c3378b35aa7832/png.py |
if index+3 < len(text) \ | elif index+3 < len(text) \ | def parseRange(text, start): r""" >>> parseRange('[a]b', 1) (<RegexRange '[a]'>, 3) >>> parseRange('[a-z]b', 1) (<RegexRange '[a-z]'>, 5) >>> parseRange('[^a-z-]b', 1) (<RegexRange '[^a-z-]'>, 7) >>> parseRange('[^]-]b', 1) (<RegexRange '[^]-]'>, 5) """ index = start char_range = [] exclude = False if text[index] == '^... | 5f9ba2af1fc98421432c133b649b495baaf09706 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/5f9ba2af1fc98421432c133b649b495baaf09706/parser.py |
if "Duration/float" in info \ and "TimecodeScale/unsigned" in info \ and 0 < info["Duration/float"].value: try: seconds = info["Duration/float"].value * info["TimecodeScale/unsigned"].value * 1e-9 self.duration = timedelta(seconds=seconds) except OverflowError: pass | if "TimecodeScale/unsigned" in info: timecode_scale = info["TimecodeScale/unsigned"].value * 1e-9 if "Duration/float" in info: self.readDuration(info["Duration/float"].value, timecode_scale) elif "Duration/double" in info: self.readDuration(info["Duration/double"].value, timecode_scale) | def processInfo(self, info): if "Duration/float" in info \ and "TimecodeScale/unsigned" in info \ and 0 < info["Duration/float"].value: try: seconds = info["Duration/float"].value * info["TimecodeScale/unsigned"].value * 1e-9 self.duration = timedelta(seconds=seconds) except OverflowError: # Catch OverflowError for tim... | 08b95015934a813dcf9110191169cc4a80e23117 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/08b95015934a813dcf9110191169cc4a80e23117/video.py |
install_options["install_requires"] = "hachoir-core>=1.2.1" | install_options["install_requires"] = "hachoir-core>=1.3" | def main(): if "--setuptools" in argv: argv.remove("--setuptools") from setuptools import setup use_setuptools = True else: from distutils.core import setup use_setuptools = False hachoir_parser = load_source("version", path.join("hachoir_parser", "version.py")) PACKAGES = {"hachoir_parser": "hachoir_parser"} for nam... | 220337ca35c626beed28bc42e10c8f1ff75e4a85 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9327/220337ca35c626beed28bc42e10c8f1ff75e4a85/setup.py |
title=title, desc=desc, tags=tags, search_hidden=not visible, safety=safety, is_public=is_public, is_family=is_family, is_friend=is_friend, content_type=content_type) | title=title, desc=desc, tags=tags, search_hidden=not visible, safety=safety, is_public=is_public, is_family=is_family, is_friend=is_friend, content_type=content_type, progress_tracker=self.upload_progress_tracker) | def upload(self, response=None): """Upload worker function, called by the File->Upload callback. As this calls itself in the deferred callback, it takes a response argument.""" | d35662d034077396062db82bd5cff8ad3142aee6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11023/d35662d034077396062db82bd5cff8ad3142aee6/postr.py |
search_hidden=not visible, safety=safety, is_public=is_public, is_family=is_family, is_friend=is_friend, content_type=content_type) | search_hidden=not visible, safety=safety, is_public=is_public, is_family=is_family, is_friend=is_friend, content_type=content_type, progress_tracker=self.upload_progress_tracker) | def upload(self, response=None): """Upload worker function, called by the File->Upload callback. As this calls itself in the deferred callback, it takes a response argument.""" | d35662d034077396062db82bd5cff8ad3142aee6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11023/d35662d034077396062db82bd5cff8ad3142aee6/postr.py |
lock = "org.gtk.PyUnique.lock" | lock = "%s.lock" % name | def __init__(self, name, startup_id=None): gobject.GObject.__init__(self) self._is_running = False self._name = name self._screen = gdk.screen_get_default() | d24d90e1a73ed3b8d5069a3c15df983cd4119f2d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11023/d24d90e1a73ed3b8d5069a3c15df983cd4119f2d/PyUnique.py |
import logging LOG_FILENAME = '/tmp/openid.log' logging.basicConfig(filename=LOG_FILENAME,level=logging.DEBUG) logging.debug("%s %s" % (identity_url, openid.openid)) | def openid_is_authorized(req, identity_url, trust_root): """ Check that they own the given identity URL, and that the trust_root is in their whitelist of trusted sites. """ if not req.user.is_authenticated(): return None openid = openid_get_identity(req, identity_url) if openid is None: return None import logging LOG... | 997a2f2657d1af4f794786b9a8fca8fc73c59ca5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3362/997a2f2657d1af4f794786b9a8fca8fc73c59ca5/views.py | |
url = '%s?%s=%s' % (login_url, REDIRECT_FIELD_NAME, path) | url = '%s?%s=%s' % (login_url, REDIRECT_FIELD_NAME, urlquote(path)) | def landing_page(request, orequest): """ The page shown when the user attempts to sign in somewhere using OpenID but is not authenticated with the site. For idproxy.net, a message telling them to log in manually is displayed. """ request.session['OPENID_REQUEST'] = orequest login_url = settings.LOGIN_URL path = request... | 60007c126a62178e2e9fc03ba51bdef45d68e0e1 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3362/60007c126a62178e2e9fc03ba51bdef45d68e0e1/views.py |
self._check_for_errors(stderr) | def _download(self): self.log = '' self.information['status'] = DownloadStatus.RUNNING if self.information['download_type'] == DownloadTypes.TORRENT: # download torrent if necessary torrent_filename = os.path.join(self._config.get('general', 'folder_new_otrkeys'), self.filename + '.torrent') if not os.path.exists(tor... | 40ba3da8572c68666ad2bddd9b87704d260d2b86 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11048/40ba3da8572c68666ad2bddd9b87704d260d2b86/downloader.py | |
This path is by default <mfm_lib_path>/../data/ in trunk and /usr/share/mfm in an installed version but this path | This path is by default <otrverwaltung_lib_path>/../data/ in trunk and /usr/share/otrverwaltung in an installed version but this path | def getdatapath(*args): """Retrieve otrverwaltung data path This path is by default <mfm_lib_path>/../data/ in trunk and /usr/share/mfm in an installed version but this path is specified at installation time. """ return os.path.join(os.path.dirname(__file__), data_dir, *args) | c5db042b0f3a8def985930d720d6c9e7b8ee0837 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11048/c5db042b0f3a8def985930d720d6c9e7b8ee0837/path.py |
self.combobox_archive.fill(archive_directory) self.combobox_archive.set_active(0) self.combobox_archive.connect('changed', self._on_combobox_archive_changed) | if action != Action.DECODE: self.combobox_archive.fill(archive_directory) self.combobox_archive.set_active(0) self.combobox_archive.connect('changed', self._on_combobox_archive_changed) | def _run(self, file_conclusions, action, rename_by_schema, archive_directory): self.action = action self.rename_by_schema = rename_by_schema self.__file_conclusions = file_conclusions self.forward_clicks = 0 self.show_all() self.combobox_archive.fill(archive_directory) self.combobox_archive.set_active(0) self.combobo... | 270edd8c959fe53e05f88f1d050c4a41c66345e2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11048/270edd8c959fe53e05f88f1d050c4a41c66345e2/ConclusionDialog.py |
self.builder.get_object('button_play_cut').props.visible = cut_ok | self.builder.get_object('button_play_cut').props.visible = cut_ok self.builder.get_object('box_archive').props.visible = cut_ok | def show_conclusion(self, new_iter): self.conclusion_iter = new_iter self.file_conclusion = self.__file_conclusions[self.conclusion_iter] self.builder.get_object('label_count').set_text("Zeige Datei %s/%s" % (str(new_iter + 1), len(self.__file_conclusions))) | 270edd8c959fe53e05f88f1d050c4a41c66345e2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11048/270edd8c959fe53e05f88f1d050c4a41c66345e2/ConclusionDialog.py |
url = "%sgetxml.php?ofsb=%s" % (server, str(size)) | urls = ["%sgetxml.php?ofsb=%s" % (server, str(size)), "%sgetxml.php?ofsb=%s" % (server, str((size+2*1024**3)%(4*1024**3)- 2*1024**3))] | def download_cutlists(filename, server, choose_cutlists_by, cutlist_mp4_as_hq, error_cb=None, cutlist_found_cb=None): """ Downloads all cutlists for the given file. filename - movie filename server - cutlist server choose_cutlists_by - 0 by size, 1 by name error_cb - callback: an err... | 7bfd0d18856929e717091080e4f935f644384838 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11048/7bfd0d18856929e717091080e4f935f644384838/cutlists.py |
url = "%sgetxml.php?name=%s" % (server, root) print url try: handle = urllib.urlopen(url) except IOError: if error_cb: error_cb("Verbindungsprobleme") return "Verbindungsprobleme", None | urls = ["%sgetxml.php?name=%s" % (server, root)] cutlists = [] for url in urls: print "[Cutlists] Download by : %s" % url try: handle = urllib.urlopen(url) except IOError: if error_cb: error_cb("Verbindungsprobleme") return "Verbindungsprobleme", None try: dom_cutlists = xml.dom.minidom.parse(handle) handle.close() ... | def download_cutlists(filename, server, choose_cutlists_by, cutlist_mp4_as_hq, error_cb=None, cutlist_found_cb=None): """ Downloads all cutlists for the given file. filename - movie filename server - cutlist server choose_cutlists_by - 0 by size, 1 by name error_cb - callback: an err... | 7bfd0d18856929e717091080e4f935f644384838 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11048/7bfd0d18856929e717091080e4f935f644384838/cutlists.py |
try: dom_cutlists = xml.dom.minidom.parse(handle) handle.close() dom_cutlists = dom_cutlists.getElementsByTagName('cutlist') except: if error_cb: error_cb("Keine Cutlists gefunden") return "Keine Cutlists gefunden", None cutlists = [] for cutlist in dom_cutlists: c = Cutlist() c.id = __read_v... | cutlists.append(c) | def download_cutlists(filename, server, choose_cutlists_by, cutlist_mp4_as_hq, error_cb=None, cutlist_found_cb=None): """ Downloads all cutlists for the given file. filename - movie filename server - cutlist server choose_cutlists_by - 0 by size, 1 by name error_cb - callback: an err... | 7bfd0d18856929e717091080e4f935f644384838 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11048/7bfd0d18856929e717091080e4f935f644384838/cutlists.py |
def foreach(model, path, iter, data=None): index = model.get_value(iter, 0) stamp = self.app.planned_broadcasts[index].datetime if stamp < now: selection.select_iter(iter) self.builder.get_object('treeview_planning').get_model().foreach(foreach) | for row in self.builder.get_object('treeview_planning').get_model(): if row[0].datetime < now: selection.select_iter(row.iter) | def foreach(model, path, iter, data=None): index = model.get_value(iter, 0) stamp = self.app.planned_broadcasts[index].datetime | f56f6cfdeb741159c9a70341d049ed00110310e4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11048/f56f6cfdeb741159c9a70341d049ed00110310e4/MainWindow.py |
return "%s %s" % (self.user.username, self.group) | if self.user: username = self.user.username else: username = 'anonymous return "%s %s" % (username, self.group) | def __unicode__(self): return "%s %s" % (self.user.username, self.group) | 7b6e2c7c64ef0ff59d59e18b1cd9498d2d38889a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8053/7b6e2c7c64ef0ff59d59e18b1cd9498d2d38889a/models.py |
raise | if settings.DEBUG: raise l.warning("Can't find the GoalType named %s" % goal_name) | def record(cls, goal_name, experiment_user): try: return cls._record(goal_name, experiment_user) except GoalType.DoesNotExist: raise except Exception, e: l.error("Unexpected exception in GoalRecord.record:\n" "%s" % traceback.format_exc) | e2556a89a4896c445138be295a28133b724376d2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8053/e2556a89a4896c445138be295a28133b724376d2/models.py |
goal_types = [GoalType.objects.create(name=i) for i in range(3)] | goal_types = [GoalType.objects.create(name=str(i)) for i in range(3)] | def testParticipantConversionCalculator(self): goal_types = [GoalType.objects.create(name=i) for i in range(3)] anonymous_visitor = AnonymousVisitor.objects.create() participant = self.create_participant( anonymous_visitor=anonymous_visitor, experiment=self.experiment, enrollment_date=self.experiment.start_date + timed... | 4d2693ab2ff9f9ff588c67317843b363d670f0b3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8053/4d2693ab2ff9f9ff588c67317843b363d670f0b3/test_daily_report.py |
l.error("Unexpected exception in GoalRecord.record:\n" "%s" % traceback.format_exc) | l.exception("Unexpected exception in GoalRecord.record") | def record(cls, goal_name, experiment_user): try: return cls._record(goal_name, experiment_user) except GoalType.DoesNotExist: if settings.DEBUG: raise l.warning("Can't find the GoalType named %s" % goal_name) except Exception, e: l.error("Unexpected exception in GoalRecord.record:\n" "%s" % traceback.format_exc) | a4abc5c3ab2bce477fe2a191bfefa08b542e09b3 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8053/a4abc5c3ab2bce477fe2a191bfefa08b542e09b3/models.py |
scores a, and b. From Numerical Recipies, p.483. If printit=1, results are printed to the screen. If printit='filename', the results are output to 'filename' using the given writemode (default=append). Returns t-value, and prob. Originally written by Gary Strangman. Usage: lttest_ind(a,b,printit=0,name1='Samp1'... | scores a, and b. Returns t-value, and prob. Originally written by Gary Strangman. Usage: lttest_ind(a,b) | def ttest_ind(a, b): """ Calculates the t-obtained T-test on TWO INDEPENDENT samples of scores a, and b. From Numerical Recipies, p.483. If printit=1, results are printed to the screen. If printit='filename', the results are output to 'filename' using the given writemode (default=append). Returns t-value, and prob.... | ac91143735ceeb00eab3b860e39630663bf9dd5b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8053/ac91143735ceeb00eab3b860e39630663bf9dd5b/stats.py |
x1 = mean(a) x2 = mean(b) v1 = stdev(a)**2 v2 = stdev(b)**2 n1 = len(a) n2 = len(b) | x1, x2 = mean(a), mean(b) v1, v2 = stdev(a)**2, stdev(b)**2 n1, n2 = len(a), len(b) | def ttest_ind(a, b): """ Calculates the t-obtained T-test on TWO INDEPENDENT samples of scores a, and b. From Numerical Recipies, p.483. If printit=1, results are printed to the screen. If printit='filename', the results are output to 'filename' using the given writemode (default=append). Returns t-value, and prob.... | ac91143735ceeb00eab3b860e39630663bf9dd5b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8053/ac91143735ceeb00eab3b860e39630663bf9dd5b/stats.py |
svar = ((n1-1)*v1+(n2-1)*v2)/float(df) t = (x1-x2)/sqrt(svar*(1.0/n1 + 1.0/n2)) | try: svar = ((n1-1)*v1+(n2-1)*v2)/float(df) except ZeroDivisionError: return float('nan'), float('nan') try: t = (x1-x2)/sqrt(svar*(1.0/n1 + 1.0/n2)) except ZeroDivisionError: t = 1.0 | def ttest_ind(a, b): """ Calculates the t-obtained T-test on TWO INDEPENDENT samples of scores a, and b. From Numerical Recipies, p.483. If printit=1, results are printed to the screen. If printit='filename', the results are output to 'filename' using the given writemode (default=append). Returns t-value, and prob.... | ac91143735ceeb00eab3b860e39630663bf9dd5b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8053/ac91143735ceeb00eab3b860e39630663bf9dd5b/stats.py |
p = subprocess.Popen(['git', 'log', '-1'], stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=(sys.platform=='win32')) | proc = subprocess.Popen(['git', 'log', '-999'], stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=(sys.platform=='win32')) for line in proc.stdout: match = git_re.search(line) if match: id = match.group(2) if id: proc.stdout.close() return id | def git_fetch_id(): """ Fetch the GIT identifier for the local tree. Errors are swallowed. """ try: p = subprocess.Popen(['git', 'log', '-1'], stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=(sys.platform=='win32')) except OSError: # 'git' is apparently either not installed or not executable. return None id = No... | 34035cf0ee36bfd3adf17a0949c62a9ce7409a8b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/34035cf0ee36bfd3adf17a0949c62a9ce7409a8b/lastchange.py |
return None id = None if p: git_re = re.compile('^\s*git-svn-id:\s+(\S+)@(\d+)', re.M) m = git_re.search(p.stdout.read()) if m: id = m.group(2) return id | pass return None | def git_fetch_id(): """ Fetch the GIT identifier for the local tree. Errors are swallowed. """ try: p = subprocess.Popen(['git', 'log', '-1'], stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=(sys.platform=='win32')) except OSError: # 'git' is apparently either not installed or not executable. return None id = No... | 34035cf0ee36bfd3adf17a0949c62a9ce7409a8b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/34035cf0ee36bfd3adf17a0949c62a9ce7409a8b/lastchange.py |
urllib.urlretrieve(download_url, BUILD_ZIP_NAME) | urllib.urlretrieve(download_url, BUILD_ZIP_NAME, _Reporthook) print | def TryRevision(rev, profile, args): """Downloads revision |rev|, unzips it, and opens it for the user to test. |profile| is the profile to use.""" # Do this in a temp dir so we don't collide with user files. cwd = os.getcwd() tempdir = tempfile.mkdtemp(prefix='bisect_tmp') os.chdir(tempdir) # Download the file. downl... | 5145b406f16c6106365dbfe4012045da97ddf6aa /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/5145b406f16c6106365dbfe4012045da97ddf6aa/build-bisect.py |
def RewritePath(path, sysroot): """Rewrites a path by prefixing it with the sysroot if it is absolute.""" | def RewritePath(path, opts): """Rewrites a path by stripping the prefix and prepending the sysroot.""" sysroot = opts.sysroot prefix = opts.strip_prefix | def RewritePath(path, sysroot): """Rewrites a path by prefixing it with the sysroot if it is absolute.""" if os.path.isabs(path) and not path.startswith(sysroot): path = path.lstrip('/') return os.path.join(sysroot, path) else: return path | 789e97c6a87570a68b1d21d2be3b108097fe22f8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/789e97c6a87570a68b1d21d2be3b108097fe22f8/rewrite_dirs.py |
def RewriteLine(line, sysroot): | def RewriteLine(line, opts): | def RewriteLine(line, sysroot): """Rewrites all the paths in recognized options.""" args = line.split() count = len(args) i = 0 while i < count: for prefix in REWRITE_PREFIX: # The option can be either in the form "-I /path/to/dir" or # "-I/path/to/dir" so handle both. if args[i] == prefix: i += 1 try: args[i] = Rewrit... | 789e97c6a87570a68b1d21d2be3b108097fe22f8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/789e97c6a87570a68b1d21d2be3b108097fe22f8/rewrite_dirs.py |
args[i] = RewritePath(args[i], sysroot) | args[i] = RewritePath(args[i], opts) | def RewriteLine(line, sysroot): """Rewrites all the paths in recognized options.""" args = line.split() count = len(args) i = 0 while i < count: for prefix in REWRITE_PREFIX: # The option can be either in the form "-I /path/to/dir" or # "-I/path/to/dir" so handle both. if args[i] == prefix: i += 1 try: args[i] = Rewrit... | 789e97c6a87570a68b1d21d2be3b108097fe22f8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/789e97c6a87570a68b1d21d2be3b108097fe22f8/rewrite_dirs.py |
args[i] = prefix + RewritePath(args[i][len(prefix):], sysroot) | args[i] = prefix + RewritePath(args[i][len(prefix):], opts) | def RewriteLine(line, sysroot): """Rewrites all the paths in recognized options.""" args = line.split() count = len(args) i = 0 while i < count: for prefix in REWRITE_PREFIX: # The option can be either in the form "-I /path/to/dir" or # "-I/path/to/dir" so handle both. if args[i] == prefix: i += 1 try: args[i] = Rewrit... | 789e97c6a87570a68b1d21d2be3b108097fe22f8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/789e97c6a87570a68b1d21d2be3b108097fe22f8/rewrite_dirs.py |
try: sysroot = argv[1] except IndexError: sys.stderr.write('usage: %s /path/to/sysroot\n' % argv[0]) return 1 | parser = optparse.OptionParser() parser.add_option('-s', '--sysroot', default='/', help='sysroot to prepend') parser.add_option('-p', '--strip-prefix', default='', help='prefix to strip') opts, args = parser.parse_args(argv[1:]) | def main(argv): try: sysroot = argv[1] except IndexError: sys.stderr.write('usage: %s /path/to/sysroot\n' % argv[0]) return 1 for line in sys.stdin.readlines(): line = RewriteLine(line.strip(), sysroot) print line return 0 | 789e97c6a87570a68b1d21d2be3b108097fe22f8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/789e97c6a87570a68b1d21d2be3b108097fe22f8/rewrite_dirs.py |
line = RewriteLine(line.strip(), sysroot) | line = RewriteLine(line.strip(), opts) | def main(argv): try: sysroot = argv[1] except IndexError: sys.stderr.write('usage: %s /path/to/sysroot\n' % argv[0]) return 1 for line in sys.stdin.readlines(): line = RewriteLine(line.strip(), sysroot) print line return 0 | 789e97c6a87570a68b1d21d2be3b108097fe22f8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/789e97c6a87570a68b1d21d2be3b108097fe22f8/rewrite_dirs.py |
BUILD_EXE_NAME = "chrome" | BUILD_EXE_NAME = 'chrome' | def SetArchiveVars(archive): """Set a bunch of global variables appropriate for the specified archive.""" global BUILD_ARCHIVE_TYPE global BUILD_ARCHIVE_DIR global BUILD_ZIP_NAME global BUILD_DIR_NAME global BUILD_EXE_NAME global BUILD_BASE_URL BUILD_ARCHIVE_TYPE = archive BUILD_ARCHIVE_DIR = 'chromium-rel-' + BUILD_A... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
BUILD_EXE_NAME = "Chromium.app/Contents/MacOS/Chromium" | BUILD_EXE_NAME = 'Chromium.app/Contents/MacOS/Chromium' | def SetArchiveVars(archive): """Set a bunch of global variables appropriate for the specified archive.""" global BUILD_ARCHIVE_TYPE global BUILD_ARCHIVE_DIR global BUILD_ZIP_NAME global BUILD_DIR_NAME global BUILD_EXE_NAME global BUILD_BASE_URL BUILD_ARCHIVE_TYPE = archive BUILD_ARCHIVE_DIR = 'chromium-rel-' + BUILD_A... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
BUILD_EXE_NAME = "chrome.exe" | BUILD_EXE_NAME = 'chrome.exe' | def SetArchiveVars(archive): """Set a bunch of global variables appropriate for the specified archive.""" global BUILD_ARCHIVE_TYPE global BUILD_ARCHIVE_DIR global BUILD_ZIP_NAME global BUILD_DIR_NAME global BUILD_EXE_NAME global BUILD_BASE_URL BUILD_ARCHIVE_TYPE = archive BUILD_ARCHIVE_DIR = 'chromium-rel-' + BUILD_A... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
print("Could not retrieve the download. Sorry.") | print('Could not retrieve the download. Sorry.') | def TryRevision(rev, profile, args): """Downloads revision |rev|, unzips it, and opens it for the user to test. |profile| is the profile to use.""" # Do this in a temp dir so we don't collide with user files. cwd = os.getcwd() tempdir = tempfile.mkdtemp(prefix='bisect_tmp') os.chdir(tempdir) # Download the file. downl... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
print 'Unzipping ...' os.system("unzip -q %s" % BUILD_ZIP_NAME) | print 'Unziping ...' UnzipFilenameToDir(BUILD_ZIP_NAME, os.curdir) | def TryRevision(rev, profile, args): """Downloads revision |rev|, unzips it, and opens it for the user to test. |profile| is the profile to use.""" # Do this in a temp dir so we don't collide with user files. cwd = os.getcwd() tempdir = tempfile.mkdtemp(prefix='bisect_tmp') os.chdir(tempdir) # Download the file. downl... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
print 'Running %s/%s/%s %s' % (os.getcwd(), BUILD_DIR_NAME, BUILD_EXE_NAME, flags) if BUILD_ARCHIVE_TYPE in ('linux', 'linux-64', 'mac'): os.system("%s/%s %s" % (BUILD_DIR_NAME, BUILD_EXE_NAME, flags)) elif BUILD_ARCHIVE_TYPE in ('xp'): os.system("%s/%s %s" % (BUILD_DIR_NAME, BUILD_EXE_NAME, flags)) | exe = os.path.join(os.getcwd(), BUILD_DIR_NAME, BUILD_EXE_NAME) cmd = '%s %s' % (exe, flags) print 'Running %s' % cmd os.system(cmd) | def TryRevision(rev, profile, args): """Downloads revision |rev|, unzips it, and opens it for the user to test. |profile| is the profile to use.""" # Do this in a temp dir so we don't collide with user files. cwd = os.getcwd() tempdir = tempfile.mkdtemp(prefix='bisect_tmp') os.chdir(tempdir) # Download the file. downl... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
response = raw_input("\nBuild %d is [(g)ood/(b)ad]: " % int(rev)) if response and response in ("g", "b"): return response == "g" | response = raw_input('\nBuild %d is [(g)ood/(b)ad]: ' % int(rev)) if response and response in ('g', 'b'): return response == 'g' | def AskIsGoodBuild(rev): """Ask the user whether build |rev| is good or bad.""" # Loop until we get a response that we can parse. while True: response = raw_input("\nBuild %d is [(g)ood/(b)ad]: " % int(rev)) if response and response in ("g", "b"): return response == "g" | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
BUILD_LATEST_URL = "%s/LATEST" % (BUILD_BASE_URL) | BUILD_LATEST_URL = '%s/LATEST' % (BUILD_BASE_URL) | def main(): usage = ('%prog [options] [-- chromium-options]\n' 'Perform binary search on the snapshot builds.') parser = optparse.OptionParser(usage=usage) # Strangely, the default help output doesn't include the choice list. choices = ['mac', 'xp', 'linux', 'linux-64'] parser.add_option('-a', '--archive', choices = ch... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
bad_rev = raw_input("Bad revision [HEAD:%d]: " % latest) if (bad_rev == ""): | bad_rev = raw_input('Bad revision [HEAD:%d]: ' % latest) if (bad_rev == ''): | def main(): usage = ('%prog [options] [-- chromium-options]\n' 'Perform binary search on the snapshot builds.') parser = optparse.OptionParser(usage=usage) # Strangely, the default help output doesn't include the choice list. choices = ['mac', 'xp', 'linux', 'linux-64'] parser.add_option('-a', '--archive', choices = ch... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
print("Could not determine latest revision. This could be bad...") bad_rev = int(raw_input("Bad revision: ")) | print('Could not determine latest revision. This could be bad...') bad_rev = int(raw_input('Bad revision: ')) | def main(): usage = ('%prog [options] [-- chromium-options]\n' 'Perform binary search on the snapshot builds.') parser = optparse.OptionParser(usage=usage) # Strangely, the default help output doesn't include the choice list. choices = ['mac', 'xp', 'linux', 'linux-64'] parser.add_option('-a', '--archive', choices = ch... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
good_rev = int(raw_input("Last known good [0]: ")) | good_rev = int(raw_input('Last known good [0]: ')) | def main(): usage = ('%prog [options] [-- chromium-options]\n' 'Perform binary search on the snapshot builds.') parser = optparse.OptionParser(usage=usage) # Strangely, the default help output doesn't include the choice list. choices = ['mac', 'xp', 'linux', 'linux-64'] parser.add_option('-a', '--archive', choices = ch... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
print "We don't have enough builds to bisect. revlist: %s" % revlist | print 'We don\'t have enough builds to bisect. revlist: %s' % revlist | def main(): usage = ('%prog [options] [-- chromium-options]\n' 'Perform binary search on the snapshot builds.') parser = optparse.OptionParser(usage=usage) # Strangely, the default help output doesn't include the choice list. choices = ['mac', 'xp', 'linux', 'linux-64'] parser.add_option('-a', '--archive', choices = ch... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
print("%d candidates. %d tries left." % | print('%d candidates. %d tries left.' % | def main(): usage = ('%prog [options] [-- chromium-options]\n' 'Perform binary search on the snapshot builds.') parser = optparse.OptionParser(usage=usage) # Strangely, the default help output doesn't include the choice list. choices = ['mac', 'xp', 'linux', 'linux-64'] parser.add_option('-a', '--archive', choices = ch... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
print("Candidates: %s" % revlist[good:bad]) | print('Candidates: %s' % revlist[good:bad]) | def main(): usage = ('%prog [options] [-- chromium-options]\n' 'Perform binary search on the snapshot builds.') parser = optparse.OptionParser(usage=usage) # Strangely, the default help output doesn't include the choice list. choices = ['mac', 'xp', 'linux', 'linux-64'] parser.add_option('-a', '--archive', choices = ch... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
print("You are probably looking for build %d." % revlist[bad]) print("CHANGELOG URL:") | print('You are probably looking for build %d.' % revlist[bad]) print('CHANGELOG URL:') | def main(): usage = ('%prog [options] [-- chromium-options]\n' 'Perform binary search on the snapshot builds.') parser = optparse.OptionParser(usage=usage) # Strangely, the default help output doesn't include the choice list. choices = ['mac', 'xp', 'linux', 'linux-64'] parser.add_option('-a', '--archive', choices = ch... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
print("Built at revision:") | print('Built at revision:') | def main(): usage = ('%prog [options] [-- chromium-options]\n' 'Perform binary search on the snapshot builds.') parser = optparse.OptionParser(usage=usage) # Strangely, the default help output doesn't include the choice list. choices = ['mac', 'xp', 'linux', 'linux-64'] parser.add_option('-a', '--archive', choices = ch... | 68071845423fd40c390833991c2dd5f14df43ab8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/68071845423fd40c390833991c2dd5f14df43ab8/build-bisect.py |
if BUILD_ARCHIVE_TYPE in ('linux', 'linux-64'): | if BUILD_ARCHIVE_TYPE in ('linux', 'linux-64', 'linux-chromiumos'): | def SetArchiveVars(archive): """Set a bunch of global variables appropriate for the specified archive.""" global BUILD_ARCHIVE_TYPE global BUILD_ARCHIVE_DIR global BUILD_ZIP_NAME global BUILD_DIR_NAME global BUILD_EXE_NAME global BUILD_BASE_URL BUILD_ARCHIVE_TYPE = archive BUILD_ARCHIVE_DIR = 'chromium-rel-' + BUILD_A... | bb46e0d24a1c87d3a627ae0b4832c44d184888e0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/bb46e0d24a1c87d3a627ae0b4832c44d184888e0/build-bisect.py |
choices = ['mac', 'xp', 'linux', 'linux-64'] | choices = ['mac', 'xp', 'linux', 'linux-64', 'linux-chromiumos'] | def main(): usage = ('%prog [options] [-- chromium-options]\n' 'Perform binary search on the snapshot builds.') parser = optparse.OptionParser(usage=usage) # Strangely, the default help output doesn't include the choice list. choices = ['mac', 'xp', 'linux', 'linux-64'] parser.add_option('-a', '--archive', choices = ch... | bb46e0d24a1c87d3a627ae0b4832c44d184888e0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6097/bb46e0d24a1c87d3a627ae0b4832c44d184888e0/build-bisect.py |
arg = len(names) | arg = len(proposals) | def code_assist(self, prefix): proposals = self._calculate_proposals() if prefix is not None: arg = self.env.prefix_value(prefix) if arg == 0: arg = len(names) common_start = self._calculate_prefix(proposals[:arg]) self.env.insert(common_start[self.offset - self.starting_offset:]) self._starting = common_start self._of... | a597b84aff7d22ed14c01ab3c54a0919fb04d92d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8285/a597b84aff7d22ed14c01ab3c54a0919fb04d92d/interface.py |
proposals = codeassist.sorted_proposals(proposals) | if self.env.get('sorted_completions', True): proposals = codeassist.sorted_proposals(proposals) | def _calculate_proposals(self): self.interface._check_project() resource = self.interface.resource maxfixes = self.env.get('codeassist_maxfixes') proposals = codeassist.code_assist( self.interface.project, self.source, self.offset, resource, maxfixes=maxfixes) proposals = codeassist.sorted_proposals(proposals) if self.... | 00d74e27b1db60cd70fc51f9bfcbf2420cda2f65 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8285/00d74e27b1db60cd70fc51f9bfcbf2420cda2f65/interface.py |
response = self.requestor.request('execute_cql_query', request_params) | try: response = self.requestor.request('execute_cql_query', request_params) except AvroRemoteException, are: raise CQLException(are) | def execute(self, query, compression=None): compress = compression is None and DEFAULT_COMPRESSION \ or compression.upper() if not compress in COMPRESSION_SCHEMES: raise InvalidCompressionScheme(compress) compressed_query = Connection.compress_query(query, compress) request_params = dict(query=compressed_query, compre... | 4d8937a8690423e01777877132c1feb2e7e3340c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5004/4d8937a8690423e01777877132c1feb2e7e3340c/__init__.py |
newSnpData = SNPData(col_id_ls=copy.deepcopy(snpData.col_id_ls), row_id_ls=[]) newSnpData.data_matrix = num.zeros([no_of_rows, no_of_cols], num.int8) | new_col_id_ls = copy.deepcopy(snpData.col_id_ls) new_row_id_ls = [] new_data_matrix = num.zeros([no_of_rows, no_of_cols], num.int8) | def keepRowsByRowID(cls, snpData, row_id_ls): """ 2009-05-19 keep certain rows in snpData given row_id_ls """ sys.stderr.write("Keeping rows given row_id_ls ...") no_of_rows = len(row_id_ls) row_id_wanted_set = set(row_id_ls) no_of_cols = len(snpData.col_id_ls) newSnpData = SNPData(col_id_ls=copy.deepcopy(snpData.col_... | 254211f49fa61cb0238f804701edac97473331a0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/254211f49fa61cb0238f804701edac97473331a0/SNP.py |
newSnpData.row_id_ls.append(row_id) newSnpData.data_matrix[row_index] = snpData.data_matrix[i] | new_row_id_ls.append(row_id) new_data_matrix[row_index] = snpData.data_matrix[i] | def keepRowsByRowID(cls, snpData, row_id_ls): """ 2009-05-19 keep certain rows in snpData given row_id_ls """ sys.stderr.write("Keeping rows given row_id_ls ...") no_of_rows = len(row_id_ls) row_id_wanted_set = set(row_id_ls) no_of_cols = len(snpData.col_id_ls) newSnpData = SNPData(col_id_ls=copy.deepcopy(snpData.col_... | 254211f49fa61cb0238f804701edac97473331a0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/254211f49fa61cb0238f804701edac97473331a0/SNP.py |
report=1, run_type=1): """ | cofactor_phenotype_id_ls=[], report=1, run_type=1,): """ 2010-1-11 add argument cofactor_phenotype_id_ls to have the functionality of treating some phenotypes as cofactors | def cofactorLM(cls, genotype_fname, phenotype_fname, phenotype_method_id_ls, output_fname_prefix, start_snp, stop_snp, cofactors=[],\ report=1, run_type=1): """ 2009-8-26 run_type 1: pure linear model by python run_type 2: EMMA run_type 3: pure linear model by R (Field test shows run_type 3 is same as 1.) start_snp an... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
run_type 1: pure linear model by python run_type 2: EMMA run_type 3: pure linear model by R (Field test shows run_type 3 is same as 1.) | one phenotype at a time: 1. create a new SNP matrix which includes accessions whose phenotypes (this phenotype + cofactor_phenotype_id_ls) are non-NA 2. one SNP at a time 1. add cofactor SNP matrix if cofactors are present 2. add cofactor phenotype matrix if cofactor_phenotype_id_ls exist 3. run association parameter ... | def cofactorLM(cls, genotype_fname, phenotype_fname, phenotype_method_id_ls, output_fname_prefix, start_snp, stop_snp, cofactors=[],\ report=1, run_type=1): """ 2009-8-26 run_type 1: pure linear model by python run_type 2: EMMA run_type 3: pure linear model by R (Field test shows run_type 3 is same as 1.) start_snp an... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
chromosome, start_pos = start_snp.split('_')[:2] | start_chr, start_pos = start_snp.split('_')[:2] start_chr = int(start_chr) | def cofactorLM(cls, genotype_fname, phenotype_fname, phenotype_method_id_ls, output_fname_prefix, start_snp, stop_snp, cofactors=[],\ report=1, run_type=1): """ 2009-8-26 run_type 1: pure linear model by python run_type 2: EMMA run_type 3: pure linear model by R (Field test shows run_type 3 is same as 1.) start_snp an... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
stop_pos = int(stop_snp.split('_')[1]) | stop_chr, stop_pos = stop_snp.split('_')[:2] stop_chr = int(stop_chr) stop_pos = int(stop_pos) | def cofactorLM(cls, genotype_fname, phenotype_fname, phenotype_method_id_ls, output_fname_prefix, start_snp, stop_snp, cofactors=[],\ report=1, run_type=1): """ 2009-8-26 run_type 1: pure linear model by python run_type 2: EMMA run_type 3: pure linear model by R (Field test shows run_type 3 is same as 1.) start_snp an... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
if not numpy.isnan(phenotype_ls[i]): | this_row_has_NA_phenotype = False if cofactor_phenotype_index_ls: for phenotype_index in cofactor_phenotype_index_ls: if numpy.isnan(initData.phenData.data_matrix[i, phenotype_index]): this_row_has_NA_phenotype = True break if numpy.isnan(phenotype_ls[i]): this_row_has_NA_phenotype = True if not this_row_has_NA_pheno... | def cofactorLM(cls, genotype_fname, phenotype_fname, phenotype_method_id_ls, output_fname_prefix, start_snp, stop_snp, cofactors=[],\ report=1, run_type=1): """ 2009-8-26 run_type 1: pure linear model by python run_type 2: EMMA run_type 3: pure linear model by R (Field test shows run_type 3 is same as 1.) start_snp an... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
if chr==chromosome and pos>=start_pos and pos<=stop_pos: | if chr>=start_chr and chr<=stop_chr and pos>=start_pos and pos<=stop_pos: | def cofactorLM(cls, genotype_fname, phenotype_fname, phenotype_method_id_ls, output_fname_prefix, start_snp, stop_snp, cofactors=[],\ report=1, run_type=1): """ 2009-8-26 run_type 1: pure linear model by python run_type 2: EMMA run_type 3: pure linear model by R (Field test shows run_type 3 is same as 1.) start_snp an... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
phenotype_method_id_ls = [43] | phenotype_method_id_ls = [285] | def cofactorLM(cls, genotype_fname, phenotype_fname, phenotype_method_id_ls, output_fname_prefix, start_snp, stop_snp, cofactors=[],\ report=1, run_type=1): """ 2009-8-26 run_type 1: pure linear model by python run_type 2: EMMA run_type 3: pure linear model by R (Field test shows run_type 3 is same as 1.) start_snp an... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
GWA.cofactorLM(genotype_fname, phenotype_fname, phenotype_method_id_ls, output_fname_prefix, start_snp, stop_snp, cofactors=cofactors) | cofactor_phenotype_id_ls = [77] GWA.cofactorLM(genotype_fname, phenotype_fname, phenotype_method_id_ls, output_fname_prefix, start_snp, stop_snp, \ cofactors=cofactors, cofactor_phenotype_id_ls=cofactor_phenotype_id_ls) | def cofactorLM(cls, genotype_fname, phenotype_fname, phenotype_method_id_ls, output_fname_prefix, start_snp, stop_snp, cofactors=[],\ report=1, run_type=1): """ 2009-8-26 run_type 1: pure linear model by python run_type 2: EMMA run_type 3: pure linear model by R (Field test shows run_type 3 is same as 1.) start_snp an... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
max_diff_perc=0.10, min_no_of_probes=5, count_embedded_segment_as_match=False): """ | max_diff_perc=0.10, min_no_of_probes=5, count_embedded_segment_as_match=False, \ min_reciprocal_overlap=0.6, report=True): """ 2010-1-26 value of the ecotype_id2cnv_qc_call_data dictionary is a RBDict (RBTree dictionary) structure. 2009-12-8 add argument min_reciprocal_overlap | def compareCNVSegmentsAgainstQCHandler(cls, input_fname_ls, ecotype_id2cnv_qc_call_data, function_handler, param_obj, \ deletion_cutoff=None, max_boundary_diff=10000, \ max_diff_perc=0.10, min_no_of_probes=5, count_embedded_segment_as_match=False): """ 2009-11-4 a general handler to compare CNV segments from input_fnam... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
if boundary_diff1<=max_boundary_diff and boundary_diff2<=max_boundary_diff and diff1_perc<=max_diff_perc and \ diff2_perc<=max_diff_perc: | is_overlap = is_reciprocal_overlap([segment_start_pos, segment_stop_pos], [qc_start, qc_stop], \ min_reciprocal_overlap=min_reciprocal_overlap) if is_overlap: | def compareCNVSegmentsAgainstQCHandler(cls, input_fname_ls, ecotype_id2cnv_qc_call_data, function_handler, param_obj, \ deletion_cutoff=None, max_boundary_diff=10000, \ max_diff_perc=0.10, min_no_of_probes=5, count_embedded_segment_as_match=False): """ 2009-11-4 a general handler to compare CNV segments from input_fnam... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
elif count_embedded_segment_as_match and segment_start_pos>=qc_start and segment_stop_pos<=qc_stop: no_of_valid_deletions += 1 valid_match = True | def compareCNVSegmentsAgainstQCHandler(cls, input_fname_ls, ecotype_id2cnv_qc_call_data, function_handler, param_obj, \ deletion_cutoff=None, max_boundary_diff=10000, \ max_diff_perc=0.10, min_no_of_probes=5, count_embedded_segment_as_match=False): """ 2009-11-4 a general handler to compare CNV segments from input_fnam... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py | |
cnv_segment_obj = PassingData(ecotype_id=cnv_ecotype_id, start_probe=start_probe, stop_probe=stop_probe,\ no_of_probes=no_of_probes, amplitude=amplitude, segment_length=segment_length,\ segment_chromosome=segment_chromosome, ) function_handler(cnv_segment_obj, cnv_qc_call, param_obj) | function_handler(param_obj, cnv_segment_obj, cnv_qc_call, ) | def compareCNVSegmentsAgainstQCHandler(cls, input_fname_ls, ecotype_id2cnv_qc_call_data, function_handler, param_obj, \ deletion_cutoff=None, max_boundary_diff=10000, \ max_diff_perc=0.10, min_no_of_probes=5, count_embedded_segment_as_match=False): """ 2009-11-4 a general handler to compare CNV segments from input_fnam... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
if counter%10000==0: | """ if report and counter%10000==0: | def compareCNVSegmentsAgainstQCHandler(cls, input_fname_ls, ecotype_id2cnv_qc_call_data, function_handler, param_obj, \ deletion_cutoff=None, max_boundary_diff=10000, \ max_diff_perc=0.10, min_no_of_probes=5, count_embedded_segment_as_match=False): """ 2009-11-4 a general handler to compare CNV segments from input_fnam... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
def getCNVQCDataFromDB(cls, data_source_id=1, ecotype_id=None, cnv_type_id=1, \ min_QC_segment_size=200, min_no_of_probes=None): """ | def getCNVQCDataFromDB(cls, data_source_id=1, ecotype_id=None, cnv_type_id=None, \ min_QC_segment_size=None, min_no_of_probes=None, min_reciprocal_overlap=0.6): """ 2010-1-26 replace the list structure of cnv_qc_call_data in ecotype_id2cnv_qc_call_data with binary_tree structure 2009-12-9 add no_of_probes_covered into ... | def getCNVQCDataFromDB(cls, data_source_id=1, ecotype_id=None, cnv_type_id=1, \ min_QC_segment_size=200, min_no_of_probes=None): """ 2009-11-4 get CNV QC data from database """ sys.stderr.write("Getting CNV QC data ... \n") import Stock_250kDB sql_string = "select a.ecotype_id, c.chromosome, c.start, c.stop, c.size_aff... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
sql_string = "select a.ecotype_id, c.chromosome, c.start, c.stop, c.size_affected, c.id from %s c,\ %s a where c.accession_id=a.id and a.data_source_id=%s and c.size_affected>=%s \ and c.cnv_type_id=%s"%\ (Stock_250kDB.CNVQCCalls.table.name, Stock_250kDB.CNVQCAccession.table.name, data_source_id,\ min_QC_segment_size, ... | sql_string = "select a.ecotype_id, c.chromosome, c.start, c.stop, c.size_affected, c.no_of_probes_covered, c.copy_number, c.id from %s c,\ %s a where c.accession_id=a.id and a.data_source_id=%s order by RAND()"%\ (Stock_250kDB.CNVQCCalls.table.name, Stock_250kDB.CNVQCAccession.table.name, data_source_id) if cnv_type_i... | def getCNVQCDataFromDB(cls, data_source_id=1, ecotype_id=None, cnv_type_id=1, \ min_QC_segment_size=200, min_no_of_probes=None): """ 2009-11-4 get CNV QC data from database """ sys.stderr.write("Getting CNV QC data ... \n") import Stock_250kDB sql_string = "select a.ecotype_id, c.chromosome, c.start, c.stop, c.size_aff... | 1440ce5092ec2e1bfdd9b1e187e40f109bf0000b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9645/1440ce5092ec2e1bfdd9b1e187e40f109bf0000b/misc.py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.