_id
stringlengths
2
7
title
stringlengths
1
88
partition
stringclasses
3 values
text
stringlengths
31
13.1k
language
stringclasses
1 value
meta_information
dict
q14400
setup_decoder
train
def setup_decoder(codec, dparams): """Wraps openjp2 library function opj_setup_decoder. Setup the decoder with decompression parameters. Parameters ---------- codec: CODEC_TYPE Codec initialized by create_compress function. dparams: DecompressionParametersType Decompression parameters. Raises ------ RuntimeError If the
python
{ "resource": "" }
q14401
setup_encoder
train
def setup_encoder(codec, cparams, image): """Wraps openjp2 library function opj_setup_encoder. Setup the encoder parameters using the current image and using user parameters. Parameters ---------- codec : CODEC_TYPE codec initialized by create_compress function cparams : CompressionParametersType compression parameters image : ImageType input-filled image Raises ------ RuntimeError If the OpenJPEG library routine opj_setup_encoder fails. """ ARGTYPES = [CODEC_TYPE,
python
{ "resource": "" }
q14402
start_compress
train
def start_compress(codec, image, stream): """Wraps openjp2 library function opj_start_compress. Start to compress the current image. Parameters ---------- codec : CODEC_TYPE Compressor handle. image : pointer to ImageType Input filled image. stream : STREAM_TYPE_P
python
{ "resource": "" }
q14403
stream_create_default_file_stream
train
def stream_create_default_file_stream(fname, isa_read_stream): """Wraps openjp2 library function opj_stream_create_default_vile_stream. Sets the stream to be a file stream. This function is only valid for the 2.1 version of the openjp2 library. Parameters ---------- fname : str Specifies a file. isa_read_stream: bool True (read) or False (write) Returns ------- stream : stream_t An OpenJPEG file stream.
python
{ "resource": "" }
q14404
stream_destroy
train
def stream_destroy(stream): """Wraps openjp2 library function opj_stream_destroy. Destroys the stream created by create_stream. Parameters
python
{ "resource": "" }
q14405
write_tile
train
def write_tile(codec, tile_index, data, data_size, stream): """Wraps openjp2 library function opj_write_tile. Write a tile into an image. Parameters ---------- codec : CODEC_TYPE The jpeg2000 codec tile_index : int The index of the tile to write, zero-indexing assumed data : array Image data arranged in usual C-order data_size : int Size of a tile in bytes stream : STREAM_TYPE_P The stream to write data to Raises ------ RuntimeError If the OpenJPEG library routine opj_write_tile fails. """ OPENJP2.opj_write_tile.argtypes = [CODEC_TYPE, ctypes.c_uint32, ctypes.POINTER(ctypes.c_uint8),
python
{ "resource": "" }
q14406
removeZeroLenPadding
train
def removeZeroLenPadding(str, blocksize=AES_blocksize): 'Remove Padding with zeroes + last byte equal to the number of padding bytes' try: pad_len = ord(str[-1]) # last byte contains number of padding bytes except TypeError:
python
{ "resource": "" }
q14407
appendNullPadding
train
def appendNullPadding(str, blocksize=AES_blocksize): 'Pad with null bytes' pad_len =
python
{ "resource": "" }
q14408
removeNullPadding
train
def removeNullPadding(str, blocksize=AES_blocksize): 'Remove padding with null bytes' pad_len = 0 for char in str[::-1]: # str[::-1] reverses string if char ==
python
{ "resource": "" }
q14409
appendSpacePadding
train
def appendSpacePadding(str, blocksize=AES_blocksize): 'Pad with spaces' pad_len = paddingLength(len(str), blocksize)
python
{ "resource": "" }
q14410
removeSpacePadding
train
def removeSpacePadding(str, blocksize=AES_blocksize): 'Remove padding with spaces' pad_len = 0 for char in
python
{ "resource": "" }
q14411
_default_error_handler
train
def _default_error_handler(msg, _): """Default error handler callback for libopenjp2.""" msg = "OpenJPEG library error:
python
{ "resource": "" }
q14412
_default_warning_handler
train
def _default_warning_handler(library_msg, _): """Default warning handler callback.""" library_msg = library_msg.decode('utf-8').rstrip() msg = "OpenJPEG
python
{ "resource": "" }
q14413
Jp2k.parse
train
def parse(self): """Parses the JPEG 2000 file. Raises ------ IOError The file was not JPEG 2000. """ self.length = os.path.getsize(self.filename) with open(self.filename, 'rb') as fptr: # Make sure we have a JPEG2000 file. It could be either JP2 or # J2C. Check for J2C first, single box in that case. read_buffer = fptr.read(2) signature, = struct.unpack('>H', read_buffer) if signature == 0xff4f: self._codec_format = opj2.CODEC_J2K # That's it, we're done. The codestream object is only # produced upon explicit request. return self._codec_format = opj2.CODEC_JP2 # Should be JP2. # First 4 bytes should be 12, the length of the 'jP ' box. # 2nd 4 bytes should be the box ID ('jP '). # 3rd 4 bytes should be the box signature (13, 10, 135, 10). fptr.seek(0)
python
{ "resource": "" }
q14414
Jp2k._validate
train
def _validate(self): """Validate the JPEG 2000 outermost superbox. These checks must be done at a file level. """ # A JP2 file must contain certain boxes. The 2nd box must be a file # type box. if not isinstance(self.box[1], FileTypeBox): msg = "{filename} does not contain a valid File Type box." msg = msg.format(filename=self.filename) raise IOError(msg) # A jp2-branded file cannot contain an "any ICC profile ftyp = self.box[1] if ftyp.brand == 'jp2 ': jp2h = [box for box in self.box if box.box_id == 'jp2h'][0] colrs = [box for box in jp2h.box
python
{ "resource": "" }
q14415
Jp2k._set_cinema_params
train
def _set_cinema_params(self, cinema_mode, fps): """Populate compression parameters structure for cinema2K. Parameters ---------- params : ctypes struct Corresponds to compression parameters structure used by the library. cinema_mode : {'cinema2k', 'cinema4k} Use either Cinema2K or Cinema4K profile. fps : {24, 48} Frames per second. """ if re.match("1.5|2.0", version.openjpeg_version) is not None: msg = ("Writing Cinema2K or Cinema4K files is not supported with " "OpenJPEG library versions less than 2.1.0. The installed " "version of OpenJPEG is {version}.") msg = msg.format(version=version.openjpeg_version) raise IOError(msg) # Cinema modes imply MCT. self._cparams.tcp_mct = 1 if cinema_mode == 'cinema2k': if fps not in [24, 48]: msg = 'Cinema2K frame rate must be either 24 or 48.'
python
{ "resource": "" }
q14416
Jp2k._write_openjpeg
train
def _write_openjpeg(self, img_array, verbose=False): """ Write JPEG 2000 file using OpenJPEG 1.5 interface. """ if img_array.ndim == 2: # Force the image to be 3D. Just makes things easier later on. img_array = img_array.reshape(img_array.shape[0], img_array.shape[1], 1) self._populate_comptparms(img_array) with ExitStack() as stack: image = opj.image_create(self._comptparms, self._colorspace) stack.callback(opj.image_destroy, image) numrows, numcols, numlayers = img_array.shape # set image offset and reference grid image.contents.x0 = self._cparams.image_offset_x0 image.contents.y0 = self._cparams.image_offset_y0 image.contents.x1 = (image.contents.x0 + (numcols - 1) * self._cparams.subsampling_dx + 1) image.contents.y1 = (image.contents.y0 + (numrows - 1) * self._cparams.subsampling_dy + 1) # Stage the image data to the openjpeg data structure. for k in range(0, numlayers): layer = np.ascontiguousarray(img_array[:, :, k], dtype=np.int32) dest = image.contents.comps[k].data src = layer.ctypes.data ctypes.memmove(dest, src, layer.nbytes) cinfo = opj.create_compress(self._cparams.codec_fmt) stack.callback(opj.destroy_compress, cinfo) # Setup the info, warning, and error handlers. # Always use the warning and error handler. Use of an info #
python
{ "resource": "" }
q14417
Jp2k._validate_j2k_colorspace
train
def _validate_j2k_colorspace(self, cparams, colorspace): """ Cannot specify a colorspace with J2K. """ if cparams.codec_fmt == opj2.CODEC_J2K and colorspace is not None:
python
{ "resource": "" }
q14418
Jp2k._validate_codeblock_size
train
def _validate_codeblock_size(self, cparams): """ Code block dimensions must satisfy certain restrictions. They must both be a power of 2 and the total area defined by the width and height cannot be either too great or too small for the codec. """ if cparams.cblockw_init != 0 and cparams.cblockh_init != 0: # These fields ARE zero if uninitialized. width = cparams.cblockw_init height = cparams.cblockh_init if height * width > 4096 or height < 4 or width < 4: msg = ("The code block area is specified as " "{height} x {width} = {area} square pixels. " "Code block area cannot exceed 4096 square pixels. " "Code block height and width dimensions must be larger " "than 4 pixels.")
python
{ "resource": "" }
q14419
Jp2k._validate_precinct_size
train
def _validate_precinct_size(self, cparams): """ Precinct dimensions must satisfy certain restrictions if specified. They must both be a power of 2 and must both be at least twice the size of their codeblock size counterparts. """ code_block_specified = False if cparams.cblockw_init != 0 and cparams.cblockh_init != 0: code_block_specified = True if cparams.res_spec != 0: # precinct size was not specified if this field is zero. for j in range(cparams.res_spec): prch = cparams.prch_init[j] prcw = cparams.prcw_init[j] if j == 0 and code_block_specified: height, width = cparams.cblockh_init, cparams.cblockw_init if prch < height * 2 or prcw < width * 2: msg = ("The highest resolution precinct size " "({prch} x {prcw}) must be at least twice that "
python
{ "resource": "" }
q14420
Jp2k._validate_image_rank
train
def _validate_image_rank(self, img_array): """ Images must be either 2D or 3D. """
python
{ "resource": "" }
q14421
Jp2k._validate_image_datatype
train
def _validate_image_datatype(self, img_array): """ Only uint8 and uint16 images are currently supported. """ if img_array.dtype != np.uint8 and img_array.dtype != np.uint16:
python
{ "resource": "" }
q14422
Jp2k._validate_compression_params
train
def _validate_compression_params(self, img_array, cparams, colorspace): """Check that the compression parameters are valid. Parameters ---------- img_array : ndarray Image data to be written to file. cparams : CompressionParametersType(ctypes.Structure) Corresponds to cparameters_t type in openjp2 headers. """
python
{ "resource": "" }
q14423
Jp2k._determine_colorspace
train
def _determine_colorspace(self, colorspace=None, **kwargs): """Determine the colorspace from the supplied inputs. Parameters ---------- colorspace : str, optional Either 'rgb' or 'gray'. """ if colorspace is None: # Must infer the colorspace from the image dimensions. if len(self.shape) < 3: # A single channel image is grayscale. self._colorspace = opj2.CLRSPC_GRAY elif self.shape[2] == 1 or self.shape[2] == 2: # A single channel image or an image with two channels is going # to be greyscale. self._colorspace = opj2.CLRSPC_GRAY else: # Anything else must be RGB, right? self._colorspace = opj2.CLRSPC_SRGB else: if colorspace.lower() not in ('rgb', 'grey', 'gray'): msg = 'Invalid colorspace "{0}".'.format(colorspace)
python
{ "resource": "" }
q14424
Jp2k._write_openjp2
train
def _write_openjp2(self, img_array, verbose=False): """ Write JPEG 2000 file using OpenJPEG 2.x interface. """ if img_array.ndim == 2: # Force the image to be 3D. Just makes things easier later on. numrows, numcols = img_array.shape img_array = img_array.reshape(numrows, numcols, 1) self._populate_comptparms(img_array) with ExitStack() as stack: image = opj2.image_create(self._comptparms, self._colorspace) stack.callback(opj2.image_destroy, image) self._populate_image_struct(image, img_array)
python
{ "resource": "" }
q14425
Jp2k.append
train
def append(self, box): """Append a JP2 box to the file in-place. Parameters ---------- box : Jp2Box Instance of a JP2 box. Only UUID and XML boxes can currently be appended. """ if self._codec_format == opj2.CODEC_J2K: msg = "Only JP2 files can currently have boxes appended to them." raise IOError(msg) if not ((box.box_id == 'xml ') or (box.box_id == 'uuid' and box.uuid == UUID('be7acfcb-97a9-42e8-9c71-999491e3afac'))): msg = ("Only XML boxes and XMP UUID boxes can currently be " "appended.") raise IOError(msg) # Check the last box. If the length field is zero, then rewrite # the length field to reflect the true length of the box. with open(self.filename, 'rb') as ifile: offset = self.box[-1].offset ifile.seek(offset) read_buffer = ifile.read(4) box_length, = struct.unpack('>I',
python
{ "resource": "" }
q14426
Jp2k._write_wrapped_codestream
train
def _write_wrapped_codestream(self, ofile, box): """Write wrapped codestream.""" # Codestreams require a bit more care. # Am I a raw codestream? if len(self.box) == 0: # Yes, just write the codestream box header plus all # of myself out to file. ofile.write(struct.pack('>I', self.length + 8)) ofile.write(b'jp2c') with open(self.filename, 'rb') as ifile: ofile.write(ifile.read()) return # OK, I'm a jp2/jpx file. Need to find out where the raw codestream # actually starts. offset = box.offset if offset == -1: if self.box[1].brand == 'jpx ': msg = ("The codestream box must have its offset and length " "attributes fully specified if the file type brand is " "JPX.") raise IOError(msg) # Find the first codestream in the file. jp2c = [_box for _box in self.box if _box.box_id == 'jp2c'] offset = jp2c[0].offset # Ready to write the codestream. with open(self.filename, 'rb') as ifile:
python
{ "resource": "" }
q14427
Jp2k._get_default_jp2_boxes
train
def _get_default_jp2_boxes(self): """Create a default set of JP2 boxes.""" # Try to create a reasonable default. boxes = [JPEG2000SignatureBox(), FileTypeBox(), JP2HeaderBox(), ContiguousCodestreamBox()] height = self.codestream.segment[1].ysiz width = self.codestream.segment[1].xsiz num_components = len(self.codestream.segment[1].xrsiz) if num_components < 3: colorspace = core.GREYSCALE else: if len(self.box) == 0: # Best guess is SRGB colorspace = core.SRGB else: # Take whatever the first jp2 header / color specification # says.
python
{ "resource": "" }
q14428
Jp2k._remove_ellipsis
train
def _remove_ellipsis(self, index, numrows, numcols, numbands): """ resolve the first ellipsis in the index so that it references the image Parameters ---------- index : tuple tuple of index arguments, presumably one of them is the Ellipsis numrows, numcols, numbands : int image dimensions Returns ------- tuple Same as index, except that the first Ellipsis is replaced with a proper slice whose start and stop members are not None """ # Remove the first ellipsis we find. rows = slice(0, numrows) cols = slice(0, numcols) bands = slice(0, numbands) if index[0] is Ellipsis: if len(index) == 2: # jp2k[..., other_slice] newindex = (rows, cols, index[1]) else:
python
{ "resource": "" }
q14429
Jp2k._subsampling_sanity_check
train
def _subsampling_sanity_check(self): """Check for differing subsample factors. """ dxs = np.array(self.codestream.segment[1].xrsiz) dys = np.array(self.codestream.segment[1].yrsiz) if np.any(dxs - dxs[0]) or np.any(dys - dys[0]): msg = ("The read_bands method should be used with the subsampling
python
{ "resource": "" }
q14430
Jp2k._read_openjpeg
train
def _read_openjpeg(self, rlevel=0, verbose=False, area=None): """Read a JPEG 2000 image using libopenjpeg. Parameters ---------- rlevel : int, optional Factor by which to rlevel output resolution. Use -1 to get the lowest resolution thumbnail. verbose : bool, optional Print informational messages produced by the OpenJPEG library. area : tuple, optional Specifies decoding image area, (first_row, first_col, last_row, last_col) Returns ------- ndarray The image data. Raises ------ RuntimeError If the image has differing subsample factors. """ self._subsampling_sanity_check() self._populate_dparams(rlevel) with ExitStack() as stack: try: self._dparams.decod_format = self._codec_format dinfo = opj.create_decompress(self._dparams.decod_format) event_mgr = opj.EventMgrType() handler = ctypes.cast(_INFO_CALLBACK, ctypes.c_void_p) event_mgr.info_handler = handler if self.verbose else None event_mgr.warning_handler = ctypes.cast(_WARNING_CALLBACK, ctypes.c_void_p) event_mgr.error_handler = ctypes.cast(_ERROR_CALLBACK, ctypes.c_void_p) opj.set_event_mgr(dinfo, ctypes.byref(event_mgr)) opj.setup_decoder(dinfo, self._dparams) with open(self.filename, 'rb') as fptr:
python
{ "resource": "" }
q14431
Jp2k._populate_dparams
train
def _populate_dparams(self, rlevel, tile=None, area=None): """Populate decompression structure with appropriate input parameters. Parameters ---------- rlevel : int Factor by which to rlevel output resolution. area : tuple Specifies decoding image area, (first_row, first_col, last_row, last_col) tile : int Number of tile to decode. """ if opj2.OPENJP2 is not None: dparam = opj2.set_default_decoder_parameters() else: dparam = opj.DecompressionParametersType() opj.set_default_decoder_parameters(ctypes.byref(dparam)) infile = self.filename.encode() nelts = opj2.PATH_LEN - len(infile) infile += b'0' * nelts dparam.infile = infile # Return raw codestream components instead of "interpolating" the # colormap? dparam.flags |= 1 if self.ignore_pclr_cmap_cdef else 0 dparam.decod_format = self._codec_format dparam.cp_layer = self._layer # Must check the specified rlevel against the maximum. if rlevel != 0: # Must check the specified rlevel against the maximum. max_rlevel = self.codestream.segment[2].num_res if rlevel == -1: # -1 is shorthand for the largest rlevel rlevel = max_rlevel elif rlevel < -1 or rlevel > max_rlevel: msg = ("rlevel must be in the range [-1, {max_rlevel}] " "for this image.") msg = msg.format(max_rlevel=max_rlevel) raise IOError(msg) dparam.cp_reduce = rlevel if area is not None: if area[0] < 0 or area[1] < 0
python
{ "resource": "" }
q14432
Jp2k._extract_image
train
def _extract_image(self, raw_image): """ Extract unequally-sized image bands. Parameters ---------- raw_image : reference to openjpeg ImageType instance The image structure initialized with image characteristics. Returns ------- list or ndarray If the JPEG 2000 image has unequally-sized components, they are extracted into a list, otherwise a numpy array. """ ncomps = raw_image.contents.numcomps # Make a pass thru the image, see if any of the band datatypes or # dimensions differ. dtypes, nrows, ncols = [], [], [] for k in range(raw_image.contents.numcomps): component = raw_image.contents.comps[k] dtypes.append(self._component2dtype(component)) nrows.append(component.h) ncols.append(component.w) is_cube = all(r == nrows[0] and c == ncols[0] and d == dtypes[0] for r, c, d in zip(nrows, ncols, dtypes)) if is_cube: image = np.zeros((nrows[0], ncols[0], ncomps), dtypes[0]) else: image = [] for k in range(raw_image.contents.numcomps): component = raw_image.contents.comps[k]
python
{ "resource": "" }
q14433
Jp2k._component2dtype
train
def _component2dtype(self, component): """Determin the appropriate numpy datatype for an OpenJPEG component. Parameters ---------- component : ctypes pointer to ImageCompType (image_comp_t) single image component structure. Returns ------- builtins.type numpy datatype to be used to construct an image array """ if component.prec > 16: msg = "Unhandled precision: {0} bits.".format(component.prec)
python
{ "resource": "" }
q14434
Jp2k.get_codestream
train
def get_codestream(self, header_only=True): """Retrieve codestream. Parameters ---------- header_only : bool, optional If True, only marker segments in the main header are parsed. Supplying False may impose a large performance penalty. Returns ------- Codestream Object describing the codestream syntax. Examples -------- >>> import glymur >>> jfile = glymur.data.nemo() >>> jp2 = glymur.Jp2k(jfile) >>> codestream = jp2.get_codestream() >>> print(codestream.segment[1]) SIZ marker segment @ (3233, 47) Profile: no profile Reference Grid Height, Width: (1456 x 2592) Vertical, Horizontal Reference Grid Offset: (0 x 0) Reference Tile Height, Width: (1456 x 2592) Vertical, Horizontal Reference Tile Offset:
python
{ "resource": "" }
q14435
Jp2k._populate_image_struct
train
def _populate_image_struct(self, image, imgdata): """Populates image struct needed for compression. Parameters ---------- image : ImageType(ctypes.Structure) Corresponds to image_t type in openjp2 headers. img_array : ndarray Image data to be written to file. """ numrows, numcols, num_comps = imgdata.shape for k in range(num_comps): self._validate_nonzero_image_size(numrows, numcols, k) # set image offset and reference grid image.contents.x0 = self._cparams.image_offset_x0 image.contents.y0 = self._cparams.image_offset_y0 image.contents.x1 = (image.contents.x0 + (numcols - 1) * self._cparams.subsampling_dx + 1) image.contents.y1 = (image.contents.y0 + (numrows - 1) * self._cparams.subsampling_dy + 1) # Stage the image data to the openjpeg data structure. for k in range(0, num_comps):
python
{ "resource": "" }
q14436
Jp2k._populate_comptparms
train
def _populate_comptparms(self, img_array): """Instantiate and populate comptparms structure. This structure defines the image components. Parameters ---------- img_array : ndarray Image data to be written to file. """ # Only two precisions are possible. if img_array.dtype == np.uint8: comp_prec = 8 else: comp_prec = 16 numrows, numcols, num_comps = img_array.shape if version.openjpeg_version_tuple[0] == 1: comptparms = (opj.ImageComptParmType * num_comps)() else: comptparms = (opj2.ImageComptParmType * num_comps)() for j in range(num_comps):
python
{ "resource": "" }
q14437
Jp2k._validate_nonzero_image_size
train
def _validate_nonzero_image_size(self, nrows, ncols, component_index): """The image cannot have area of zero. """ if nrows == 0 or ncols == 0: # Letting this situation continue would segfault openjpeg. msg
python
{ "resource": "" }
q14438
Jp2k._validate_jp2_box_sequence
train
def _validate_jp2_box_sequence(self, boxes): """Run through series of tests for JP2 box legality. This is non-exhaustive. """ JP2_IDS = ['colr', 'cdef', 'cmap', 'jp2c', 'ftyp', 'ihdr', 'jp2h', 'jP ', 'pclr', 'res ', 'resc', 'resd', 'xml ', 'ulst', 'uinf', 'url ', 'uuid'] self._validate_signature_compatibility(boxes) self._validate_jp2h(boxes) self._validate_jp2c(boxes) if boxes[1].brand == 'jpx ': self._validate_jpx_box_sequence(boxes) else: # Validate the JP2 box IDs.
python
{ "resource": "" }
q14439
Jp2k._validate_jp2_colr
train
def _validate_jp2_colr(self, boxes): """ Validate JP2 requirements on colour specification boxes. """ lst = [box for box in boxes if box.box_id == 'jp2h'] jp2h = lst[0] for colr in [box for box in jp2h.box if box.box_id == 'colr']:
python
{ "resource": "" }
q14440
Jp2k._validate_jpx_box_sequence
train
def _validate_jpx_box_sequence(self, boxes): """Run through series of tests for JPX box legality.""" self._validate_label(boxes) self._validate_jpx_compatibility(boxes,
python
{ "resource": "" }
q14441
Jp2k._validate_signature_compatibility
train
def _validate_signature_compatibility(self, boxes): """Validate the file signature and compatibility status.""" # Check for a bad sequence of boxes. # 1st two boxes must be 'jP ' and 'ftyp' if boxes[0].box_id != 'jP ' or boxes[1].box_id != 'ftyp': msg = ("The first box must be the signature box and the second " "must be the file type box.") raise IOError(msg)
python
{ "resource": "" }
q14442
Jp2k._validate_jp2c
train
def _validate_jp2c(self, boxes): """Validate the codestream box in relation to other boxes.""" # jp2c must be preceeded by jp2h jp2h_lst = [idx for (idx, box) in enumerate(boxes) if box.box_id == 'jp2h'] jp2h_idx = jp2h_lst[0] jp2c_lst = [idx for (idx, box) in enumerate(boxes) if box.box_id == 'jp2c'] if len(jp2c_lst) == 0: msg = ("A codestream box must be defined in the outermost "
python
{ "resource": "" }
q14443
Jp2k._validate_jp2h
train
def _validate_jp2h(self, boxes): """Validate the JP2 Header box.""" self._check_jp2h_child_boxes(boxes, 'top-level') jp2h_lst = [box for box in boxes if box.box_id == 'jp2h'] jp2h = jp2h_lst[0] # 1st jp2 header box cannot be empty. if len(jp2h.box) == 0: msg = "The JP2 header superbox cannot be empty." raise IOError(msg) # 1st jp2 header box must be ihdr if jp2h.box[0].box_id != 'ihdr': msg = ("The first box in the jp2 header box must be the image " "header box.")
python
{ "resource": "" }
q14444
Jp2k._validate_channel_definition
train
def _validate_channel_definition(self, jp2h, colr): """Validate the channel definition box.""" cdef_lst = [j for (j, box) in enumerate(jp2h.box) if box.box_id == 'cdef'] if len(cdef_lst) > 1: msg = ("Only one channel definition box is allowed in the " "JP2 header.") raise IOError(msg) elif len(cdef_lst) == 1: cdef = jp2h.box[cdef_lst[0]]
python
{ "resource": "" }
q14445
Jp2k._check_jp2h_child_boxes
train
def _check_jp2h_child_boxes(self, boxes, parent_box_name): """Certain boxes can only reside in the JP2 header.""" JP2H_CHILDREN = set(['bpcc', 'cdef', 'cmap', 'ihdr', 'pclr']) box_ids = set([box.box_id for box in boxes]) intersection = box_ids.intersection(JP2H_CHILDREN) if len(intersection) > 0 and parent_box_name not in ['jp2h', 'jpch']: msg = "A {0} box can
python
{ "resource": "" }
q14446
Jp2k._collect_box_count
train
def _collect_box_count(self, boxes): """Count the occurences of each box type.""" count = Counter([box.box_id for box in boxes]) # Add the counts in the superboxes. for box in boxes:
python
{ "resource": "" }
q14447
Jp2k._validate_singletons
train
def _validate_singletons(self, boxes): """Several boxes can only occur once.""" count = self._collect_box_count(boxes) # Which boxes occur more than once? multiples = [box_id for
python
{ "resource": "" }
q14448
Jp2k._validate_jpx_compatibility
train
def _validate_jpx_compatibility(self, boxes, compatibility_list): """ If there is a JPX box then the compatibility list must also contain 'jpx '. """ JPX_IDS = ['asoc', 'nlst'] jpx_cl = set(compatibility_list) for box in boxes: if box.box_id in JPX_IDS: if len(set(['jpx ', 'jpxb']).intersection(jpx_cl)) == 0: msg = ("A JPX box requires that either 'jpx ' or 'jpxb' "
python
{ "resource": "" }
q14449
Jp2k._validate_label
train
def _validate_label(self, boxes): """ Label boxes can only be inside association, codestream headers, or compositing layer header boxes. """ for box in boxes: if box.box_id != 'asoc': if hasattr(box, 'box'): for boxi in box.box: if boxi.box_id == 'lbl ':
python
{ "resource": "" }
q14450
KeenApi.fulfill
train
def fulfill(self, method, *args, **kwargs): """ Fulfill an
python
{ "resource": "" }
q14451
KeenApi._order_by_is_valid_or_none
train
def _order_by_is_valid_or_none(self, params): """ Validates that a given order_by has proper syntax. :param params: Query params. :return: Returns True if either no order_by is present, or if the order_by is well-formed. """ if not "order_by" in params or not params["order_by"]: return True def _order_by_dict_is_not_well_formed(d): if not isinstance(d, dict): # Bad type. return True if "property_name" in d and d["property_name"]: if "direction" in d and not direction.is_valid_direction(d["direction"]): # Bad direction provided. return True for k in d: if k != "property_name" and k != "direction": # Unexpected key.
python
{ "resource": "" }
q14452
KeenApi._limit_is_valid_or_none
train
def _limit_is_valid_or_none(self, params): """ Validates that a given limit is not present or is well-formed. :param params: Query params. :return: Returns True if a limit is present or is well-formed. """ if not "limit" in params or not params["limit"]:
python
{ "resource": "" }
q14453
KeenApi.query
train
def query(self, analysis_type, params, all_keys=False): """ Performs a query using the Keen IO analysis API. A read key must be set first. """ if not self._order_by_is_valid_or_none(params): raise ValueError("order_by given is invalid or is missing required group_by.") if not self._limit_is_valid_or_none(params): raise ValueError("limit given is invalid or is missing required order_by.") url = "{0}/{1}/projects/{2}/queries/{3}".format(self.base_url, self.api_version,
python
{ "resource": "" }
q14454
KeenApi.delete_events
train
def delete_events(self, event_collection, params): """ Deletes events via the Keen IO API. A master key must be set first. :param event_collection: string, the event collection from which event are being deleted """ url = "{0}/{1}/projects/{2}/events/{3}".format(self.base_url, self.api_version, self.project_id,
python
{ "resource": "" }
q14455
KeenApi.get_collection
train
def get_collection(self, event_collection): """ Extracts info about a collection using the Keen IO API. A master key must be set first. :param event_collection: the name of the collection to retrieve info for """ url = "{0}/{1}/projects/{2}/events/{3}".format(self.base_url, self.api_version,
python
{ "resource": "" }
q14456
KeenApi._update_access_key_pair
train
def _update_access_key_pair(self, access_key_id, key, val): """ Helper for updating access keys in a DRY fashion. """ # Get current state via HTTPS. current_access_key = self.get_access_key(access_key_id) # Copy and only change the single parameter.
python
{ "resource": "" }
q14457
KeenApi.add_access_key_permissions
train
def add_access_key_permissions(self, access_key_id, permissions): """ Adds to the existing list of permissions on this key with the contents of this list. Will not remove any existing permissions or modify the remainder of the key. :param access_key_id: the 'key' value of the access key to add permissions to :param permissions: the new permissions to add to the existing list of permissions """ # Get current state via HTTPS. current_access_key = self.get_access_key(access_key_id) # Copy and only change the single parameter. payload_dict = KeenApi._build_access_key_dict(current_access_key) # Turn
python
{ "resource": "" }
q14458
KeenApi.remove_access_key_permissions
train
def remove_access_key_permissions(self, access_key_id, permissions): """ Removes a list of permissions from the existing list of permissions. Will not remove all existing permissions unless all such permissions are included in this list. Not to be confused with key revocation. See also: revoke_access_key() :param access_key_id: the 'key' value of the access key to remove some permissions from :param permissions: the permissions you wish to remove from this access key """ # Get current state via HTTPS. current_access_key = self.get_access_key(access_key_id) # Copy and only change the single parameter. payload_dict = KeenApi._build_access_key_dict(current_access_key)
python
{ "resource": "" }
q14459
KeenApi._error_handling
train
def _error_handling(self, res): """ Helper function to do the error handling :params res: the response from a request """ # making the error handling generic so if an status_code starting with 2 doesn't exist, we raise the
python
{ "resource": "" }
q14460
KeenApi._get_response_json
train
def _get_response_json(self, res): """ Helper function to extract the JSON body out of a response OR throw an exception. :param res: the response from a request :return: the JSON body OR throws an exception """ try: error = res.json() except ValueError: error = {
python
{ "resource": "" }
q14461
CachedDatasetsInterface.all
train
def all(self): """ Fetch all Cached Datasets for a Project. Read key must be set. """ return self._get_json(HTTPMethods.GET,
python
{ "resource": "" }
q14462
CachedDatasetsInterface.get
train
def get(self, dataset_name): """ Fetch a single Cached Dataset for a Project. Read key must be set. :param dataset_name: Name of Cached Dataset (not `display_name`) """
python
{ "resource": "" }
q14463
CachedDatasetsInterface.create
train
def create(self, dataset_name, query, index_by, display_name): """ Create a Cached Dataset for a Project. Master key must be set. """ url = "{0}/{1}".format(self._cached_datasets_url, dataset_name) payload = { "query": query,
python
{ "resource": "" }
q14464
CachedDatasetsInterface.results
train
def results(self, dataset_name, index_by, timeframe): """ Retrieve results from a Cached Dataset. Read key must be set. """ url = "{0}/{1}/results".format(self._cached_datasets_url, dataset_name) index_by = index_by if isinstance(index_by, str) else json.dumps(index_by) timeframe = timeframe if isinstance(timeframe, str) else json.dumps(timeframe) query_params = {
python
{ "resource": "" }
q14465
CachedDatasetsInterface.delete
train
def delete(self, dataset_name): """ Delete a Cached Dataset. Master Key must be set. """
python
{ "resource": "" }
q14466
switch.match
train
def match(self, *args): """Whether or not to enter a given case statement""" self.fall = self.fall or not args
python
{ "resource": "" }
q14467
glymurrc_fname
train
def glymurrc_fname(): """Return the path to the configuration file. Search order: 1) current working directory 2) environ var XDG_CONFIG_HOME 3) $HOME/.config/glymur/glymurrc """ # Current directory. fname = os.path.join(os.getcwd(), 'glymurrc') if os.path.exists(fname): return fname confdir = get_configdir()
python
{ "resource": "" }
q14468
load_library_handle
train
def load_library_handle(libname, path): """Load the library, return the ctypes handle.""" if path is None or path in ['None', 'none']: # Either could not find a library via ctypes or # user-configuration-file, or we could not find it in any of the # default locations, or possibly the user intentionally does not want # one of the libraries to load. return None try: if os.name == "nt": opj_lib = ctypes.windll.LoadLibrary(path) else:
python
{ "resource": "" }
q14469
read_config_file
train
def read_config_file(libname): """ Extract library locations from a configuration file. Parameters ---------- libname : str One of either 'openjp2' or 'openjpeg' Returns ------- path : None or str None if no location is specified, otherwise a path to the
python
{ "resource": "" }
q14470
glymur_config
train
def glymur_config(): """ Try to ascertain locations of openjp2, openjpeg libraries. Returns ------- tuple tuple of library handles """ handles = (load_openjpeg_library(x) for x in ['openjp2', 'openjpeg']) handles = tuple(handles)
python
{ "resource": "" }
q14471
get_configdir
train
def get_configdir(): """Return string representing the configuration directory. Default is $HOME/.config/glymur. You can override this with the XDG_CONFIG_HOME environment variable. """ if 'XDG_CONFIG_HOME' in os.environ: return os.path.join(os.environ['XDG_CONFIG_HOME'], 'glymur') if 'HOME' in os.environ and os.name != 'nt': # HOME is set by WinPython to something unusual, so we don't
python
{ "resource": "" }
q14472
set_option
train
def set_option(key, value): """Set the value of the specified option. Available options: parse.full_codestream print.xml print.codestream print.short Parameters ---------- key : str Name of a single option. value : New value of option. Option Descriptions ------------------- parse.full_codestream : bool When False, only the codestream header is parsed for metadata. This can results in faster JP2/JPX parsing. When True, the entire codestream is parsed. [default: False] print.codestream : bool When False, the codestream segments are not printed. Otherwise the segments are printed depending on the value of the parse.full_codestream option. [default: True]
python
{ "resource": "" }
q14473
reset_option
train
def reset_option(key): """ Reset one or more options to their default value. Pass "all" as argument to reset all options. Available options: parse.full_codestream print.xml print.codestream
python
{ "resource": "" }
q14474
set_printoptions
train
def set_printoptions(**kwargs): """Set printing options. These options determine the way JPEG 2000 boxes are displayed. Parameters ---------- short : bool, optional When True, only the box ID, offset, and length are displayed. Useful for displaying only the basic structure or skeleton of a JPEG 2000 file. xml : bool, optional When False, printing of the XML contents of any XML boxes or UUID XMP boxes is suppressed. codestream : bool, optional When False, the codestream segments are not printed. Otherwise the segments are printed depending on how set_parseoptions has been used. See also -------- get_printoptions Examples --------
python
{ "resource": "" }
q14475
get_printoptions
train
def get_printoptions(): """Return the current print options. Returns ------- dict Dictionary of current print options with keys - short : bool - xml : bool - codestream : bool For a full description of these options, see `set_printoptions`. See also -------- set_printoptions
python
{ "resource": "" }
q14476
main
train
def main(): """ Entry point for console script jp2dump. """ kwargs = {'description': 'Print JPEG2000 metadata.', 'formatter_class': argparse.ArgumentDefaultsHelpFormatter} parser = argparse.ArgumentParser(**kwargs) parser.add_argument('-x', '--noxml', help='suppress XML', action='store_true') parser.add_argument('-s', '--short', help='only print box id, offset, and length', action='store_true') chelp = 'Level of codestream information. 0 suppresses all details, ' chelp += '1 prints the main header, 2 prints the full codestream.' parser.add_argument('-c', '--codestream', help=chelp, metavar='LEVEL', nargs=1, type=int, default=[1]) parser.add_argument('filename') args = parser.parse_args() if args.noxml: set_option('print.xml', False) if args.short: set_option('print.short', True) codestream_level = args.codestream[0] if codestream_level not in [0, 1, 2]: raise ValueError("Invalid level of codestream information specified.") if codestream_level == 0: set_option('print.codestream', False) elif codestream_level == 2: set_option('parse.full_codestream', True) filename = args.filename # JP2 metadata can be extensive, so don't print any warnings until we # are done with the metadata. with warnings.catch_warnings(record=True) as wctx:
python
{ "resource": "" }
q14477
CerberusClient._set_token
train
def _set_token(self): """Set the Cerberus token based on auth type""" try: self.token = os.environ['CERBERUS_TOKEN'] if self.verbose: print("Overriding Cerberus token with environment variable.", file=sys.stderr) logger.info("Overriding Cerberus token with environment
python
{ "resource": "" }
q14478
CerberusClient.get_role
train
def get_role(self, key): """Return id of named role.""" json_resp = self.get_roles() for item in json_resp: if key in item["name"]:
python
{ "resource": "" }
q14479
CerberusClient.get_categories
train
def get_categories(self): """ Return a list of categories that a safe deposit box can belong to""" sdb_resp = get_with_retry(self.cerberus_url + '/v1/category',
python
{ "resource": "" }
q14480
CerberusClient.create_sdb
train
def create_sdb(self, name, category_id, owner, description="", user_group_permissions=None, iam_principal_permissions=None): """Create a safe deposit box. You need to refresh your token before the iam role is granted permission to the new safe deposit box. Keyword arguments: name (string) -- name of the safe deposit box category_id (string) -- category id that determines where to store the sdb. (ex: shared, applications) owner (string) -- AD group that owns the safe deposit box description (string) -- Description of the safe deposit box user_group_permissions (list) -- list of dictionaries containing the key name and maybe role_id iam_principal_permissions (list) -- list of dictionaries containing the key name iam_principal_arn and role_id """ # Do some sanity checking if user_group_permissions is None: user_group_permissions = [] if iam_principal_permissions is None: iam_principal_permissions = [] if list != type(user_group_permissions): raise(TypeError('Expected list, but got ' + str(type(user_group_permissions)))) if list != type(iam_principal_permissions):
python
{ "resource": "" }
q14481
CerberusClient.delete_sdb
train
def delete_sdb(self, sdb_id): """ Delete a safe deposit box specified by id Keyword arguments: sdb_id -- this is the id of the safe deposit box, not the path.""" sdb_resp = delete_with_retry(self.cerberus_url + '/v2/safe-deposit-box/' +
python
{ "resource": "" }
q14482
CerberusClient.get_sdb_path
train
def get_sdb_path(self, sdb): """Return the path for a SDB""" sdb_id = self.get_sdb_id(sdb) sdb_resp = get_with_retry(
python
{ "resource": "" }
q14483
CerberusClient.get_sdb_keys
train
def get_sdb_keys(self, path): """Return the keys for a SDB, which are need for the full secure data path""" list_resp = get_with_retry(
python
{ "resource": "" }
q14484
CerberusClient.get_sdb_id
train
def get_sdb_id(self, sdb): """ Return the ID for the given safe deposit box. Keyword arguments: sdb -- This is the name of the safe deposit box, not the path"""
python
{ "resource": "" }
q14485
CerberusClient.get_sdb_id_by_path
train
def get_sdb_id_by_path(self, sdb_path): """ Given the path, return the ID for the given safe deposit box.""" json_resp = self.get_sdbs() # Deal with the supplied path possibly missing an ending slash
python
{ "resource": "" }
q14486
CerberusClient.get_sdb_by_id
train
def get_sdb_by_id(self, sdb_id): """ Return the details for the given safe deposit box id Keyword arguments: sdb_id -- this is the id of the safe deposit box, not the path. """
python
{ "resource": "" }
q14487
CerberusClient.get_sdb_secret_version_paths
train
def get_sdb_secret_version_paths(self, sdb_id): """ Get SDB secret version paths. This function takes the sdb_id """ sdb_resp = get_with_retry(str.join('', [self.cerberus_url, '/v1/sdb-secret-version-paths/', sdb_id]),
python
{ "resource": "" }
q14488
CerberusClient.list_sdbs
train
def list_sdbs(self): """ Return sdbs by Name """ sdb_raw = self.get_sdbs() sdbs = []
python
{ "resource": "" }
q14489
CerberusClient.update_sdb
train
def update_sdb(self, sdb_id, owner=None, description=None, user_group_permissions=None, iam_principal_permissions=None): """ Update a safe deposit box. Keyword arguments: owner (string) -- AD group that owns the safe deposit box description (string) -- Description of the safe deposit box user_group_permissions (list) -- list of dictionaries containing the key name and maybe role_id iam_principal_permissions (list) -- list of dictionaries containing the key name iam_principal_arn and role_id """ # Grab current data old_data = self.get_sdb_by_id(sdb_id) # Assemble information to update temp_data = {} keys = ('owner', 'description', 'iam_principal_permissions', 'user_group_permissions') for k in keys: if k in old_data: temp_data[k] = old_data[k] if owner is not None: temp_data["owner"] = owner if description is not None: temp_data["description"] = description
python
{ "resource": "" }
q14490
CerberusClient.delete_file
train
def delete_file(self, secure_data_path): """Delete a file at the given secure data path""" secret_resp = delete_with_retry(self.cerberus_url + '/v1/secure-file/'
python
{ "resource": "" }
q14491
CerberusClient.get_file_metadata
train
def get_file_metadata(self, secure_data_path, version=None): """Get just the metadata for a file, not the content""" if not version: version = "CURRENT" payload = {'versionId': str(version)} secret_resp = head_with_retry(str.join('', [self.cerberus_url, '/v1/secure-file/', secure_data_path]),
python
{ "resource": "" }
q14492
CerberusClient._parse_metadata_filename
train
def _parse_metadata_filename(self, metadata): """ Parse the header metadata to pull out the filename and then store it under the key
python
{ "resource": "" }
q14493
CerberusClient.get_file_versions
train
def get_file_versions(self, secure_data_path, limit=None, offset=None): """ Get versions of a particular file This is just a shim to get_secret_versions secure_data_path -- full path to the file in the safety deposit box limit -- Default(100), limits how many records to be returned from the api at once.
python
{ "resource": "" }
q14494
CerberusClient._get_all_file_versions
train
def _get_all_file_versions(self, secure_data_path, limit=None): """ Convenience function that returns a generator yielding the contents of all versions of a file and its version info secure_data_path -- full path to the file in the safety deposit box limit -- Default(100), limits how many records to be returned from the api at once.
python
{ "resource": "" }
q14495
CerberusClient.list_files
train
def list_files(self, secure_data_path, limit=None, offset=None): """Return the list of files in the path. May need to be paginated""" # Make sure that limit and offset are in range. # Set the normal defaults if not limit or limit <= 0: limit = 100 if not offset or offset < 0: offset = 0
python
{ "resource": "" }
q14496
CerberusClient.put_file
train
def put_file(self, secure_data_path, filehandle, content_type=None): """ Upload a file to a secure data path provided Keyword arguments: secure_data_path -- full path in the safety deposit box that contains the file key to store things under filehandle -- Pass an opened filehandle to the file you want to upload. Make sure that the file was opened in binary mode, otherwise the size calculations can be off for text files. content_type -- Optional. Set the Mime type of the file you're uploading. """ # Parse out the filename from the path filename = secure_data_path.rsplit('/', 1) if content_type:
python
{ "resource": "" }
q14497
CerberusClient.get_secret_versions
train
def get_secret_versions(self, secure_data_path, limit=None, offset=None): """ Get versions of a particular secret key secure_data_path -- full path to the key in the safety deposit box limit -- Default(100), limits how many records to be returned from the api at once. offset -- Default(0), used for pagination. Will request records from the given offset. """ # Make sure that limit and offset are in range. # Set the normal defaults if not limit or limit <= 0: limit =
python
{ "resource": "" }
q14498
CerberusClient.list_secrets
train
def list_secrets(self, secure_data_path): """Return json secrets based on the secure_data_path, this will list keys in a folder""" # Because of the addition of versionId and the way URLs are constructed, secure_data_path should # always end in a '/'. secure_data_path = self._add_slash(secure_data_path) secret_resp
python
{ "resource": "" }
q14499
AWSAuth._get_v4_signed_headers
train
def _get_v4_signed_headers(self): """Returns V4 signed get-caller-identity request headers""" if self.aws_session is None: boto_session = session.Session() creds = boto_session.get_credentials() else: creds = self.aws_session.get_credentials() if creds is None: raise CerberusClientException("Unable to locate AWS credentials") readonly_credentials = creds.get_frozen_credentials() # hardcode get-caller-identity request
python
{ "resource": "" }