id int32 0 252k | repo stringlengths 7 55 | path stringlengths 4 127 | func_name stringlengths 1 88 | original_string stringlengths 75 19.8k | language stringclasses 1
value | code stringlengths 51 19.8k | code_tokens list | docstring stringlengths 3 17.3k | docstring_tokens list | sha stringlengths 40 40 | url stringlengths 87 242 |
|---|---|---|---|---|---|---|---|---|---|---|---|
14,400 | quintusdias/glymur | glymur/lib/openjp2.py | setup_decoder | def setup_decoder(codec, dparams):
"""Wraps openjp2 library function opj_setup_decoder.
Setup the decoder with decompression parameters.
Parameters
----------
codec: CODEC_TYPE
Codec initialized by create_compress function.
dparams: DecompressionParametersType
Decompression parameters.
Raises
------
RuntimeError
If the OpenJPEG library routine opj_setup_decoder fails.
"""
ARGTYPES = [CODEC_TYPE, ctypes.POINTER(DecompressionParametersType)]
OPENJP2.opj_setup_decoder.argtypes = ARGTYPES
OPENJP2.opj_setup_decoder.restype = check_error
OPENJP2.opj_setup_decoder(codec, ctypes.byref(dparams)) | python | def setup_decoder(codec, dparams):
ARGTYPES = [CODEC_TYPE, ctypes.POINTER(DecompressionParametersType)]
OPENJP2.opj_setup_decoder.argtypes = ARGTYPES
OPENJP2.opj_setup_decoder.restype = check_error
OPENJP2.opj_setup_decoder(codec, ctypes.byref(dparams)) | [
"def",
"setup_decoder",
"(",
"codec",
",",
"dparams",
")",
":",
"ARGTYPES",
"=",
"[",
"CODEC_TYPE",
",",
"ctypes",
".",
"POINTER",
"(",
"DecompressionParametersType",
")",
"]",
"OPENJP2",
".",
"opj_setup_decoder",
".",
"argtypes",
"=",
"ARGTYPES",
"OPENJP2",
"... | Wraps openjp2 library function opj_setup_decoder.
Setup the decoder with decompression parameters.
Parameters
----------
codec: CODEC_TYPE
Codec initialized by create_compress function.
dparams: DecompressionParametersType
Decompression parameters.
Raises
------
RuntimeError
If the OpenJPEG library routine opj_setup_decoder fails. | [
"Wraps",
"openjp2",
"library",
"function",
"opj_setup_decoder",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/lib/openjp2.py#L1168-L1189 |
14,401 | quintusdias/glymur | glymur/lib/openjp2.py | setup_encoder | def setup_encoder(codec, cparams, image):
"""Wraps openjp2 library function opj_setup_encoder.
Setup the encoder parameters using the current image and using user
parameters.
Parameters
----------
codec : CODEC_TYPE
codec initialized by create_compress function
cparams : CompressionParametersType
compression parameters
image : ImageType
input-filled image
Raises
------
RuntimeError
If the OpenJPEG library routine opj_setup_encoder fails.
"""
ARGTYPES = [CODEC_TYPE,
ctypes.POINTER(CompressionParametersType),
ctypes.POINTER(ImageType)]
OPENJP2.opj_setup_encoder.argtypes = ARGTYPES
OPENJP2.opj_setup_encoder.restype = check_error
OPENJP2.opj_setup_encoder(codec, ctypes.byref(cparams), image) | python | def setup_encoder(codec, cparams, image):
ARGTYPES = [CODEC_TYPE,
ctypes.POINTER(CompressionParametersType),
ctypes.POINTER(ImageType)]
OPENJP2.opj_setup_encoder.argtypes = ARGTYPES
OPENJP2.opj_setup_encoder.restype = check_error
OPENJP2.opj_setup_encoder(codec, ctypes.byref(cparams), image) | [
"def",
"setup_encoder",
"(",
"codec",
",",
"cparams",
",",
"image",
")",
":",
"ARGTYPES",
"=",
"[",
"CODEC_TYPE",
",",
"ctypes",
".",
"POINTER",
"(",
"CompressionParametersType",
")",
",",
"ctypes",
".",
"POINTER",
"(",
"ImageType",
")",
"]",
"OPENJP2",
".... | Wraps openjp2 library function opj_setup_encoder.
Setup the encoder parameters using the current image and using user
parameters.
Parameters
----------
codec : CODEC_TYPE
codec initialized by create_compress function
cparams : CompressionParametersType
compression parameters
image : ImageType
input-filled image
Raises
------
RuntimeError
If the OpenJPEG library routine opj_setup_encoder fails. | [
"Wraps",
"openjp2",
"library",
"function",
"opj_setup_encoder",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/lib/openjp2.py#L1192-L1217 |
14,402 | quintusdias/glymur | glymur/lib/openjp2.py | start_compress | def start_compress(codec, image, stream):
"""Wraps openjp2 library function opj_start_compress.
Start to compress the current image.
Parameters
----------
codec : CODEC_TYPE
Compressor handle.
image : pointer to ImageType
Input filled image.
stream : STREAM_TYPE_P
Input stream.
Raises
------
RuntimeError
If the OpenJPEG library routine opj_start_compress fails.
"""
OPENJP2.opj_start_compress.argtypes = [CODEC_TYPE,
ctypes.POINTER(ImageType),
STREAM_TYPE_P]
OPENJP2.opj_start_compress.restype = check_error
OPENJP2.opj_start_compress(codec, image, stream) | python | def start_compress(codec, image, stream):
OPENJP2.opj_start_compress.argtypes = [CODEC_TYPE,
ctypes.POINTER(ImageType),
STREAM_TYPE_P]
OPENJP2.opj_start_compress.restype = check_error
OPENJP2.opj_start_compress(codec, image, stream) | [
"def",
"start_compress",
"(",
"codec",
",",
"image",
",",
"stream",
")",
":",
"OPENJP2",
".",
"opj_start_compress",
".",
"argtypes",
"=",
"[",
"CODEC_TYPE",
",",
"ctypes",
".",
"POINTER",
"(",
"ImageType",
")",
",",
"STREAM_TYPE_P",
"]",
"OPENJP2",
".",
"o... | Wraps openjp2 library function opj_start_compress.
Start to compress the current image.
Parameters
----------
codec : CODEC_TYPE
Compressor handle.
image : pointer to ImageType
Input filled image.
stream : STREAM_TYPE_P
Input stream.
Raises
------
RuntimeError
If the OpenJPEG library routine opj_start_compress fails. | [
"Wraps",
"openjp2",
"library",
"function",
"opj_start_compress",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/lib/openjp2.py#L1220-L1244 |
14,403 | quintusdias/glymur | glymur/lib/openjp2.py | stream_create_default_file_stream | def stream_create_default_file_stream(fname, isa_read_stream):
"""Wraps openjp2 library function opj_stream_create_default_vile_stream.
Sets the stream to be a file stream. This function is only valid for the
2.1 version of the openjp2 library.
Parameters
----------
fname : str
Specifies a file.
isa_read_stream: bool
True (read) or False (write)
Returns
-------
stream : stream_t
An OpenJPEG file stream.
"""
ARGTYPES = [ctypes.c_char_p, ctypes.c_int32]
OPENJP2.opj_stream_create_default_file_stream.argtypes = ARGTYPES
OPENJP2.opj_stream_create_default_file_stream.restype = STREAM_TYPE_P
read_stream = 1 if isa_read_stream else 0
file_argument = ctypes.c_char_p(fname.encode())
stream = OPENJP2.opj_stream_create_default_file_stream(file_argument,
read_stream)
return stream | python | def stream_create_default_file_stream(fname, isa_read_stream):
ARGTYPES = [ctypes.c_char_p, ctypes.c_int32]
OPENJP2.opj_stream_create_default_file_stream.argtypes = ARGTYPES
OPENJP2.opj_stream_create_default_file_stream.restype = STREAM_TYPE_P
read_stream = 1 if isa_read_stream else 0
file_argument = ctypes.c_char_p(fname.encode())
stream = OPENJP2.opj_stream_create_default_file_stream(file_argument,
read_stream)
return stream | [
"def",
"stream_create_default_file_stream",
"(",
"fname",
",",
"isa_read_stream",
")",
":",
"ARGTYPES",
"=",
"[",
"ctypes",
".",
"c_char_p",
",",
"ctypes",
".",
"c_int32",
"]",
"OPENJP2",
".",
"opj_stream_create_default_file_stream",
".",
"argtypes",
"=",
"ARGTYPES"... | Wraps openjp2 library function opj_stream_create_default_vile_stream.
Sets the stream to be a file stream. This function is only valid for the
2.1 version of the openjp2 library.
Parameters
----------
fname : str
Specifies a file.
isa_read_stream: bool
True (read) or False (write)
Returns
-------
stream : stream_t
An OpenJPEG file stream. | [
"Wraps",
"openjp2",
"library",
"function",
"opj_stream_create_default_vile_stream",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/lib/openjp2.py#L1247-L1272 |
14,404 | quintusdias/glymur | glymur/lib/openjp2.py | stream_destroy | def stream_destroy(stream):
"""Wraps openjp2 library function opj_stream_destroy.
Destroys the stream created by create_stream.
Parameters
----------
stream : STREAM_TYPE_P
The file stream.
"""
OPENJP2.opj_stream_destroy.argtypes = [STREAM_TYPE_P]
OPENJP2.opj_stream_destroy.restype = ctypes.c_void_p
OPENJP2.opj_stream_destroy(stream) | python | def stream_destroy(stream):
OPENJP2.opj_stream_destroy.argtypes = [STREAM_TYPE_P]
OPENJP2.opj_stream_destroy.restype = ctypes.c_void_p
OPENJP2.opj_stream_destroy(stream) | [
"def",
"stream_destroy",
"(",
"stream",
")",
":",
"OPENJP2",
".",
"opj_stream_destroy",
".",
"argtypes",
"=",
"[",
"STREAM_TYPE_P",
"]",
"OPENJP2",
".",
"opj_stream_destroy",
".",
"restype",
"=",
"ctypes",
".",
"c_void_p",
"OPENJP2",
".",
"opj_stream_destroy",
"... | Wraps openjp2 library function opj_stream_destroy.
Destroys the stream created by create_stream.
Parameters
----------
stream : STREAM_TYPE_P
The file stream. | [
"Wraps",
"openjp2",
"library",
"function",
"opj_stream_destroy",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/lib/openjp2.py#L1275-L1287 |
14,405 | quintusdias/glymur | glymur/lib/openjp2.py | write_tile | def write_tile(codec, tile_index, data, data_size, stream):
"""Wraps openjp2 library function opj_write_tile.
Write a tile into an image.
Parameters
----------
codec : CODEC_TYPE
The jpeg2000 codec
tile_index : int
The index of the tile to write, zero-indexing assumed
data : array
Image data arranged in usual C-order
data_size : int
Size of a tile in bytes
stream : STREAM_TYPE_P
The stream to write data to
Raises
------
RuntimeError
If the OpenJPEG library routine opj_write_tile fails.
"""
OPENJP2.opj_write_tile.argtypes = [CODEC_TYPE,
ctypes.c_uint32,
ctypes.POINTER(ctypes.c_uint8),
ctypes.c_uint32,
STREAM_TYPE_P]
OPENJP2.opj_write_tile.restype = check_error
datap = data.ctypes.data_as(ctypes.POINTER(ctypes.c_uint8))
OPENJP2.opj_write_tile(codec,
ctypes.c_uint32(int(tile_index)),
datap,
ctypes.c_uint32(int(data_size)),
stream) | python | def write_tile(codec, tile_index, data, data_size, stream):
OPENJP2.opj_write_tile.argtypes = [CODEC_TYPE,
ctypes.c_uint32,
ctypes.POINTER(ctypes.c_uint8),
ctypes.c_uint32,
STREAM_TYPE_P]
OPENJP2.opj_write_tile.restype = check_error
datap = data.ctypes.data_as(ctypes.POINTER(ctypes.c_uint8))
OPENJP2.opj_write_tile(codec,
ctypes.c_uint32(int(tile_index)),
datap,
ctypes.c_uint32(int(data_size)),
stream) | [
"def",
"write_tile",
"(",
"codec",
",",
"tile_index",
",",
"data",
",",
"data_size",
",",
"stream",
")",
":",
"OPENJP2",
".",
"opj_write_tile",
".",
"argtypes",
"=",
"[",
"CODEC_TYPE",
",",
"ctypes",
".",
"c_uint32",
",",
"ctypes",
".",
"POINTER",
"(",
"... | Wraps openjp2 library function opj_write_tile.
Write a tile into an image.
Parameters
----------
codec : CODEC_TYPE
The jpeg2000 codec
tile_index : int
The index of the tile to write, zero-indexing assumed
data : array
Image data arranged in usual C-order
data_size : int
Size of a tile in bytes
stream : STREAM_TYPE_P
The stream to write data to
Raises
------
RuntimeError
If the OpenJPEG library routine opj_write_tile fails. | [
"Wraps",
"openjp2",
"library",
"function",
"opj_write_tile",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/lib/openjp2.py#L1290-L1325 |
14,406 | keenlabs/KeenClient-Python | keen/Padding.py | removeZeroLenPadding | def removeZeroLenPadding(str, blocksize=AES_blocksize):
'Remove Padding with zeroes + last byte equal to the number of padding bytes'
try:
pad_len = ord(str[-1]) # last byte contains number of padding bytes
except TypeError:
pad_len = str[-1]
assert pad_len < blocksize, 'padding error'
assert pad_len < len(str), 'padding error'
return str[:-pad_len] | python | def removeZeroLenPadding(str, blocksize=AES_blocksize):
'Remove Padding with zeroes + last byte equal to the number of padding bytes'
try:
pad_len = ord(str[-1]) # last byte contains number of padding bytes
except TypeError:
pad_len = str[-1]
assert pad_len < blocksize, 'padding error'
assert pad_len < len(str), 'padding error'
return str[:-pad_len] | [
"def",
"removeZeroLenPadding",
"(",
"str",
",",
"blocksize",
"=",
"AES_blocksize",
")",
":",
"try",
":",
"pad_len",
"=",
"ord",
"(",
"str",
"[",
"-",
"1",
"]",
")",
"# last byte contains number of padding bytes",
"except",
"TypeError",
":",
"pad_len",
"=",
"st... | Remove Padding with zeroes + last byte equal to the number of padding bytes | [
"Remove",
"Padding",
"with",
"zeroes",
"+",
"last",
"byte",
"equal",
"to",
"the",
"number",
"of",
"padding",
"bytes"
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/Padding.py#L156-L165 |
14,407 | keenlabs/KeenClient-Python | keen/Padding.py | appendNullPadding | def appendNullPadding(str, blocksize=AES_blocksize):
'Pad with null bytes'
pad_len = paddingLength(len(str), blocksize)
padding = '\0'*pad_len
return str + padding | python | def appendNullPadding(str, blocksize=AES_blocksize):
'Pad with null bytes'
pad_len = paddingLength(len(str), blocksize)
padding = '\0'*pad_len
return str + padding | [
"def",
"appendNullPadding",
"(",
"str",
",",
"blocksize",
"=",
"AES_blocksize",
")",
":",
"pad_len",
"=",
"paddingLength",
"(",
"len",
"(",
"str",
")",
",",
"blocksize",
")",
"padding",
"=",
"'\\0'",
"*",
"pad_len",
"return",
"str",
"+",
"padding"
] | Pad with null bytes | [
"Pad",
"with",
"null",
"bytes"
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/Padding.py#L167-L173 |
14,408 | keenlabs/KeenClient-Python | keen/Padding.py | removeNullPadding | def removeNullPadding(str, blocksize=AES_blocksize):
'Remove padding with null bytes'
pad_len = 0
for char in str[::-1]: # str[::-1] reverses string
if char == '\0':
pad_len += 1
else:
break
str = str[:-pad_len]
return str | python | def removeNullPadding(str, blocksize=AES_blocksize):
'Remove padding with null bytes'
pad_len = 0
for char in str[::-1]: # str[::-1] reverses string
if char == '\0':
pad_len += 1
else:
break
str = str[:-pad_len]
return str | [
"def",
"removeNullPadding",
"(",
"str",
",",
"blocksize",
"=",
"AES_blocksize",
")",
":",
"pad_len",
"=",
"0",
"for",
"char",
"in",
"str",
"[",
":",
":",
"-",
"1",
"]",
":",
"# str[::-1] reverses string",
"if",
"char",
"==",
"'\\0'",
":",
"pad_len",
"+="... | Remove padding with null bytes | [
"Remove",
"padding",
"with",
"null",
"bytes"
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/Padding.py#L175-L184 |
14,409 | keenlabs/KeenClient-Python | keen/Padding.py | appendSpacePadding | def appendSpacePadding(str, blocksize=AES_blocksize):
'Pad with spaces'
pad_len = paddingLength(len(str), blocksize)
padding = '\0'*pad_len
return str + padding | python | def appendSpacePadding(str, blocksize=AES_blocksize):
'Pad with spaces'
pad_len = paddingLength(len(str), blocksize)
padding = '\0'*pad_len
return str + padding | [
"def",
"appendSpacePadding",
"(",
"str",
",",
"blocksize",
"=",
"AES_blocksize",
")",
":",
"pad_len",
"=",
"paddingLength",
"(",
"len",
"(",
"str",
")",
",",
"blocksize",
")",
"padding",
"=",
"'\\0'",
"*",
"pad_len",
"return",
"str",
"+",
"padding"
] | Pad with spaces | [
"Pad",
"with",
"spaces"
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/Padding.py#L186-L192 |
14,410 | keenlabs/KeenClient-Python | keen/Padding.py | removeSpacePadding | def removeSpacePadding(str, blocksize=AES_blocksize):
'Remove padding with spaces'
pad_len = 0
for char in str[::-1]: # str[::-1] reverses string
if char == ' ':
pad_len += 1
else:
break
str = str[:-pad_len]
return str | python | def removeSpacePadding(str, blocksize=AES_blocksize):
'Remove padding with spaces'
pad_len = 0
for char in str[::-1]: # str[::-1] reverses string
if char == ' ':
pad_len += 1
else:
break
str = str[:-pad_len]
return str | [
"def",
"removeSpacePadding",
"(",
"str",
",",
"blocksize",
"=",
"AES_blocksize",
")",
":",
"pad_len",
"=",
"0",
"for",
"char",
"in",
"str",
"[",
":",
":",
"-",
"1",
"]",
":",
"# str[::-1] reverses string",
"if",
"char",
"==",
"' '",
":",
"pad_len",
"+=",... | Remove padding with spaces | [
"Remove",
"padding",
"with",
"spaces"
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/Padding.py#L194-L205 |
14,411 | quintusdias/glymur | glymur/jp2k.py | _default_error_handler | def _default_error_handler(msg, _):
"""Default error handler callback for libopenjp2."""
msg = "OpenJPEG library error: {0}".format(msg.decode('utf-8').rstrip())
opj2.set_error_message(msg) | python | def _default_error_handler(msg, _):
msg = "OpenJPEG library error: {0}".format(msg.decode('utf-8').rstrip())
opj2.set_error_message(msg) | [
"def",
"_default_error_handler",
"(",
"msg",
",",
"_",
")",
":",
"msg",
"=",
"\"OpenJPEG library error: {0}\"",
".",
"format",
"(",
"msg",
".",
"decode",
"(",
"'utf-8'",
")",
".",
"rstrip",
"(",
")",
")",
"opj2",
".",
"set_error_message",
"(",
"msg",
")"
... | Default error handler callback for libopenjp2. | [
"Default",
"error",
"handler",
"callback",
"for",
"libopenjp2",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1874-L1877 |
14,412 | quintusdias/glymur | glymur/jp2k.py | _default_warning_handler | def _default_warning_handler(library_msg, _):
"""Default warning handler callback."""
library_msg = library_msg.decode('utf-8').rstrip()
msg = "OpenJPEG library warning: {0}".format(library_msg)
warnings.warn(msg, UserWarning) | python | def _default_warning_handler(library_msg, _):
library_msg = library_msg.decode('utf-8').rstrip()
msg = "OpenJPEG library warning: {0}".format(library_msg)
warnings.warn(msg, UserWarning) | [
"def",
"_default_warning_handler",
"(",
"library_msg",
",",
"_",
")",
":",
"library_msg",
"=",
"library_msg",
".",
"decode",
"(",
"'utf-8'",
")",
".",
"rstrip",
"(",
")",
"msg",
"=",
"\"OpenJPEG library warning: {0}\"",
".",
"format",
"(",
"library_msg",
")",
... | Default warning handler callback. | [
"Default",
"warning",
"handler",
"callback",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1885-L1889 |
14,413 | quintusdias/glymur | glymur/jp2k.py | Jp2k.parse | def parse(self):
"""Parses the JPEG 2000 file.
Raises
------
IOError
The file was not JPEG 2000.
"""
self.length = os.path.getsize(self.filename)
with open(self.filename, 'rb') as fptr:
# Make sure we have a JPEG2000 file. It could be either JP2 or
# J2C. Check for J2C first, single box in that case.
read_buffer = fptr.read(2)
signature, = struct.unpack('>H', read_buffer)
if signature == 0xff4f:
self._codec_format = opj2.CODEC_J2K
# That's it, we're done. The codestream object is only
# produced upon explicit request.
return
self._codec_format = opj2.CODEC_JP2
# Should be JP2.
# First 4 bytes should be 12, the length of the 'jP ' box.
# 2nd 4 bytes should be the box ID ('jP ').
# 3rd 4 bytes should be the box signature (13, 10, 135, 10).
fptr.seek(0)
read_buffer = fptr.read(12)
values = struct.unpack('>I4s4B', read_buffer)
box_length = values[0]
box_id = values[1]
signature = values[2:]
if (((box_length != 12) or (box_id != b'jP ') or
(signature != (13, 10, 135, 10)))):
msg = '{filename} is not a JPEG 2000 file.'
msg = msg.format(filename=self.filename)
raise IOError(msg)
# Back up and start again, we know we have a superbox (box of
# boxes) here.
fptr.seek(0)
self.box = self.parse_superbox(fptr)
self._validate() | python | def parse(self):
self.length = os.path.getsize(self.filename)
with open(self.filename, 'rb') as fptr:
# Make sure we have a JPEG2000 file. It could be either JP2 or
# J2C. Check for J2C first, single box in that case.
read_buffer = fptr.read(2)
signature, = struct.unpack('>H', read_buffer)
if signature == 0xff4f:
self._codec_format = opj2.CODEC_J2K
# That's it, we're done. The codestream object is only
# produced upon explicit request.
return
self._codec_format = opj2.CODEC_JP2
# Should be JP2.
# First 4 bytes should be 12, the length of the 'jP ' box.
# 2nd 4 bytes should be the box ID ('jP ').
# 3rd 4 bytes should be the box signature (13, 10, 135, 10).
fptr.seek(0)
read_buffer = fptr.read(12)
values = struct.unpack('>I4s4B', read_buffer)
box_length = values[0]
box_id = values[1]
signature = values[2:]
if (((box_length != 12) or (box_id != b'jP ') or
(signature != (13, 10, 135, 10)))):
msg = '{filename} is not a JPEG 2000 file.'
msg = msg.format(filename=self.filename)
raise IOError(msg)
# Back up and start again, we know we have a superbox (box of
# boxes) here.
fptr.seek(0)
self.box = self.parse_superbox(fptr)
self._validate() | [
"def",
"parse",
"(",
"self",
")",
":",
"self",
".",
"length",
"=",
"os",
".",
"path",
".",
"getsize",
"(",
"self",
".",
"filename",
")",
"with",
"open",
"(",
"self",
".",
"filename",
",",
"'rb'",
")",
"as",
"fptr",
":",
"# Make sure we have a JPEG2000 ... | Parses the JPEG 2000 file.
Raises
------
IOError
The file was not JPEG 2000. | [
"Parses",
"the",
"JPEG",
"2000",
"file",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L262-L306 |
14,414 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate | def _validate(self):
"""Validate the JPEG 2000 outermost superbox. These checks must be
done at a file level.
"""
# A JP2 file must contain certain boxes. The 2nd box must be a file
# type box.
if not isinstance(self.box[1], FileTypeBox):
msg = "{filename} does not contain a valid File Type box."
msg = msg.format(filename=self.filename)
raise IOError(msg)
# A jp2-branded file cannot contain an "any ICC profile
ftyp = self.box[1]
if ftyp.brand == 'jp2 ':
jp2h = [box for box in self.box if box.box_id == 'jp2h'][0]
colrs = [box for box in jp2h.box if box.box_id == 'colr']
for colr in colrs:
if colr.method not in (core.ENUMERATED_COLORSPACE,
core.RESTRICTED_ICC_PROFILE):
msg = ("Color Specification box method must specify "
"either an enumerated colorspace or a restricted "
"ICC profile if the file type box brand is 'jp2 '.")
warnings.warn(msg, UserWarning) | python | def _validate(self):
# A JP2 file must contain certain boxes. The 2nd box must be a file
# type box.
if not isinstance(self.box[1], FileTypeBox):
msg = "{filename} does not contain a valid File Type box."
msg = msg.format(filename=self.filename)
raise IOError(msg)
# A jp2-branded file cannot contain an "any ICC profile
ftyp = self.box[1]
if ftyp.brand == 'jp2 ':
jp2h = [box for box in self.box if box.box_id == 'jp2h'][0]
colrs = [box for box in jp2h.box if box.box_id == 'colr']
for colr in colrs:
if colr.method not in (core.ENUMERATED_COLORSPACE,
core.RESTRICTED_ICC_PROFILE):
msg = ("Color Specification box method must specify "
"either an enumerated colorspace or a restricted "
"ICC profile if the file type box brand is 'jp2 '.")
warnings.warn(msg, UserWarning) | [
"def",
"_validate",
"(",
"self",
")",
":",
"# A JP2 file must contain certain boxes. The 2nd box must be a file",
"# type box.",
"if",
"not",
"isinstance",
"(",
"self",
".",
"box",
"[",
"1",
"]",
",",
"FileTypeBox",
")",
":",
"msg",
"=",
"\"{filename} does not contai... | Validate the JPEG 2000 outermost superbox. These checks must be
done at a file level. | [
"Validate",
"the",
"JPEG",
"2000",
"outermost",
"superbox",
".",
"These",
"checks",
"must",
"be",
"done",
"at",
"a",
"file",
"level",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L308-L330 |
14,415 | quintusdias/glymur | glymur/jp2k.py | Jp2k._set_cinema_params | def _set_cinema_params(self, cinema_mode, fps):
"""Populate compression parameters structure for cinema2K.
Parameters
----------
params : ctypes struct
Corresponds to compression parameters structure used by the
library.
cinema_mode : {'cinema2k', 'cinema4k}
Use either Cinema2K or Cinema4K profile.
fps : {24, 48}
Frames per second.
"""
if re.match("1.5|2.0", version.openjpeg_version) is not None:
msg = ("Writing Cinema2K or Cinema4K files is not supported with "
"OpenJPEG library versions less than 2.1.0. The installed "
"version of OpenJPEG is {version}.")
msg = msg.format(version=version.openjpeg_version)
raise IOError(msg)
# Cinema modes imply MCT.
self._cparams.tcp_mct = 1
if cinema_mode == 'cinema2k':
if fps not in [24, 48]:
msg = 'Cinema2K frame rate must be either 24 or 48.'
raise IOError(msg)
if fps == 24:
self._cparams.rsiz = core.OPJ_PROFILE_CINEMA_2K
self._cparams.max_comp_size = core.OPJ_CINEMA_24_COMP
self._cparams.max_cs_size = core.OPJ_CINEMA_24_CS
else:
self._cparams.rsiz = core.OPJ_PROFILE_CINEMA_2K
self._cparams.max_comp_size = core.OPJ_CINEMA_48_COMP
self._cparams.max_cs_size = core.OPJ_CINEMA_48_CS
else:
# cinema4k
self._cparams.rsiz = core.OPJ_PROFILE_CINEMA_4K | python | def _set_cinema_params(self, cinema_mode, fps):
if re.match("1.5|2.0", version.openjpeg_version) is not None:
msg = ("Writing Cinema2K or Cinema4K files is not supported with "
"OpenJPEG library versions less than 2.1.0. The installed "
"version of OpenJPEG is {version}.")
msg = msg.format(version=version.openjpeg_version)
raise IOError(msg)
# Cinema modes imply MCT.
self._cparams.tcp_mct = 1
if cinema_mode == 'cinema2k':
if fps not in [24, 48]:
msg = 'Cinema2K frame rate must be either 24 or 48.'
raise IOError(msg)
if fps == 24:
self._cparams.rsiz = core.OPJ_PROFILE_CINEMA_2K
self._cparams.max_comp_size = core.OPJ_CINEMA_24_COMP
self._cparams.max_cs_size = core.OPJ_CINEMA_24_CS
else:
self._cparams.rsiz = core.OPJ_PROFILE_CINEMA_2K
self._cparams.max_comp_size = core.OPJ_CINEMA_48_COMP
self._cparams.max_cs_size = core.OPJ_CINEMA_48_CS
else:
# cinema4k
self._cparams.rsiz = core.OPJ_PROFILE_CINEMA_4K | [
"def",
"_set_cinema_params",
"(",
"self",
",",
"cinema_mode",
",",
"fps",
")",
":",
"if",
"re",
".",
"match",
"(",
"\"1.5|2.0\"",
",",
"version",
".",
"openjpeg_version",
")",
"is",
"not",
"None",
":",
"msg",
"=",
"(",
"\"Writing Cinema2K or Cinema4K files is ... | Populate compression parameters structure for cinema2K.
Parameters
----------
params : ctypes struct
Corresponds to compression parameters structure used by the
library.
cinema_mode : {'cinema2k', 'cinema4k}
Use either Cinema2K or Cinema4K profile.
fps : {24, 48}
Frames per second. | [
"Populate",
"compression",
"parameters",
"structure",
"for",
"cinema2K",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L332-L371 |
14,416 | quintusdias/glymur | glymur/jp2k.py | Jp2k._write_openjpeg | def _write_openjpeg(self, img_array, verbose=False):
"""
Write JPEG 2000 file using OpenJPEG 1.5 interface.
"""
if img_array.ndim == 2:
# Force the image to be 3D. Just makes things easier later on.
img_array = img_array.reshape(img_array.shape[0],
img_array.shape[1],
1)
self._populate_comptparms(img_array)
with ExitStack() as stack:
image = opj.image_create(self._comptparms, self._colorspace)
stack.callback(opj.image_destroy, image)
numrows, numcols, numlayers = img_array.shape
# set image offset and reference grid
image.contents.x0 = self._cparams.image_offset_x0
image.contents.y0 = self._cparams.image_offset_y0
image.contents.x1 = (image.contents.x0 +
(numcols - 1) * self._cparams.subsampling_dx +
1)
image.contents.y1 = (image.contents.y0 +
(numrows - 1) * self._cparams.subsampling_dy +
1)
# Stage the image data to the openjpeg data structure.
for k in range(0, numlayers):
layer = np.ascontiguousarray(img_array[:, :, k],
dtype=np.int32)
dest = image.contents.comps[k].data
src = layer.ctypes.data
ctypes.memmove(dest, src, layer.nbytes)
cinfo = opj.create_compress(self._cparams.codec_fmt)
stack.callback(opj.destroy_compress, cinfo)
# Setup the info, warning, and error handlers.
# Always use the warning and error handler. Use of an info
# handler is optional.
event_mgr = opj.EventMgrType()
_info_handler = _INFO_CALLBACK if verbose else None
event_mgr.info_handler = _info_handler
event_mgr.warning_handler = ctypes.cast(_WARNING_CALLBACK,
ctypes.c_void_p)
event_mgr.error_handler = ctypes.cast(_ERROR_CALLBACK,
ctypes.c_void_p)
opj.setup_encoder(cinfo, ctypes.byref(self._cparams), image)
cio = opj.cio_open(cinfo)
stack.callback(opj.cio_close, cio)
if not opj.encode(cinfo, cio, image):
raise IOError("Encode error.")
pos = opj.cio_tell(cio)
blob = ctypes.string_at(cio.contents.buffer, pos)
fptr = open(self.filename, 'wb')
stack.callback(fptr.close)
fptr.write(blob)
self.parse() | python | def _write_openjpeg(self, img_array, verbose=False):
if img_array.ndim == 2:
# Force the image to be 3D. Just makes things easier later on.
img_array = img_array.reshape(img_array.shape[0],
img_array.shape[1],
1)
self._populate_comptparms(img_array)
with ExitStack() as stack:
image = opj.image_create(self._comptparms, self._colorspace)
stack.callback(opj.image_destroy, image)
numrows, numcols, numlayers = img_array.shape
# set image offset and reference grid
image.contents.x0 = self._cparams.image_offset_x0
image.contents.y0 = self._cparams.image_offset_y0
image.contents.x1 = (image.contents.x0 +
(numcols - 1) * self._cparams.subsampling_dx +
1)
image.contents.y1 = (image.contents.y0 +
(numrows - 1) * self._cparams.subsampling_dy +
1)
# Stage the image data to the openjpeg data structure.
for k in range(0, numlayers):
layer = np.ascontiguousarray(img_array[:, :, k],
dtype=np.int32)
dest = image.contents.comps[k].data
src = layer.ctypes.data
ctypes.memmove(dest, src, layer.nbytes)
cinfo = opj.create_compress(self._cparams.codec_fmt)
stack.callback(opj.destroy_compress, cinfo)
# Setup the info, warning, and error handlers.
# Always use the warning and error handler. Use of an info
# handler is optional.
event_mgr = opj.EventMgrType()
_info_handler = _INFO_CALLBACK if verbose else None
event_mgr.info_handler = _info_handler
event_mgr.warning_handler = ctypes.cast(_WARNING_CALLBACK,
ctypes.c_void_p)
event_mgr.error_handler = ctypes.cast(_ERROR_CALLBACK,
ctypes.c_void_p)
opj.setup_encoder(cinfo, ctypes.byref(self._cparams), image)
cio = opj.cio_open(cinfo)
stack.callback(opj.cio_close, cio)
if not opj.encode(cinfo, cio, image):
raise IOError("Encode error.")
pos = opj.cio_tell(cio)
blob = ctypes.string_at(cio.contents.buffer, pos)
fptr = open(self.filename, 'wb')
stack.callback(fptr.close)
fptr.write(blob)
self.parse() | [
"def",
"_write_openjpeg",
"(",
"self",
",",
"img_array",
",",
"verbose",
"=",
"False",
")",
":",
"if",
"img_array",
".",
"ndim",
"==",
"2",
":",
"# Force the image to be 3D. Just makes things easier later on.",
"img_array",
"=",
"img_array",
".",
"reshape",
"(",
... | Write JPEG 2000 file using OpenJPEG 1.5 interface. | [
"Write",
"JPEG",
"2000",
"file",
"using",
"OpenJPEG",
"1",
".",
"5",
"interface",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L517-L582 |
14,417 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_j2k_colorspace | def _validate_j2k_colorspace(self, cparams, colorspace):
"""
Cannot specify a colorspace with J2K.
"""
if cparams.codec_fmt == opj2.CODEC_J2K and colorspace is not None:
msg = 'Do not specify a colorspace when writing a raw codestream.'
raise IOError(msg) | python | def _validate_j2k_colorspace(self, cparams, colorspace):
if cparams.codec_fmt == opj2.CODEC_J2K and colorspace is not None:
msg = 'Do not specify a colorspace when writing a raw codestream.'
raise IOError(msg) | [
"def",
"_validate_j2k_colorspace",
"(",
"self",
",",
"cparams",
",",
"colorspace",
")",
":",
"if",
"cparams",
".",
"codec_fmt",
"==",
"opj2",
".",
"CODEC_J2K",
"and",
"colorspace",
"is",
"not",
"None",
":",
"msg",
"=",
"'Do not specify a colorspace when writing a ... | Cannot specify a colorspace with J2K. | [
"Cannot",
"specify",
"a",
"colorspace",
"with",
"J2K",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L584-L590 |
14,418 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_codeblock_size | def _validate_codeblock_size(self, cparams):
"""
Code block dimensions must satisfy certain restrictions.
They must both be a power of 2 and the total area defined by the width
and height cannot be either too great or too small for the codec.
"""
if cparams.cblockw_init != 0 and cparams.cblockh_init != 0:
# These fields ARE zero if uninitialized.
width = cparams.cblockw_init
height = cparams.cblockh_init
if height * width > 4096 or height < 4 or width < 4:
msg = ("The code block area is specified as "
"{height} x {width} = {area} square pixels. "
"Code block area cannot exceed 4096 square pixels. "
"Code block height and width dimensions must be larger "
"than 4 pixels.")
msg = msg.format(height=height, width=width,
area=height * width)
raise IOError(msg)
if ((math.log(height, 2) != math.floor(math.log(height, 2)) or
math.log(width, 2) != math.floor(math.log(width, 2)))):
msg = ("Bad code block size ({height} x {width}). "
"The dimensions must be powers of 2.")
msg = msg.format(height=height, width=width)
raise IOError(msg) | python | def _validate_codeblock_size(self, cparams):
if cparams.cblockw_init != 0 and cparams.cblockh_init != 0:
# These fields ARE zero if uninitialized.
width = cparams.cblockw_init
height = cparams.cblockh_init
if height * width > 4096 or height < 4 or width < 4:
msg = ("The code block area is specified as "
"{height} x {width} = {area} square pixels. "
"Code block area cannot exceed 4096 square pixels. "
"Code block height and width dimensions must be larger "
"than 4 pixels.")
msg = msg.format(height=height, width=width,
area=height * width)
raise IOError(msg)
if ((math.log(height, 2) != math.floor(math.log(height, 2)) or
math.log(width, 2) != math.floor(math.log(width, 2)))):
msg = ("Bad code block size ({height} x {width}). "
"The dimensions must be powers of 2.")
msg = msg.format(height=height, width=width)
raise IOError(msg) | [
"def",
"_validate_codeblock_size",
"(",
"self",
",",
"cparams",
")",
":",
"if",
"cparams",
".",
"cblockw_init",
"!=",
"0",
"and",
"cparams",
".",
"cblockh_init",
"!=",
"0",
":",
"# These fields ARE zero if uninitialized.",
"width",
"=",
"cparams",
".",
"cblockw_in... | Code block dimensions must satisfy certain restrictions.
They must both be a power of 2 and the total area defined by the width
and height cannot be either too great or too small for the codec. | [
"Code",
"block",
"dimensions",
"must",
"satisfy",
"certain",
"restrictions",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L592-L617 |
14,419 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_precinct_size | def _validate_precinct_size(self, cparams):
"""
Precinct dimensions must satisfy certain restrictions if specified.
They must both be a power of 2 and must both be at least twice the
size of their codeblock size counterparts.
"""
code_block_specified = False
if cparams.cblockw_init != 0 and cparams.cblockh_init != 0:
code_block_specified = True
if cparams.res_spec != 0:
# precinct size was not specified if this field is zero.
for j in range(cparams.res_spec):
prch = cparams.prch_init[j]
prcw = cparams.prcw_init[j]
if j == 0 and code_block_specified:
height, width = cparams.cblockh_init, cparams.cblockw_init
if prch < height * 2 or prcw < width * 2:
msg = ("The highest resolution precinct size "
"({prch} x {prcw}) must be at least twice that "
"of the code block size "
"({cbh} x {cbw}).")
msg = msg.format(prch=prch, prcw=prcw,
cbh=height, cbw=width)
raise IOError(msg)
if ((math.log(prch, 2) != math.floor(math.log(prch, 2)) or
math.log(prcw, 2) != math.floor(math.log(prcw, 2)))):
msg = ("Bad precinct size ({height} x {width}). "
"Precinct dimensions must be powers of 2.")
msg = msg.format(height=prch, width=prcw)
raise IOError(msg) | python | def _validate_precinct_size(self, cparams):
code_block_specified = False
if cparams.cblockw_init != 0 and cparams.cblockh_init != 0:
code_block_specified = True
if cparams.res_spec != 0:
# precinct size was not specified if this field is zero.
for j in range(cparams.res_spec):
prch = cparams.prch_init[j]
prcw = cparams.prcw_init[j]
if j == 0 and code_block_specified:
height, width = cparams.cblockh_init, cparams.cblockw_init
if prch < height * 2 or prcw < width * 2:
msg = ("The highest resolution precinct size "
"({prch} x {prcw}) must be at least twice that "
"of the code block size "
"({cbh} x {cbw}).")
msg = msg.format(prch=prch, prcw=prcw,
cbh=height, cbw=width)
raise IOError(msg)
if ((math.log(prch, 2) != math.floor(math.log(prch, 2)) or
math.log(prcw, 2) != math.floor(math.log(prcw, 2)))):
msg = ("Bad precinct size ({height} x {width}). "
"Precinct dimensions must be powers of 2.")
msg = msg.format(height=prch, width=prcw)
raise IOError(msg) | [
"def",
"_validate_precinct_size",
"(",
"self",
",",
"cparams",
")",
":",
"code_block_specified",
"=",
"False",
"if",
"cparams",
".",
"cblockw_init",
"!=",
"0",
"and",
"cparams",
".",
"cblockh_init",
"!=",
"0",
":",
"code_block_specified",
"=",
"True",
"if",
"c... | Precinct dimensions must satisfy certain restrictions if specified.
They must both be a power of 2 and must both be at least twice the
size of their codeblock size counterparts. | [
"Precinct",
"dimensions",
"must",
"satisfy",
"certain",
"restrictions",
"if",
"specified",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L619-L650 |
14,420 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_image_rank | def _validate_image_rank(self, img_array):
"""
Images must be either 2D or 3D.
"""
if img_array.ndim == 1 or img_array.ndim > 3:
msg = "{0}D imagery is not allowed.".format(img_array.ndim)
raise IOError(msg) | python | def _validate_image_rank(self, img_array):
if img_array.ndim == 1 or img_array.ndim > 3:
msg = "{0}D imagery is not allowed.".format(img_array.ndim)
raise IOError(msg) | [
"def",
"_validate_image_rank",
"(",
"self",
",",
"img_array",
")",
":",
"if",
"img_array",
".",
"ndim",
"==",
"1",
"or",
"img_array",
".",
"ndim",
">",
"3",
":",
"msg",
"=",
"\"{0}D imagery is not allowed.\"",
".",
"format",
"(",
"img_array",
".",
"ndim",
... | Images must be either 2D or 3D. | [
"Images",
"must",
"be",
"either",
"2D",
"or",
"3D",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L652-L658 |
14,421 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_image_datatype | def _validate_image_datatype(self, img_array):
"""
Only uint8 and uint16 images are currently supported.
"""
if img_array.dtype != np.uint8 and img_array.dtype != np.uint16:
msg = ("Only uint8 and uint16 datatypes are currently supported "
"when writing.")
raise RuntimeError(msg) | python | def _validate_image_datatype(self, img_array):
if img_array.dtype != np.uint8 and img_array.dtype != np.uint16:
msg = ("Only uint8 and uint16 datatypes are currently supported "
"when writing.")
raise RuntimeError(msg) | [
"def",
"_validate_image_datatype",
"(",
"self",
",",
"img_array",
")",
":",
"if",
"img_array",
".",
"dtype",
"!=",
"np",
".",
"uint8",
"and",
"img_array",
".",
"dtype",
"!=",
"np",
".",
"uint16",
":",
"msg",
"=",
"(",
"\"Only uint8 and uint16 datatypes are cur... | Only uint8 and uint16 images are currently supported. | [
"Only",
"uint8",
"and",
"uint16",
"images",
"are",
"currently",
"supported",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L660-L667 |
14,422 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_compression_params | def _validate_compression_params(self, img_array, cparams, colorspace):
"""Check that the compression parameters are valid.
Parameters
----------
img_array : ndarray
Image data to be written to file.
cparams : CompressionParametersType(ctypes.Structure)
Corresponds to cparameters_t type in openjp2 headers.
"""
self._validate_j2k_colorspace(cparams, colorspace)
self._validate_codeblock_size(cparams)
self._validate_precinct_size(cparams)
self._validate_image_rank(img_array)
self._validate_image_datatype(img_array) | python | def _validate_compression_params(self, img_array, cparams, colorspace):
self._validate_j2k_colorspace(cparams, colorspace)
self._validate_codeblock_size(cparams)
self._validate_precinct_size(cparams)
self._validate_image_rank(img_array)
self._validate_image_datatype(img_array) | [
"def",
"_validate_compression_params",
"(",
"self",
",",
"img_array",
",",
"cparams",
",",
"colorspace",
")",
":",
"self",
".",
"_validate_j2k_colorspace",
"(",
"cparams",
",",
"colorspace",
")",
"self",
".",
"_validate_codeblock_size",
"(",
"cparams",
")",
"self"... | Check that the compression parameters are valid.
Parameters
----------
img_array : ndarray
Image data to be written to file.
cparams : CompressionParametersType(ctypes.Structure)
Corresponds to cparameters_t type in openjp2 headers. | [
"Check",
"that",
"the",
"compression",
"parameters",
"are",
"valid",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L669-L683 |
14,423 | quintusdias/glymur | glymur/jp2k.py | Jp2k._determine_colorspace | def _determine_colorspace(self, colorspace=None, **kwargs):
"""Determine the colorspace from the supplied inputs.
Parameters
----------
colorspace : str, optional
Either 'rgb' or 'gray'.
"""
if colorspace is None:
# Must infer the colorspace from the image dimensions.
if len(self.shape) < 3:
# A single channel image is grayscale.
self._colorspace = opj2.CLRSPC_GRAY
elif self.shape[2] == 1 or self.shape[2] == 2:
# A single channel image or an image with two channels is going
# to be greyscale.
self._colorspace = opj2.CLRSPC_GRAY
else:
# Anything else must be RGB, right?
self._colorspace = opj2.CLRSPC_SRGB
else:
if colorspace.lower() not in ('rgb', 'grey', 'gray'):
msg = 'Invalid colorspace "{0}".'.format(colorspace)
raise IOError(msg)
elif colorspace.lower() == 'rgb' and self.shape[2] < 3:
msg = 'RGB colorspace requires at least 3 components.'
raise IOError(msg)
# Turn the colorspace from a string to the enumerated value that
# the library expects.
COLORSPACE_MAP = {'rgb': opj2.CLRSPC_SRGB,
'gray': opj2.CLRSPC_GRAY,
'grey': opj2.CLRSPC_GRAY,
'ycc': opj2.CLRSPC_YCC}
self._colorspace = COLORSPACE_MAP[colorspace.lower()] | python | def _determine_colorspace(self, colorspace=None, **kwargs):
if colorspace is None:
# Must infer the colorspace from the image dimensions.
if len(self.shape) < 3:
# A single channel image is grayscale.
self._colorspace = opj2.CLRSPC_GRAY
elif self.shape[2] == 1 or self.shape[2] == 2:
# A single channel image or an image with two channels is going
# to be greyscale.
self._colorspace = opj2.CLRSPC_GRAY
else:
# Anything else must be RGB, right?
self._colorspace = opj2.CLRSPC_SRGB
else:
if colorspace.lower() not in ('rgb', 'grey', 'gray'):
msg = 'Invalid colorspace "{0}".'.format(colorspace)
raise IOError(msg)
elif colorspace.lower() == 'rgb' and self.shape[2] < 3:
msg = 'RGB colorspace requires at least 3 components.'
raise IOError(msg)
# Turn the colorspace from a string to the enumerated value that
# the library expects.
COLORSPACE_MAP = {'rgb': opj2.CLRSPC_SRGB,
'gray': opj2.CLRSPC_GRAY,
'grey': opj2.CLRSPC_GRAY,
'ycc': opj2.CLRSPC_YCC}
self._colorspace = COLORSPACE_MAP[colorspace.lower()] | [
"def",
"_determine_colorspace",
"(",
"self",
",",
"colorspace",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"colorspace",
"is",
"None",
":",
"# Must infer the colorspace from the image dimensions.",
"if",
"len",
"(",
"self",
".",
"shape",
")",
"<",
"... | Determine the colorspace from the supplied inputs.
Parameters
----------
colorspace : str, optional
Either 'rgb' or 'gray'. | [
"Determine",
"the",
"colorspace",
"from",
"the",
"supplied",
"inputs",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L685-L720 |
14,424 | quintusdias/glymur | glymur/jp2k.py | Jp2k._write_openjp2 | def _write_openjp2(self, img_array, verbose=False):
"""
Write JPEG 2000 file using OpenJPEG 2.x interface.
"""
if img_array.ndim == 2:
# Force the image to be 3D. Just makes things easier later on.
numrows, numcols = img_array.shape
img_array = img_array.reshape(numrows, numcols, 1)
self._populate_comptparms(img_array)
with ExitStack() as stack:
image = opj2.image_create(self._comptparms, self._colorspace)
stack.callback(opj2.image_destroy, image)
self._populate_image_struct(image, img_array)
codec = opj2.create_compress(self._cparams.codec_fmt)
stack.callback(opj2.destroy_codec, codec)
if self._verbose or verbose:
info_handler = _INFO_CALLBACK
else:
info_handler = None
opj2.set_info_handler(codec, info_handler)
opj2.set_warning_handler(codec, _WARNING_CALLBACK)
opj2.set_error_handler(codec, _ERROR_CALLBACK)
opj2.setup_encoder(codec, self._cparams, image)
strm = opj2.stream_create_default_file_stream(self.filename,
False)
stack.callback(opj2.stream_destroy, strm)
opj2.start_compress(codec, image, strm)
opj2.encode(codec, strm)
opj2.end_compress(codec, strm)
# Refresh the metadata.
self.parse() | python | def _write_openjp2(self, img_array, verbose=False):
if img_array.ndim == 2:
# Force the image to be 3D. Just makes things easier later on.
numrows, numcols = img_array.shape
img_array = img_array.reshape(numrows, numcols, 1)
self._populate_comptparms(img_array)
with ExitStack() as stack:
image = opj2.image_create(self._comptparms, self._colorspace)
stack.callback(opj2.image_destroy, image)
self._populate_image_struct(image, img_array)
codec = opj2.create_compress(self._cparams.codec_fmt)
stack.callback(opj2.destroy_codec, codec)
if self._verbose or verbose:
info_handler = _INFO_CALLBACK
else:
info_handler = None
opj2.set_info_handler(codec, info_handler)
opj2.set_warning_handler(codec, _WARNING_CALLBACK)
opj2.set_error_handler(codec, _ERROR_CALLBACK)
opj2.setup_encoder(codec, self._cparams, image)
strm = opj2.stream_create_default_file_stream(self.filename,
False)
stack.callback(opj2.stream_destroy, strm)
opj2.start_compress(codec, image, strm)
opj2.encode(codec, strm)
opj2.end_compress(codec, strm)
# Refresh the metadata.
self.parse() | [
"def",
"_write_openjp2",
"(",
"self",
",",
"img_array",
",",
"verbose",
"=",
"False",
")",
":",
"if",
"img_array",
".",
"ndim",
"==",
"2",
":",
"# Force the image to be 3D. Just makes things easier later on.",
"numrows",
",",
"numcols",
"=",
"img_array",
".",
"sh... | Write JPEG 2000 file using OpenJPEG 2.x interface. | [
"Write",
"JPEG",
"2000",
"file",
"using",
"OpenJPEG",
"2",
".",
"x",
"interface",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L722-L762 |
14,425 | quintusdias/glymur | glymur/jp2k.py | Jp2k.append | def append(self, box):
"""Append a JP2 box to the file in-place.
Parameters
----------
box : Jp2Box
Instance of a JP2 box. Only UUID and XML boxes can currently be
appended.
"""
if self._codec_format == opj2.CODEC_J2K:
msg = "Only JP2 files can currently have boxes appended to them."
raise IOError(msg)
if not ((box.box_id == 'xml ') or
(box.box_id == 'uuid' and
box.uuid == UUID('be7acfcb-97a9-42e8-9c71-999491e3afac'))):
msg = ("Only XML boxes and XMP UUID boxes can currently be "
"appended.")
raise IOError(msg)
# Check the last box. If the length field is zero, then rewrite
# the length field to reflect the true length of the box.
with open(self.filename, 'rb') as ifile:
offset = self.box[-1].offset
ifile.seek(offset)
read_buffer = ifile.read(4)
box_length, = struct.unpack('>I', read_buffer)
if box_length == 0:
# Reopen the file in write mode and rewrite the length field.
true_box_length = os.path.getsize(ifile.name) - offset
with open(self.filename, 'r+b') as ofile:
ofile.seek(offset)
write_buffer = struct.pack('>I', true_box_length)
ofile.write(write_buffer)
# Can now safely append the box.
with open(self.filename, 'ab') as ofile:
box.write(ofile)
self.parse() | python | def append(self, box):
if self._codec_format == opj2.CODEC_J2K:
msg = "Only JP2 files can currently have boxes appended to them."
raise IOError(msg)
if not ((box.box_id == 'xml ') or
(box.box_id == 'uuid' and
box.uuid == UUID('be7acfcb-97a9-42e8-9c71-999491e3afac'))):
msg = ("Only XML boxes and XMP UUID boxes can currently be "
"appended.")
raise IOError(msg)
# Check the last box. If the length field is zero, then rewrite
# the length field to reflect the true length of the box.
with open(self.filename, 'rb') as ifile:
offset = self.box[-1].offset
ifile.seek(offset)
read_buffer = ifile.read(4)
box_length, = struct.unpack('>I', read_buffer)
if box_length == 0:
# Reopen the file in write mode and rewrite the length field.
true_box_length = os.path.getsize(ifile.name) - offset
with open(self.filename, 'r+b') as ofile:
ofile.seek(offset)
write_buffer = struct.pack('>I', true_box_length)
ofile.write(write_buffer)
# Can now safely append the box.
with open(self.filename, 'ab') as ofile:
box.write(ofile)
self.parse() | [
"def",
"append",
"(",
"self",
",",
"box",
")",
":",
"if",
"self",
".",
"_codec_format",
"==",
"opj2",
".",
"CODEC_J2K",
":",
"msg",
"=",
"\"Only JP2 files can currently have boxes appended to them.\"",
"raise",
"IOError",
"(",
"msg",
")",
"if",
"not",
"(",
"("... | Append a JP2 box to the file in-place.
Parameters
----------
box : Jp2Box
Instance of a JP2 box. Only UUID and XML boxes can currently be
appended. | [
"Append",
"a",
"JP2",
"box",
"to",
"the",
"file",
"in",
"-",
"place",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L764-L803 |
14,426 | quintusdias/glymur | glymur/jp2k.py | Jp2k._write_wrapped_codestream | def _write_wrapped_codestream(self, ofile, box):
"""Write wrapped codestream."""
# Codestreams require a bit more care.
# Am I a raw codestream?
if len(self.box) == 0:
# Yes, just write the codestream box header plus all
# of myself out to file.
ofile.write(struct.pack('>I', self.length + 8))
ofile.write(b'jp2c')
with open(self.filename, 'rb') as ifile:
ofile.write(ifile.read())
return
# OK, I'm a jp2/jpx file. Need to find out where the raw codestream
# actually starts.
offset = box.offset
if offset == -1:
if self.box[1].brand == 'jpx ':
msg = ("The codestream box must have its offset and length "
"attributes fully specified if the file type brand is "
"JPX.")
raise IOError(msg)
# Find the first codestream in the file.
jp2c = [_box for _box in self.box if _box.box_id == 'jp2c']
offset = jp2c[0].offset
# Ready to write the codestream.
with open(self.filename, 'rb') as ifile:
ifile.seek(offset)
# Verify that the specified codestream is right.
read_buffer = ifile.read(8)
L, T = struct.unpack_from('>I4s', read_buffer, 0)
if T != b'jp2c':
msg = "Unable to locate the specified codestream."
raise IOError(msg)
if L == 0:
# The length of the box is presumed to last until the end of
# the file. Compute the effective length of the box.
L = os.path.getsize(ifile.name) - ifile.tell() + 8
elif L == 1:
# The length of the box is in the XL field, a 64-bit value.
read_buffer = ifile.read(8)
L, = struct.unpack('>Q', read_buffer)
ifile.seek(offset)
read_buffer = ifile.read(L)
ofile.write(read_buffer) | python | def _write_wrapped_codestream(self, ofile, box):
# Codestreams require a bit more care.
# Am I a raw codestream?
if len(self.box) == 0:
# Yes, just write the codestream box header plus all
# of myself out to file.
ofile.write(struct.pack('>I', self.length + 8))
ofile.write(b'jp2c')
with open(self.filename, 'rb') as ifile:
ofile.write(ifile.read())
return
# OK, I'm a jp2/jpx file. Need to find out where the raw codestream
# actually starts.
offset = box.offset
if offset == -1:
if self.box[1].brand == 'jpx ':
msg = ("The codestream box must have its offset and length "
"attributes fully specified if the file type brand is "
"JPX.")
raise IOError(msg)
# Find the first codestream in the file.
jp2c = [_box for _box in self.box if _box.box_id == 'jp2c']
offset = jp2c[0].offset
# Ready to write the codestream.
with open(self.filename, 'rb') as ifile:
ifile.seek(offset)
# Verify that the specified codestream is right.
read_buffer = ifile.read(8)
L, T = struct.unpack_from('>I4s', read_buffer, 0)
if T != b'jp2c':
msg = "Unable to locate the specified codestream."
raise IOError(msg)
if L == 0:
# The length of the box is presumed to last until the end of
# the file. Compute the effective length of the box.
L = os.path.getsize(ifile.name) - ifile.tell() + 8
elif L == 1:
# The length of the box is in the XL field, a 64-bit value.
read_buffer = ifile.read(8)
L, = struct.unpack('>Q', read_buffer)
ifile.seek(offset)
read_buffer = ifile.read(L)
ofile.write(read_buffer) | [
"def",
"_write_wrapped_codestream",
"(",
"self",
",",
"ofile",
",",
"box",
")",
":",
"# Codestreams require a bit more care.",
"# Am I a raw codestream?",
"if",
"len",
"(",
"self",
".",
"box",
")",
"==",
"0",
":",
"# Yes, just write the codestream box header plus all",
... | Write wrapped codestream. | [
"Write",
"wrapped",
"codestream",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L853-L902 |
14,427 | quintusdias/glymur | glymur/jp2k.py | Jp2k._get_default_jp2_boxes | def _get_default_jp2_boxes(self):
"""Create a default set of JP2 boxes."""
# Try to create a reasonable default.
boxes = [JPEG2000SignatureBox(),
FileTypeBox(),
JP2HeaderBox(),
ContiguousCodestreamBox()]
height = self.codestream.segment[1].ysiz
width = self.codestream.segment[1].xsiz
num_components = len(self.codestream.segment[1].xrsiz)
if num_components < 3:
colorspace = core.GREYSCALE
else:
if len(self.box) == 0:
# Best guess is SRGB
colorspace = core.SRGB
else:
# Take whatever the first jp2 header / color specification
# says.
jp2hs = [box for box in self.box if box.box_id == 'jp2h']
colorspace = jp2hs[0].box[1].colorspace
boxes[2].box = [ImageHeaderBox(height=height, width=width,
num_components=num_components),
ColourSpecificationBox(colorspace=colorspace)]
return boxes | python | def _get_default_jp2_boxes(self):
# Try to create a reasonable default.
boxes = [JPEG2000SignatureBox(),
FileTypeBox(),
JP2HeaderBox(),
ContiguousCodestreamBox()]
height = self.codestream.segment[1].ysiz
width = self.codestream.segment[1].xsiz
num_components = len(self.codestream.segment[1].xrsiz)
if num_components < 3:
colorspace = core.GREYSCALE
else:
if len(self.box) == 0:
# Best guess is SRGB
colorspace = core.SRGB
else:
# Take whatever the first jp2 header / color specification
# says.
jp2hs = [box for box in self.box if box.box_id == 'jp2h']
colorspace = jp2hs[0].box[1].colorspace
boxes[2].box = [ImageHeaderBox(height=height, width=width,
num_components=num_components),
ColourSpecificationBox(colorspace=colorspace)]
return boxes | [
"def",
"_get_default_jp2_boxes",
"(",
"self",
")",
":",
"# Try to create a reasonable default.",
"boxes",
"=",
"[",
"JPEG2000SignatureBox",
"(",
")",
",",
"FileTypeBox",
"(",
")",
",",
"JP2HeaderBox",
"(",
")",
",",
"ContiguousCodestreamBox",
"(",
")",
"]",
"heigh... | Create a default set of JP2 boxes. | [
"Create",
"a",
"default",
"set",
"of",
"JP2",
"boxes",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L904-L930 |
14,428 | quintusdias/glymur | glymur/jp2k.py | Jp2k._remove_ellipsis | def _remove_ellipsis(self, index, numrows, numcols, numbands):
"""
resolve the first ellipsis in the index so that it references the image
Parameters
----------
index : tuple
tuple of index arguments, presumably one of them is the Ellipsis
numrows, numcols, numbands : int
image dimensions
Returns
-------
tuple
Same as index, except that the first Ellipsis is replaced with
a proper slice whose start and stop members are not None
"""
# Remove the first ellipsis we find.
rows = slice(0, numrows)
cols = slice(0, numcols)
bands = slice(0, numbands)
if index[0] is Ellipsis:
if len(index) == 2:
# jp2k[..., other_slice]
newindex = (rows, cols, index[1])
else:
# jp2k[..., cols, bands]
newindex = (rows, index[1], index[2])
elif index[1] is Ellipsis:
if len(index) == 2:
# jp2k[rows, ...]
newindex = (index[0], cols, bands)
else:
# jp2k[rows, ..., bands]
newindex = (index[0], cols, index[2])
else:
# Assume that we don't have 4D imagery, of course.
newindex = (index[0], index[1], bands)
return newindex | python | def _remove_ellipsis(self, index, numrows, numcols, numbands):
# Remove the first ellipsis we find.
rows = slice(0, numrows)
cols = slice(0, numcols)
bands = slice(0, numbands)
if index[0] is Ellipsis:
if len(index) == 2:
# jp2k[..., other_slice]
newindex = (rows, cols, index[1])
else:
# jp2k[..., cols, bands]
newindex = (rows, index[1], index[2])
elif index[1] is Ellipsis:
if len(index) == 2:
# jp2k[rows, ...]
newindex = (index[0], cols, bands)
else:
# jp2k[rows, ..., bands]
newindex = (index[0], cols, index[2])
else:
# Assume that we don't have 4D imagery, of course.
newindex = (index[0], index[1], bands)
return newindex | [
"def",
"_remove_ellipsis",
"(",
"self",
",",
"index",
",",
"numrows",
",",
"numcols",
",",
"numbands",
")",
":",
"# Remove the first ellipsis we find.",
"rows",
"=",
"slice",
"(",
"0",
",",
"numrows",
")",
"cols",
"=",
"slice",
"(",
"0",
",",
"numcols",
")... | resolve the first ellipsis in the index so that it references the image
Parameters
----------
index : tuple
tuple of index arguments, presumably one of them is the Ellipsis
numrows, numcols, numbands : int
image dimensions
Returns
-------
tuple
Same as index, except that the first Ellipsis is replaced with
a proper slice whose start and stop members are not None | [
"resolve",
"the",
"first",
"ellipsis",
"in",
"the",
"index",
"so",
"that",
"it",
"references",
"the",
"image"
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L948-L987 |
14,429 | quintusdias/glymur | glymur/jp2k.py | Jp2k._subsampling_sanity_check | def _subsampling_sanity_check(self):
"""Check for differing subsample factors.
"""
dxs = np.array(self.codestream.segment[1].xrsiz)
dys = np.array(self.codestream.segment[1].yrsiz)
if np.any(dxs - dxs[0]) or np.any(dys - dys[0]):
msg = ("The read_bands method should be used with the subsampling "
"factors are different. "
"\n\n{siz_segment}")
msg = msg.format(siz_segment=str(self.codestream.segment[1]))
raise IOError(msg) | python | def _subsampling_sanity_check(self):
dxs = np.array(self.codestream.segment[1].xrsiz)
dys = np.array(self.codestream.segment[1].yrsiz)
if np.any(dxs - dxs[0]) or np.any(dys - dys[0]):
msg = ("The read_bands method should be used with the subsampling "
"factors are different. "
"\n\n{siz_segment}")
msg = msg.format(siz_segment=str(self.codestream.segment[1]))
raise IOError(msg) | [
"def",
"_subsampling_sanity_check",
"(",
"self",
")",
":",
"dxs",
"=",
"np",
".",
"array",
"(",
"self",
".",
"codestream",
".",
"segment",
"[",
"1",
"]",
".",
"xrsiz",
")",
"dys",
"=",
"np",
".",
"array",
"(",
"self",
".",
"codestream",
".",
"segment... | Check for differing subsample factors. | [
"Check",
"for",
"differing",
"subsample",
"factors",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1144-L1154 |
14,430 | quintusdias/glymur | glymur/jp2k.py | Jp2k._read_openjpeg | def _read_openjpeg(self, rlevel=0, verbose=False, area=None):
"""Read a JPEG 2000 image using libopenjpeg.
Parameters
----------
rlevel : int, optional
Factor by which to rlevel output resolution. Use -1 to get the
lowest resolution thumbnail.
verbose : bool, optional
Print informational messages produced by the OpenJPEG library.
area : tuple, optional
Specifies decoding image area,
(first_row, first_col, last_row, last_col)
Returns
-------
ndarray
The image data.
Raises
------
RuntimeError
If the image has differing subsample factors.
"""
self._subsampling_sanity_check()
self._populate_dparams(rlevel)
with ExitStack() as stack:
try:
self._dparams.decod_format = self._codec_format
dinfo = opj.create_decompress(self._dparams.decod_format)
event_mgr = opj.EventMgrType()
handler = ctypes.cast(_INFO_CALLBACK, ctypes.c_void_p)
event_mgr.info_handler = handler if self.verbose else None
event_mgr.warning_handler = ctypes.cast(_WARNING_CALLBACK,
ctypes.c_void_p)
event_mgr.error_handler = ctypes.cast(_ERROR_CALLBACK,
ctypes.c_void_p)
opj.set_event_mgr(dinfo, ctypes.byref(event_mgr))
opj.setup_decoder(dinfo, self._dparams)
with open(self.filename, 'rb') as fptr:
src = fptr.read()
cio = opj.cio_open(dinfo, src)
raw_image = opj.decode(dinfo, cio)
stack.callback(opj.image_destroy, raw_image)
stack.callback(opj.destroy_decompress, dinfo)
stack.callback(opj.cio_close, cio)
image = self._extract_image(raw_image)
except ValueError:
opj2.check_error(0)
if area is not None:
x0, y0, x1, y1 = area
extent = 2 ** rlevel
area = [int(round(float(x) / extent + 2 ** -20)) for x in area]
rows = slice(area[0], area[2], None)
cols = slice(area[1], area[3], None)
image = image[rows, cols]
return image | python | def _read_openjpeg(self, rlevel=0, verbose=False, area=None):
self._subsampling_sanity_check()
self._populate_dparams(rlevel)
with ExitStack() as stack:
try:
self._dparams.decod_format = self._codec_format
dinfo = opj.create_decompress(self._dparams.decod_format)
event_mgr = opj.EventMgrType()
handler = ctypes.cast(_INFO_CALLBACK, ctypes.c_void_p)
event_mgr.info_handler = handler if self.verbose else None
event_mgr.warning_handler = ctypes.cast(_WARNING_CALLBACK,
ctypes.c_void_p)
event_mgr.error_handler = ctypes.cast(_ERROR_CALLBACK,
ctypes.c_void_p)
opj.set_event_mgr(dinfo, ctypes.byref(event_mgr))
opj.setup_decoder(dinfo, self._dparams)
with open(self.filename, 'rb') as fptr:
src = fptr.read()
cio = opj.cio_open(dinfo, src)
raw_image = opj.decode(dinfo, cio)
stack.callback(opj.image_destroy, raw_image)
stack.callback(opj.destroy_decompress, dinfo)
stack.callback(opj.cio_close, cio)
image = self._extract_image(raw_image)
except ValueError:
opj2.check_error(0)
if area is not None:
x0, y0, x1, y1 = area
extent = 2 ** rlevel
area = [int(round(float(x) / extent + 2 ** -20)) for x in area]
rows = slice(area[0], area[2], None)
cols = slice(area[1], area[3], None)
image = image[rows, cols]
return image | [
"def",
"_read_openjpeg",
"(",
"self",
",",
"rlevel",
"=",
"0",
",",
"verbose",
"=",
"False",
",",
"area",
"=",
"None",
")",
":",
"self",
".",
"_subsampling_sanity_check",
"(",
")",
"self",
".",
"_populate_dparams",
"(",
"rlevel",
")",
"with",
"ExitStack",
... | Read a JPEG 2000 image using libopenjpeg.
Parameters
----------
rlevel : int, optional
Factor by which to rlevel output resolution. Use -1 to get the
lowest resolution thumbnail.
verbose : bool, optional
Print informational messages produced by the OpenJPEG library.
area : tuple, optional
Specifies decoding image area,
(first_row, first_col, last_row, last_col)
Returns
-------
ndarray
The image data.
Raises
------
RuntimeError
If the image has differing subsample factors. | [
"Read",
"a",
"JPEG",
"2000",
"image",
"using",
"libopenjpeg",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1156-L1225 |
14,431 | quintusdias/glymur | glymur/jp2k.py | Jp2k._populate_dparams | def _populate_dparams(self, rlevel, tile=None, area=None):
"""Populate decompression structure with appropriate input parameters.
Parameters
----------
rlevel : int
Factor by which to rlevel output resolution.
area : tuple
Specifies decoding image area,
(first_row, first_col, last_row, last_col)
tile : int
Number of tile to decode.
"""
if opj2.OPENJP2 is not None:
dparam = opj2.set_default_decoder_parameters()
else:
dparam = opj.DecompressionParametersType()
opj.set_default_decoder_parameters(ctypes.byref(dparam))
infile = self.filename.encode()
nelts = opj2.PATH_LEN - len(infile)
infile += b'0' * nelts
dparam.infile = infile
# Return raw codestream components instead of "interpolating" the
# colormap?
dparam.flags |= 1 if self.ignore_pclr_cmap_cdef else 0
dparam.decod_format = self._codec_format
dparam.cp_layer = self._layer
# Must check the specified rlevel against the maximum.
if rlevel != 0:
# Must check the specified rlevel against the maximum.
max_rlevel = self.codestream.segment[2].num_res
if rlevel == -1:
# -1 is shorthand for the largest rlevel
rlevel = max_rlevel
elif rlevel < -1 or rlevel > max_rlevel:
msg = ("rlevel must be in the range [-1, {max_rlevel}] "
"for this image.")
msg = msg.format(max_rlevel=max_rlevel)
raise IOError(msg)
dparam.cp_reduce = rlevel
if area is not None:
if area[0] < 0 or area[1] < 0 or area[2] <= 0 or area[3] <= 0:
msg = ("The upper left corner coordinates must be nonnegative "
"and the lower right corner coordinates must be "
"positive. The specified upper left and lower right "
"coordinates are ({y0}, {x0}) and ({y1}, {x1}).")
msg = msg.format(x0=area[1], y0=area[0],
x1=area[3], y1=area[2])
raise IOError(msg)
dparam.DA_y0 = area[0]
dparam.DA_x0 = area[1]
dparam.DA_y1 = area[2]
dparam.DA_x1 = area[3]
if tile is not None:
dparam.tile_index = tile
dparam.nb_tile_to_decode = 1
self._dparams = dparam | python | def _populate_dparams(self, rlevel, tile=None, area=None):
if opj2.OPENJP2 is not None:
dparam = opj2.set_default_decoder_parameters()
else:
dparam = opj.DecompressionParametersType()
opj.set_default_decoder_parameters(ctypes.byref(dparam))
infile = self.filename.encode()
nelts = opj2.PATH_LEN - len(infile)
infile += b'0' * nelts
dparam.infile = infile
# Return raw codestream components instead of "interpolating" the
# colormap?
dparam.flags |= 1 if self.ignore_pclr_cmap_cdef else 0
dparam.decod_format = self._codec_format
dparam.cp_layer = self._layer
# Must check the specified rlevel against the maximum.
if rlevel != 0:
# Must check the specified rlevel against the maximum.
max_rlevel = self.codestream.segment[2].num_res
if rlevel == -1:
# -1 is shorthand for the largest rlevel
rlevel = max_rlevel
elif rlevel < -1 or rlevel > max_rlevel:
msg = ("rlevel must be in the range [-1, {max_rlevel}] "
"for this image.")
msg = msg.format(max_rlevel=max_rlevel)
raise IOError(msg)
dparam.cp_reduce = rlevel
if area is not None:
if area[0] < 0 or area[1] < 0 or area[2] <= 0 or area[3] <= 0:
msg = ("The upper left corner coordinates must be nonnegative "
"and the lower right corner coordinates must be "
"positive. The specified upper left and lower right "
"coordinates are ({y0}, {x0}) and ({y1}, {x1}).")
msg = msg.format(x0=area[1], y0=area[0],
x1=area[3], y1=area[2])
raise IOError(msg)
dparam.DA_y0 = area[0]
dparam.DA_x0 = area[1]
dparam.DA_y1 = area[2]
dparam.DA_x1 = area[3]
if tile is not None:
dparam.tile_index = tile
dparam.nb_tile_to_decode = 1
self._dparams = dparam | [
"def",
"_populate_dparams",
"(",
"self",
",",
"rlevel",
",",
"tile",
"=",
"None",
",",
"area",
"=",
"None",
")",
":",
"if",
"opj2",
".",
"OPENJP2",
"is",
"not",
"None",
":",
"dparam",
"=",
"opj2",
".",
"set_default_decoder_parameters",
"(",
")",
"else",
... | Populate decompression structure with appropriate input parameters.
Parameters
----------
rlevel : int
Factor by which to rlevel output resolution.
area : tuple
Specifies decoding image area,
(first_row, first_col, last_row, last_col)
tile : int
Number of tile to decode. | [
"Populate",
"decompression",
"structure",
"with",
"appropriate",
"input",
"parameters",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1306-L1370 |
14,432 | quintusdias/glymur | glymur/jp2k.py | Jp2k._extract_image | def _extract_image(self, raw_image):
"""
Extract unequally-sized image bands.
Parameters
----------
raw_image : reference to openjpeg ImageType instance
The image structure initialized with image characteristics.
Returns
-------
list or ndarray
If the JPEG 2000 image has unequally-sized components, they are
extracted into a list, otherwise a numpy array.
"""
ncomps = raw_image.contents.numcomps
# Make a pass thru the image, see if any of the band datatypes or
# dimensions differ.
dtypes, nrows, ncols = [], [], []
for k in range(raw_image.contents.numcomps):
component = raw_image.contents.comps[k]
dtypes.append(self._component2dtype(component))
nrows.append(component.h)
ncols.append(component.w)
is_cube = all(r == nrows[0] and c == ncols[0] and d == dtypes[0]
for r, c, d in zip(nrows, ncols, dtypes))
if is_cube:
image = np.zeros((nrows[0], ncols[0], ncomps), dtypes[0])
else:
image = []
for k in range(raw_image.contents.numcomps):
component = raw_image.contents.comps[k]
self._validate_nonzero_image_size(nrows[k], ncols[k], k)
addr = ctypes.addressof(component.data.contents)
with warnings.catch_warnings():
warnings.simplefilter("ignore")
nelts = nrows[k] * ncols[k]
band = np.ctypeslib.as_array(
(ctypes.c_int32 * nelts).from_address(addr))
if is_cube:
image[:, :, k] = np.reshape(band.astype(dtypes[k]),
(nrows[k], ncols[k]))
else:
image.append(np.reshape(band.astype(dtypes[k]),
(nrows[k], ncols[k])))
if is_cube and image.shape[2] == 1:
# The third dimension has just a single layer. Make the image
# data 2D instead of 3D.
image.shape = image.shape[0:2]
return image | python | def _extract_image(self, raw_image):
ncomps = raw_image.contents.numcomps
# Make a pass thru the image, see if any of the band datatypes or
# dimensions differ.
dtypes, nrows, ncols = [], [], []
for k in range(raw_image.contents.numcomps):
component = raw_image.contents.comps[k]
dtypes.append(self._component2dtype(component))
nrows.append(component.h)
ncols.append(component.w)
is_cube = all(r == nrows[0] and c == ncols[0] and d == dtypes[0]
for r, c, d in zip(nrows, ncols, dtypes))
if is_cube:
image = np.zeros((nrows[0], ncols[0], ncomps), dtypes[0])
else:
image = []
for k in range(raw_image.contents.numcomps):
component = raw_image.contents.comps[k]
self._validate_nonzero_image_size(nrows[k], ncols[k], k)
addr = ctypes.addressof(component.data.contents)
with warnings.catch_warnings():
warnings.simplefilter("ignore")
nelts = nrows[k] * ncols[k]
band = np.ctypeslib.as_array(
(ctypes.c_int32 * nelts).from_address(addr))
if is_cube:
image[:, :, k] = np.reshape(band.astype(dtypes[k]),
(nrows[k], ncols[k]))
else:
image.append(np.reshape(band.astype(dtypes[k]),
(nrows[k], ncols[k])))
if is_cube and image.shape[2] == 1:
# The third dimension has just a single layer. Make the image
# data 2D instead of 3D.
image.shape = image.shape[0:2]
return image | [
"def",
"_extract_image",
"(",
"self",
",",
"raw_image",
")",
":",
"ncomps",
"=",
"raw_image",
".",
"contents",
".",
"numcomps",
"# Make a pass thru the image, see if any of the band datatypes or",
"# dimensions differ.",
"dtypes",
",",
"nrows",
",",
"ncols",
"=",
"[",
... | Extract unequally-sized image bands.
Parameters
----------
raw_image : reference to openjpeg ImageType instance
The image structure initialized with image characteristics.
Returns
-------
list or ndarray
If the JPEG 2000 image has unequally-sized components, they are
extracted into a list, otherwise a numpy array. | [
"Extract",
"unequally",
"-",
"sized",
"image",
"bands",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1426-L1483 |
14,433 | quintusdias/glymur | glymur/jp2k.py | Jp2k._component2dtype | def _component2dtype(self, component):
"""Determin the appropriate numpy datatype for an OpenJPEG component.
Parameters
----------
component : ctypes pointer to ImageCompType (image_comp_t)
single image component structure.
Returns
-------
builtins.type
numpy datatype to be used to construct an image array
"""
if component.prec > 16:
msg = "Unhandled precision: {0} bits.".format(component.prec)
raise IOError(msg)
if component.sgnd:
dtype = np.int8 if component.prec <=8 else np.int16
else:
dtype = np.uint8 if component.prec <=8 else np.uint16
return dtype | python | def _component2dtype(self, component):
if component.prec > 16:
msg = "Unhandled precision: {0} bits.".format(component.prec)
raise IOError(msg)
if component.sgnd:
dtype = np.int8 if component.prec <=8 else np.int16
else:
dtype = np.uint8 if component.prec <=8 else np.uint16
return dtype | [
"def",
"_component2dtype",
"(",
"self",
",",
"component",
")",
":",
"if",
"component",
".",
"prec",
">",
"16",
":",
"msg",
"=",
"\"Unhandled precision: {0} bits.\"",
".",
"format",
"(",
"component",
".",
"prec",
")",
"raise",
"IOError",
"(",
"msg",
")",
"i... | Determin the appropriate numpy datatype for an OpenJPEG component.
Parameters
----------
component : ctypes pointer to ImageCompType (image_comp_t)
single image component structure.
Returns
-------
builtins.type
numpy datatype to be used to construct an image array | [
"Determin",
"the",
"appropriate",
"numpy",
"datatype",
"for",
"an",
"OpenJPEG",
"component",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1485-L1507 |
14,434 | quintusdias/glymur | glymur/jp2k.py | Jp2k.get_codestream | def get_codestream(self, header_only=True):
"""Retrieve codestream.
Parameters
----------
header_only : bool, optional
If True, only marker segments in the main header are parsed.
Supplying False may impose a large performance penalty.
Returns
-------
Codestream
Object describing the codestream syntax.
Examples
--------
>>> import glymur
>>> jfile = glymur.data.nemo()
>>> jp2 = glymur.Jp2k(jfile)
>>> codestream = jp2.get_codestream()
>>> print(codestream.segment[1])
SIZ marker segment @ (3233, 47)
Profile: no profile
Reference Grid Height, Width: (1456 x 2592)
Vertical, Horizontal Reference Grid Offset: (0 x 0)
Reference Tile Height, Width: (1456 x 2592)
Vertical, Horizontal Reference Tile Offset: (0 x 0)
Bitdepth: (8, 8, 8)
Signed: (False, False, False)
Vertical, Horizontal Subsampling: ((1, 1), (1, 1), (1, 1))
"""
with open(self.filename, 'rb') as fptr:
if self._codec_format == opj2.CODEC_J2K:
codestream = Codestream(fptr, self.length,
header_only=header_only)
else:
box = [x for x in self.box if x.box_id == 'jp2c']
fptr.seek(box[0].offset)
read_buffer = fptr.read(8)
(box_length, _) = struct.unpack('>I4s', read_buffer)
if box_length == 0:
# The length of the box is presumed to last until the end
# of the file. Compute the effective length of the box.
box_length = os.path.getsize(fptr.name) - fptr.tell() + 8
elif box_length == 1:
# Seek past the XL field.
read_buffer = fptr.read(8)
box_length, = struct.unpack('>Q', read_buffer)
codestream = Codestream(fptr, box_length - 8,
header_only=header_only)
return codestream | python | def get_codestream(self, header_only=True):
with open(self.filename, 'rb') as fptr:
if self._codec_format == opj2.CODEC_J2K:
codestream = Codestream(fptr, self.length,
header_only=header_only)
else:
box = [x for x in self.box if x.box_id == 'jp2c']
fptr.seek(box[0].offset)
read_buffer = fptr.read(8)
(box_length, _) = struct.unpack('>I4s', read_buffer)
if box_length == 0:
# The length of the box is presumed to last until the end
# of the file. Compute the effective length of the box.
box_length = os.path.getsize(fptr.name) - fptr.tell() + 8
elif box_length == 1:
# Seek past the XL field.
read_buffer = fptr.read(8)
box_length, = struct.unpack('>Q', read_buffer)
codestream = Codestream(fptr, box_length - 8,
header_only=header_only)
return codestream | [
"def",
"get_codestream",
"(",
"self",
",",
"header_only",
"=",
"True",
")",
":",
"with",
"open",
"(",
"self",
".",
"filename",
",",
"'rb'",
")",
"as",
"fptr",
":",
"if",
"self",
".",
"_codec_format",
"==",
"opj2",
".",
"CODEC_J2K",
":",
"codestream",
"... | Retrieve codestream.
Parameters
----------
header_only : bool, optional
If True, only marker segments in the main header are parsed.
Supplying False may impose a large performance penalty.
Returns
-------
Codestream
Object describing the codestream syntax.
Examples
--------
>>> import glymur
>>> jfile = glymur.data.nemo()
>>> jp2 = glymur.Jp2k(jfile)
>>> codestream = jp2.get_codestream()
>>> print(codestream.segment[1])
SIZ marker segment @ (3233, 47)
Profile: no profile
Reference Grid Height, Width: (1456 x 2592)
Vertical, Horizontal Reference Grid Offset: (0 x 0)
Reference Tile Height, Width: (1456 x 2592)
Vertical, Horizontal Reference Tile Offset: (0 x 0)
Bitdepth: (8, 8, 8)
Signed: (False, False, False)
Vertical, Horizontal Subsampling: ((1, 1), (1, 1), (1, 1)) | [
"Retrieve",
"codestream",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1509-L1560 |
14,435 | quintusdias/glymur | glymur/jp2k.py | Jp2k._populate_image_struct | def _populate_image_struct(self, image, imgdata):
"""Populates image struct needed for compression.
Parameters
----------
image : ImageType(ctypes.Structure)
Corresponds to image_t type in openjp2 headers.
img_array : ndarray
Image data to be written to file.
"""
numrows, numcols, num_comps = imgdata.shape
for k in range(num_comps):
self._validate_nonzero_image_size(numrows, numcols, k)
# set image offset and reference grid
image.contents.x0 = self._cparams.image_offset_x0
image.contents.y0 = self._cparams.image_offset_y0
image.contents.x1 = (image.contents.x0 +
(numcols - 1) * self._cparams.subsampling_dx + 1)
image.contents.y1 = (image.contents.y0 +
(numrows - 1) * self._cparams.subsampling_dy + 1)
# Stage the image data to the openjpeg data structure.
for k in range(0, num_comps):
if self._cparams.rsiz in (core.OPJ_PROFILE_CINEMA_2K,
core.OPJ_PROFILE_CINEMA_4K):
image.contents.comps[k].prec = 12
image.contents.comps[k].bpp = 12
layer = np.ascontiguousarray(imgdata[:, :, k], dtype=np.int32)
dest = image.contents.comps[k].data
src = layer.ctypes.data
ctypes.memmove(dest, src, layer.nbytes)
return image | python | def _populate_image_struct(self, image, imgdata):
numrows, numcols, num_comps = imgdata.shape
for k in range(num_comps):
self._validate_nonzero_image_size(numrows, numcols, k)
# set image offset and reference grid
image.contents.x0 = self._cparams.image_offset_x0
image.contents.y0 = self._cparams.image_offset_y0
image.contents.x1 = (image.contents.x0 +
(numcols - 1) * self._cparams.subsampling_dx + 1)
image.contents.y1 = (image.contents.y0 +
(numrows - 1) * self._cparams.subsampling_dy + 1)
# Stage the image data to the openjpeg data structure.
for k in range(0, num_comps):
if self._cparams.rsiz in (core.OPJ_PROFILE_CINEMA_2K,
core.OPJ_PROFILE_CINEMA_4K):
image.contents.comps[k].prec = 12
image.contents.comps[k].bpp = 12
layer = np.ascontiguousarray(imgdata[:, :, k], dtype=np.int32)
dest = image.contents.comps[k].data
src = layer.ctypes.data
ctypes.memmove(dest, src, layer.nbytes)
return image | [
"def",
"_populate_image_struct",
"(",
"self",
",",
"image",
",",
"imgdata",
")",
":",
"numrows",
",",
"numcols",
",",
"num_comps",
"=",
"imgdata",
".",
"shape",
"for",
"k",
"in",
"range",
"(",
"num_comps",
")",
":",
"self",
".",
"_validate_nonzero_image_size... | Populates image struct needed for compression.
Parameters
----------
image : ImageType(ctypes.Structure)
Corresponds to image_t type in openjp2 headers.
img_array : ndarray
Image data to be written to file. | [
"Populates",
"image",
"struct",
"needed",
"for",
"compression",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1562-L1597 |
14,436 | quintusdias/glymur | glymur/jp2k.py | Jp2k._populate_comptparms | def _populate_comptparms(self, img_array):
"""Instantiate and populate comptparms structure.
This structure defines the image components.
Parameters
----------
img_array : ndarray
Image data to be written to file.
"""
# Only two precisions are possible.
if img_array.dtype == np.uint8:
comp_prec = 8
else:
comp_prec = 16
numrows, numcols, num_comps = img_array.shape
if version.openjpeg_version_tuple[0] == 1:
comptparms = (opj.ImageComptParmType * num_comps)()
else:
comptparms = (opj2.ImageComptParmType * num_comps)()
for j in range(num_comps):
comptparms[j].dx = self._cparams.subsampling_dx
comptparms[j].dy = self._cparams.subsampling_dy
comptparms[j].w = numcols
comptparms[j].h = numrows
comptparms[j].x0 = self._cparams.image_offset_x0
comptparms[j].y0 = self._cparams.image_offset_y0
comptparms[j].prec = comp_prec
comptparms[j].bpp = comp_prec
comptparms[j].sgnd = 0
self._comptparms = comptparms | python | def _populate_comptparms(self, img_array):
# Only two precisions are possible.
if img_array.dtype == np.uint8:
comp_prec = 8
else:
comp_prec = 16
numrows, numcols, num_comps = img_array.shape
if version.openjpeg_version_tuple[0] == 1:
comptparms = (opj.ImageComptParmType * num_comps)()
else:
comptparms = (opj2.ImageComptParmType * num_comps)()
for j in range(num_comps):
comptparms[j].dx = self._cparams.subsampling_dx
comptparms[j].dy = self._cparams.subsampling_dy
comptparms[j].w = numcols
comptparms[j].h = numrows
comptparms[j].x0 = self._cparams.image_offset_x0
comptparms[j].y0 = self._cparams.image_offset_y0
comptparms[j].prec = comp_prec
comptparms[j].bpp = comp_prec
comptparms[j].sgnd = 0
self._comptparms = comptparms | [
"def",
"_populate_comptparms",
"(",
"self",
",",
"img_array",
")",
":",
"# Only two precisions are possible.",
"if",
"img_array",
".",
"dtype",
"==",
"np",
".",
"uint8",
":",
"comp_prec",
"=",
"8",
"else",
":",
"comp_prec",
"=",
"16",
"numrows",
",",
"numcols"... | Instantiate and populate comptparms structure.
This structure defines the image components.
Parameters
----------
img_array : ndarray
Image data to be written to file. | [
"Instantiate",
"and",
"populate",
"comptparms",
"structure",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1599-L1631 |
14,437 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_nonzero_image_size | def _validate_nonzero_image_size(self, nrows, ncols, component_index):
"""The image cannot have area of zero.
"""
if nrows == 0 or ncols == 0:
# Letting this situation continue would segfault openjpeg.
msg = "Component {0} has dimensions {1} x {2}"
msg = msg.format(component_index, nrows, ncols)
raise IOError(msg) | python | def _validate_nonzero_image_size(self, nrows, ncols, component_index):
if nrows == 0 or ncols == 0:
# Letting this situation continue would segfault openjpeg.
msg = "Component {0} has dimensions {1} x {2}"
msg = msg.format(component_index, nrows, ncols)
raise IOError(msg) | [
"def",
"_validate_nonzero_image_size",
"(",
"self",
",",
"nrows",
",",
"ncols",
",",
"component_index",
")",
":",
"if",
"nrows",
"==",
"0",
"or",
"ncols",
"==",
"0",
":",
"# Letting this situation continue would segfault openjpeg.",
"msg",
"=",
"\"Component {0} has di... | The image cannot have area of zero. | [
"The",
"image",
"cannot",
"have",
"area",
"of",
"zero",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1633-L1640 |
14,438 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_jp2_box_sequence | def _validate_jp2_box_sequence(self, boxes):
"""Run through series of tests for JP2 box legality.
This is non-exhaustive.
"""
JP2_IDS = ['colr', 'cdef', 'cmap', 'jp2c', 'ftyp', 'ihdr', 'jp2h',
'jP ', 'pclr', 'res ', 'resc', 'resd', 'xml ', 'ulst',
'uinf', 'url ', 'uuid']
self._validate_signature_compatibility(boxes)
self._validate_jp2h(boxes)
self._validate_jp2c(boxes)
if boxes[1].brand == 'jpx ':
self._validate_jpx_box_sequence(boxes)
else:
# Validate the JP2 box IDs.
count = self._collect_box_count(boxes)
for box_id in count.keys():
if box_id not in JP2_IDS:
msg = ("The presence of a '{0}' box requires that the "
"file type brand be set to 'jpx '.")
raise IOError(msg.format(box_id))
self._validate_jp2_colr(boxes) | python | def _validate_jp2_box_sequence(self, boxes):
JP2_IDS = ['colr', 'cdef', 'cmap', 'jp2c', 'ftyp', 'ihdr', 'jp2h',
'jP ', 'pclr', 'res ', 'resc', 'resd', 'xml ', 'ulst',
'uinf', 'url ', 'uuid']
self._validate_signature_compatibility(boxes)
self._validate_jp2h(boxes)
self._validate_jp2c(boxes)
if boxes[1].brand == 'jpx ':
self._validate_jpx_box_sequence(boxes)
else:
# Validate the JP2 box IDs.
count = self._collect_box_count(boxes)
for box_id in count.keys():
if box_id not in JP2_IDS:
msg = ("The presence of a '{0}' box requires that the "
"file type brand be set to 'jpx '.")
raise IOError(msg.format(box_id))
self._validate_jp2_colr(boxes) | [
"def",
"_validate_jp2_box_sequence",
"(",
"self",
",",
"boxes",
")",
":",
"JP2_IDS",
"=",
"[",
"'colr'",
",",
"'cdef'",
",",
"'cmap'",
",",
"'jp2c'",
",",
"'ftyp'",
",",
"'ihdr'",
",",
"'jp2h'",
",",
"'jP '",
",",
"'pclr'",
",",
"'res '",
",",
"'resc'",... | Run through series of tests for JP2 box legality.
This is non-exhaustive. | [
"Run",
"through",
"series",
"of",
"tests",
"for",
"JP2",
"box",
"legality",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1642-L1665 |
14,439 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_jp2_colr | def _validate_jp2_colr(self, boxes):
"""
Validate JP2 requirements on colour specification boxes.
"""
lst = [box for box in boxes if box.box_id == 'jp2h']
jp2h = lst[0]
for colr in [box for box in jp2h.box if box.box_id == 'colr']:
if colr.approximation != 0:
msg = ("A JP2 colr box cannot have a non-zero approximation "
"field.")
raise IOError(msg) | python | def _validate_jp2_colr(self, boxes):
lst = [box for box in boxes if box.box_id == 'jp2h']
jp2h = lst[0]
for colr in [box for box in jp2h.box if box.box_id == 'colr']:
if colr.approximation != 0:
msg = ("A JP2 colr box cannot have a non-zero approximation "
"field.")
raise IOError(msg) | [
"def",
"_validate_jp2_colr",
"(",
"self",
",",
"boxes",
")",
":",
"lst",
"=",
"[",
"box",
"for",
"box",
"in",
"boxes",
"if",
"box",
".",
"box_id",
"==",
"'jp2h'",
"]",
"jp2h",
"=",
"lst",
"[",
"0",
"]",
"for",
"colr",
"in",
"[",
"box",
"for",
"bo... | Validate JP2 requirements on colour specification boxes. | [
"Validate",
"JP2",
"requirements",
"on",
"colour",
"specification",
"boxes",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1667-L1677 |
14,440 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_jpx_box_sequence | def _validate_jpx_box_sequence(self, boxes):
"""Run through series of tests for JPX box legality."""
self._validate_label(boxes)
self._validate_jpx_compatibility(boxes, boxes[1].compatibility_list)
self._validate_singletons(boxes)
self._validate_top_level(boxes) | python | def _validate_jpx_box_sequence(self, boxes):
self._validate_label(boxes)
self._validate_jpx_compatibility(boxes, boxes[1].compatibility_list)
self._validate_singletons(boxes)
self._validate_top_level(boxes) | [
"def",
"_validate_jpx_box_sequence",
"(",
"self",
",",
"boxes",
")",
":",
"self",
".",
"_validate_label",
"(",
"boxes",
")",
"self",
".",
"_validate_jpx_compatibility",
"(",
"boxes",
",",
"boxes",
"[",
"1",
"]",
".",
"compatibility_list",
")",
"self",
".",
"... | Run through series of tests for JPX box legality. | [
"Run",
"through",
"series",
"of",
"tests",
"for",
"JPX",
"box",
"legality",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1679-L1684 |
14,441 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_signature_compatibility | def _validate_signature_compatibility(self, boxes):
"""Validate the file signature and compatibility status."""
# Check for a bad sequence of boxes.
# 1st two boxes must be 'jP ' and 'ftyp'
if boxes[0].box_id != 'jP ' or boxes[1].box_id != 'ftyp':
msg = ("The first box must be the signature box and the second "
"must be the file type box.")
raise IOError(msg)
# The compatibility list must contain at a minimum 'jp2 '.
if 'jp2 ' not in boxes[1].compatibility_list:
msg = "The ftyp box must contain 'jp2 ' in the compatibility list."
raise IOError(msg) | python | def _validate_signature_compatibility(self, boxes):
# Check for a bad sequence of boxes.
# 1st two boxes must be 'jP ' and 'ftyp'
if boxes[0].box_id != 'jP ' or boxes[1].box_id != 'ftyp':
msg = ("The first box must be the signature box and the second "
"must be the file type box.")
raise IOError(msg)
# The compatibility list must contain at a minimum 'jp2 '.
if 'jp2 ' not in boxes[1].compatibility_list:
msg = "The ftyp box must contain 'jp2 ' in the compatibility list."
raise IOError(msg) | [
"def",
"_validate_signature_compatibility",
"(",
"self",
",",
"boxes",
")",
":",
"# Check for a bad sequence of boxes.",
"# 1st two boxes must be 'jP ' and 'ftyp'",
"if",
"boxes",
"[",
"0",
"]",
".",
"box_id",
"!=",
"'jP '",
"or",
"boxes",
"[",
"1",
"]",
".",
"box... | Validate the file signature and compatibility status. | [
"Validate",
"the",
"file",
"signature",
"and",
"compatibility",
"status",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1686-L1698 |
14,442 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_jp2c | def _validate_jp2c(self, boxes):
"""Validate the codestream box in relation to other boxes."""
# jp2c must be preceeded by jp2h
jp2h_lst = [idx for (idx, box) in enumerate(boxes)
if box.box_id == 'jp2h']
jp2h_idx = jp2h_lst[0]
jp2c_lst = [idx for (idx, box) in enumerate(boxes)
if box.box_id == 'jp2c']
if len(jp2c_lst) == 0:
msg = ("A codestream box must be defined in the outermost "
"list of boxes.")
raise IOError(msg)
jp2c_idx = jp2c_lst[0]
if jp2h_idx >= jp2c_idx:
msg = "The codestream box must be preceeded by a jp2 header box."
raise IOError(msg) | python | def _validate_jp2c(self, boxes):
# jp2c must be preceeded by jp2h
jp2h_lst = [idx for (idx, box) in enumerate(boxes)
if box.box_id == 'jp2h']
jp2h_idx = jp2h_lst[0]
jp2c_lst = [idx for (idx, box) in enumerate(boxes)
if box.box_id == 'jp2c']
if len(jp2c_lst) == 0:
msg = ("A codestream box must be defined in the outermost "
"list of boxes.")
raise IOError(msg)
jp2c_idx = jp2c_lst[0]
if jp2h_idx >= jp2c_idx:
msg = "The codestream box must be preceeded by a jp2 header box."
raise IOError(msg) | [
"def",
"_validate_jp2c",
"(",
"self",
",",
"boxes",
")",
":",
"# jp2c must be preceeded by jp2h",
"jp2h_lst",
"=",
"[",
"idx",
"for",
"(",
"idx",
",",
"box",
")",
"in",
"enumerate",
"(",
"boxes",
")",
"if",
"box",
".",
"box_id",
"==",
"'jp2h'",
"]",
"jp2... | Validate the codestream box in relation to other boxes. | [
"Validate",
"the",
"codestream",
"box",
"in",
"relation",
"to",
"other",
"boxes",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1700-L1716 |
14,443 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_jp2h | def _validate_jp2h(self, boxes):
"""Validate the JP2 Header box."""
self._check_jp2h_child_boxes(boxes, 'top-level')
jp2h_lst = [box for box in boxes if box.box_id == 'jp2h']
jp2h = jp2h_lst[0]
# 1st jp2 header box cannot be empty.
if len(jp2h.box) == 0:
msg = "The JP2 header superbox cannot be empty."
raise IOError(msg)
# 1st jp2 header box must be ihdr
if jp2h.box[0].box_id != 'ihdr':
msg = ("The first box in the jp2 header box must be the image "
"header box.")
raise IOError(msg)
# colr must be present in jp2 header box.
colr_lst = [j for (j, box) in enumerate(jp2h.box)
if box.box_id == 'colr']
if len(colr_lst) == 0:
msg = "The jp2 header box must contain a color definition box."
raise IOError(msg)
colr = jp2h.box[colr_lst[0]]
self._validate_channel_definition(jp2h, colr) | python | def _validate_jp2h(self, boxes):
self._check_jp2h_child_boxes(boxes, 'top-level')
jp2h_lst = [box for box in boxes if box.box_id == 'jp2h']
jp2h = jp2h_lst[0]
# 1st jp2 header box cannot be empty.
if len(jp2h.box) == 0:
msg = "The JP2 header superbox cannot be empty."
raise IOError(msg)
# 1st jp2 header box must be ihdr
if jp2h.box[0].box_id != 'ihdr':
msg = ("The first box in the jp2 header box must be the image "
"header box.")
raise IOError(msg)
# colr must be present in jp2 header box.
colr_lst = [j for (j, box) in enumerate(jp2h.box)
if box.box_id == 'colr']
if len(colr_lst) == 0:
msg = "The jp2 header box must contain a color definition box."
raise IOError(msg)
colr = jp2h.box[colr_lst[0]]
self._validate_channel_definition(jp2h, colr) | [
"def",
"_validate_jp2h",
"(",
"self",
",",
"boxes",
")",
":",
"self",
".",
"_check_jp2h_child_boxes",
"(",
"boxes",
",",
"'top-level'",
")",
"jp2h_lst",
"=",
"[",
"box",
"for",
"box",
"in",
"boxes",
"if",
"box",
".",
"box_id",
"==",
"'jp2h'",
"]",
"jp2h"... | Validate the JP2 Header box. | [
"Validate",
"the",
"JP2",
"Header",
"box",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1718-L1744 |
14,444 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_channel_definition | def _validate_channel_definition(self, jp2h, colr):
"""Validate the channel definition box."""
cdef_lst = [j for (j, box) in enumerate(jp2h.box)
if box.box_id == 'cdef']
if len(cdef_lst) > 1:
msg = ("Only one channel definition box is allowed in the "
"JP2 header.")
raise IOError(msg)
elif len(cdef_lst) == 1:
cdef = jp2h.box[cdef_lst[0]]
if colr.colorspace == core.SRGB:
if any([chan + 1 not in cdef.association or
cdef.channel_type[chan] != 0 for chan in [0, 1, 2]]):
msg = ("All color channels must be defined in the "
"channel definition box.")
raise IOError(msg)
elif colr.colorspace == core.GREYSCALE:
if 0 not in cdef.channel_type:
msg = ("All color channels must be defined in the "
"channel definition box.")
raise IOError(msg) | python | def _validate_channel_definition(self, jp2h, colr):
cdef_lst = [j for (j, box) in enumerate(jp2h.box)
if box.box_id == 'cdef']
if len(cdef_lst) > 1:
msg = ("Only one channel definition box is allowed in the "
"JP2 header.")
raise IOError(msg)
elif len(cdef_lst) == 1:
cdef = jp2h.box[cdef_lst[0]]
if colr.colorspace == core.SRGB:
if any([chan + 1 not in cdef.association or
cdef.channel_type[chan] != 0 for chan in [0, 1, 2]]):
msg = ("All color channels must be defined in the "
"channel definition box.")
raise IOError(msg)
elif colr.colorspace == core.GREYSCALE:
if 0 not in cdef.channel_type:
msg = ("All color channels must be defined in the "
"channel definition box.")
raise IOError(msg) | [
"def",
"_validate_channel_definition",
"(",
"self",
",",
"jp2h",
",",
"colr",
")",
":",
"cdef_lst",
"=",
"[",
"j",
"for",
"(",
"j",
",",
"box",
")",
"in",
"enumerate",
"(",
"jp2h",
".",
"box",
")",
"if",
"box",
".",
"box_id",
"==",
"'cdef'",
"]",
"... | Validate the channel definition box. | [
"Validate",
"the",
"channel",
"definition",
"box",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1746-L1766 |
14,445 | quintusdias/glymur | glymur/jp2k.py | Jp2k._check_jp2h_child_boxes | def _check_jp2h_child_boxes(self, boxes, parent_box_name):
"""Certain boxes can only reside in the JP2 header."""
JP2H_CHILDREN = set(['bpcc', 'cdef', 'cmap', 'ihdr', 'pclr'])
box_ids = set([box.box_id for box in boxes])
intersection = box_ids.intersection(JP2H_CHILDREN)
if len(intersection) > 0 and parent_box_name not in ['jp2h', 'jpch']:
msg = "A {0} box can only be nested in a JP2 header box."
raise IOError(msg.format(list(intersection)[0]))
# Recursively check any contained superboxes.
for box in boxes:
if hasattr(box, 'box'):
self._check_jp2h_child_boxes(box.box, box.box_id) | python | def _check_jp2h_child_boxes(self, boxes, parent_box_name):
JP2H_CHILDREN = set(['bpcc', 'cdef', 'cmap', 'ihdr', 'pclr'])
box_ids = set([box.box_id for box in boxes])
intersection = box_ids.intersection(JP2H_CHILDREN)
if len(intersection) > 0 and parent_box_name not in ['jp2h', 'jpch']:
msg = "A {0} box can only be nested in a JP2 header box."
raise IOError(msg.format(list(intersection)[0]))
# Recursively check any contained superboxes.
for box in boxes:
if hasattr(box, 'box'):
self._check_jp2h_child_boxes(box.box, box.box_id) | [
"def",
"_check_jp2h_child_boxes",
"(",
"self",
",",
"boxes",
",",
"parent_box_name",
")",
":",
"JP2H_CHILDREN",
"=",
"set",
"(",
"[",
"'bpcc'",
",",
"'cdef'",
",",
"'cmap'",
",",
"'ihdr'",
",",
"'pclr'",
"]",
")",
"box_ids",
"=",
"set",
"(",
"[",
"box",
... | Certain boxes can only reside in the JP2 header. | [
"Certain",
"boxes",
"can",
"only",
"reside",
"in",
"the",
"JP2",
"header",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1768-L1781 |
14,446 | quintusdias/glymur | glymur/jp2k.py | Jp2k._collect_box_count | def _collect_box_count(self, boxes):
"""Count the occurences of each box type."""
count = Counter([box.box_id for box in boxes])
# Add the counts in the superboxes.
for box in boxes:
if hasattr(box, 'box'):
count.update(self._collect_box_count(box.box))
return count | python | def _collect_box_count(self, boxes):
count = Counter([box.box_id for box in boxes])
# Add the counts in the superboxes.
for box in boxes:
if hasattr(box, 'box'):
count.update(self._collect_box_count(box.box))
return count | [
"def",
"_collect_box_count",
"(",
"self",
",",
"boxes",
")",
":",
"count",
"=",
"Counter",
"(",
"[",
"box",
".",
"box_id",
"for",
"box",
"in",
"boxes",
"]",
")",
"# Add the counts in the superboxes.",
"for",
"box",
"in",
"boxes",
":",
"if",
"hasattr",
"(",... | Count the occurences of each box type. | [
"Count",
"the",
"occurences",
"of",
"each",
"box",
"type",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1783-L1792 |
14,447 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_singletons | def _validate_singletons(self, boxes):
"""Several boxes can only occur once."""
count = self._collect_box_count(boxes)
# Which boxes occur more than once?
multiples = [box_id for box_id, bcount in count.items() if bcount > 1]
if 'dtbl' in multiples:
raise IOError('There can only be one dtbl box in a file.') | python | def _validate_singletons(self, boxes):
count = self._collect_box_count(boxes)
# Which boxes occur more than once?
multiples = [box_id for box_id, bcount in count.items() if bcount > 1]
if 'dtbl' in multiples:
raise IOError('There can only be one dtbl box in a file.') | [
"def",
"_validate_singletons",
"(",
"self",
",",
"boxes",
")",
":",
"count",
"=",
"self",
".",
"_collect_box_count",
"(",
"boxes",
")",
"# Which boxes occur more than once?",
"multiples",
"=",
"[",
"box_id",
"for",
"box_id",
",",
"bcount",
"in",
"count",
".",
... | Several boxes can only occur once. | [
"Several",
"boxes",
"can",
"only",
"occur",
"once",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1825-L1831 |
14,448 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_jpx_compatibility | def _validate_jpx_compatibility(self, boxes, compatibility_list):
"""
If there is a JPX box then the compatibility list must also contain
'jpx '.
"""
JPX_IDS = ['asoc', 'nlst']
jpx_cl = set(compatibility_list)
for box in boxes:
if box.box_id in JPX_IDS:
if len(set(['jpx ', 'jpxb']).intersection(jpx_cl)) == 0:
msg = ("A JPX box requires that either 'jpx ' or 'jpxb' "
"be present in the ftype compatibility list.")
raise RuntimeError(msg)
if hasattr(box, 'box') != 0:
# Same set of checks on any child boxes.
self._validate_jpx_compatibility(box.box, compatibility_list) | python | def _validate_jpx_compatibility(self, boxes, compatibility_list):
JPX_IDS = ['asoc', 'nlst']
jpx_cl = set(compatibility_list)
for box in boxes:
if box.box_id in JPX_IDS:
if len(set(['jpx ', 'jpxb']).intersection(jpx_cl)) == 0:
msg = ("A JPX box requires that either 'jpx ' or 'jpxb' "
"be present in the ftype compatibility list.")
raise RuntimeError(msg)
if hasattr(box, 'box') != 0:
# Same set of checks on any child boxes.
self._validate_jpx_compatibility(box.box, compatibility_list) | [
"def",
"_validate_jpx_compatibility",
"(",
"self",
",",
"boxes",
",",
"compatibility_list",
")",
":",
"JPX_IDS",
"=",
"[",
"'asoc'",
",",
"'nlst'",
"]",
"jpx_cl",
"=",
"set",
"(",
"compatibility_list",
")",
"for",
"box",
"in",
"boxes",
":",
"if",
"box",
".... | If there is a JPX box then the compatibility list must also contain
'jpx '. | [
"If",
"there",
"is",
"a",
"JPX",
"box",
"then",
"the",
"compatibility",
"list",
"must",
"also",
"contain",
"jpx",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1833-L1848 |
14,449 | quintusdias/glymur | glymur/jp2k.py | Jp2k._validate_label | def _validate_label(self, boxes):
"""
Label boxes can only be inside association, codestream headers, or
compositing layer header boxes.
"""
for box in boxes:
if box.box_id != 'asoc':
if hasattr(box, 'box'):
for boxi in box.box:
if boxi.box_id == 'lbl ':
msg = ("A label box cannot be nested inside a "
"{0} box.")
msg = msg.format(box.box_id)
raise IOError(msg)
# Same set of checks on any child boxes.
self._validate_label(box.box) | python | def _validate_label(self, boxes):
for box in boxes:
if box.box_id != 'asoc':
if hasattr(box, 'box'):
for boxi in box.box:
if boxi.box_id == 'lbl ':
msg = ("A label box cannot be nested inside a "
"{0} box.")
msg = msg.format(box.box_id)
raise IOError(msg)
# Same set of checks on any child boxes.
self._validate_label(box.box) | [
"def",
"_validate_label",
"(",
"self",
",",
"boxes",
")",
":",
"for",
"box",
"in",
"boxes",
":",
"if",
"box",
".",
"box_id",
"!=",
"'asoc'",
":",
"if",
"hasattr",
"(",
"box",
",",
"'box'",
")",
":",
"for",
"boxi",
"in",
"box",
".",
"box",
":",
"i... | Label boxes can only be inside association, codestream headers, or
compositing layer header boxes. | [
"Label",
"boxes",
"can",
"only",
"be",
"inside",
"association",
"codestream",
"headers",
"or",
"compositing",
"layer",
"header",
"boxes",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/jp2k.py#L1850-L1865 |
14,450 | keenlabs/KeenClient-Python | keen/api.py | KeenApi.fulfill | def fulfill(self, method, *args, **kwargs):
""" Fulfill an HTTP request to Keen's API. """
return getattr(self.session, method)(*args, **kwargs) | python | def fulfill(self, method, *args, **kwargs):
return getattr(self.session, method)(*args, **kwargs) | [
"def",
"fulfill",
"(",
"self",
",",
"method",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"getattr",
"(",
"self",
".",
"session",
",",
"method",
")",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] | Fulfill an HTTP request to Keen's API. | [
"Fulfill",
"an",
"HTTP",
"request",
"to",
"Keen",
"s",
"API",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L90-L94 |
14,451 | keenlabs/KeenClient-Python | keen/api.py | KeenApi._order_by_is_valid_or_none | def _order_by_is_valid_or_none(self, params):
"""
Validates that a given order_by has proper syntax.
:param params: Query params.
:return: Returns True if either no order_by is present, or if the order_by is well-formed.
"""
if not "order_by" in params or not params["order_by"]:
return True
def _order_by_dict_is_not_well_formed(d):
if not isinstance(d, dict):
# Bad type.
return True
if "property_name" in d and d["property_name"]:
if "direction" in d and not direction.is_valid_direction(d["direction"]):
# Bad direction provided.
return True
for k in d:
if k != "property_name" and k != "direction":
# Unexpected key.
return True
# Everything looks good!
return False
# Missing required key.
return True
# order_by is converted to a list before this point if it wasn't one before.
order_by_list = json.loads(params["order_by"])
for order_by in order_by_list:
if _order_by_dict_is_not_well_formed(order_by):
return False
if not "group_by" in params or not params["group_by"]:
# We must have group_by to have order_by make sense.
return False
return True | python | def _order_by_is_valid_or_none(self, params):
if not "order_by" in params or not params["order_by"]:
return True
def _order_by_dict_is_not_well_formed(d):
if not isinstance(d, dict):
# Bad type.
return True
if "property_name" in d and d["property_name"]:
if "direction" in d and not direction.is_valid_direction(d["direction"]):
# Bad direction provided.
return True
for k in d:
if k != "property_name" and k != "direction":
# Unexpected key.
return True
# Everything looks good!
return False
# Missing required key.
return True
# order_by is converted to a list before this point if it wasn't one before.
order_by_list = json.loads(params["order_by"])
for order_by in order_by_list:
if _order_by_dict_is_not_well_formed(order_by):
return False
if not "group_by" in params or not params["group_by"]:
# We must have group_by to have order_by make sense.
return False
return True | [
"def",
"_order_by_is_valid_or_none",
"(",
"self",
",",
"params",
")",
":",
"if",
"not",
"\"order_by\"",
"in",
"params",
"or",
"not",
"params",
"[",
"\"order_by\"",
"]",
":",
"return",
"True",
"def",
"_order_by_dict_is_not_well_formed",
"(",
"d",
")",
":",
"if"... | Validates that a given order_by has proper syntax.
:param params: Query params.
:return: Returns True if either no order_by is present, or if the order_by is well-formed. | [
"Validates",
"that",
"a",
"given",
"order_by",
"has",
"proper",
"syntax",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L130-L166 |
14,452 | keenlabs/KeenClient-Python | keen/api.py | KeenApi._limit_is_valid_or_none | def _limit_is_valid_or_none(self, params):
"""
Validates that a given limit is not present or is well-formed.
:param params: Query params.
:return: Returns True if a limit is present or is well-formed.
"""
if not "limit" in params or not params["limit"]:
return True
if not isinstance(params["limit"], int) or params["limit"] < 1:
return False
if not "order_by" in params:
return False
return True | python | def _limit_is_valid_or_none(self, params):
if not "limit" in params or not params["limit"]:
return True
if not isinstance(params["limit"], int) or params["limit"] < 1:
return False
if not "order_by" in params:
return False
return True | [
"def",
"_limit_is_valid_or_none",
"(",
"self",
",",
"params",
")",
":",
"if",
"not",
"\"limit\"",
"in",
"params",
"or",
"not",
"params",
"[",
"\"limit\"",
"]",
":",
"return",
"True",
"if",
"not",
"isinstance",
"(",
"params",
"[",
"\"limit\"",
"]",
",",
"... | Validates that a given limit is not present or is well-formed.
:param params: Query params.
:return: Returns True if a limit is present or is well-formed. | [
"Validates",
"that",
"a",
"given",
"limit",
"is",
"not",
"present",
"or",
"is",
"well",
"-",
"formed",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L168-L181 |
14,453 | keenlabs/KeenClient-Python | keen/api.py | KeenApi.query | def query(self, analysis_type, params, all_keys=False):
"""
Performs a query using the Keen IO analysis API. A read key must be set first.
"""
if not self._order_by_is_valid_or_none(params):
raise ValueError("order_by given is invalid or is missing required group_by.")
if not self._limit_is_valid_or_none(params):
raise ValueError("limit given is invalid or is missing required order_by.")
url = "{0}/{1}/projects/{2}/queries/{3}".format(self.base_url, self.api_version,
self.project_id, analysis_type)
headers = utilities.headers(self.read_key)
payload = params
response = self.fulfill(HTTPMethods.GET, url, params=payload, headers=headers, timeout=self.get_timeout)
self._error_handling(response)
response = response.json()
if not all_keys:
response = response["result"]
return response | python | def query(self, analysis_type, params, all_keys=False):
if not self._order_by_is_valid_or_none(params):
raise ValueError("order_by given is invalid or is missing required group_by.")
if not self._limit_is_valid_or_none(params):
raise ValueError("limit given is invalid or is missing required order_by.")
url = "{0}/{1}/projects/{2}/queries/{3}".format(self.base_url, self.api_version,
self.project_id, analysis_type)
headers = utilities.headers(self.read_key)
payload = params
response = self.fulfill(HTTPMethods.GET, url, params=payload, headers=headers, timeout=self.get_timeout)
self._error_handling(response)
response = response.json()
if not all_keys:
response = response["result"]
return response | [
"def",
"query",
"(",
"self",
",",
"analysis_type",
",",
"params",
",",
"all_keys",
"=",
"False",
")",
":",
"if",
"not",
"self",
".",
"_order_by_is_valid_or_none",
"(",
"params",
")",
":",
"raise",
"ValueError",
"(",
"\"order_by given is invalid or is missing requi... | Performs a query using the Keen IO analysis API. A read key must be set first. | [
"Performs",
"a",
"query",
"using",
"the",
"Keen",
"IO",
"analysis",
"API",
".",
"A",
"read",
"key",
"must",
"be",
"set",
"first",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L184-L207 |
14,454 | keenlabs/KeenClient-Python | keen/api.py | KeenApi.delete_events | def delete_events(self, event_collection, params):
"""
Deletes events via the Keen IO API. A master key must be set first.
:param event_collection: string, the event collection from which event are being deleted
"""
url = "{0}/{1}/projects/{2}/events/{3}".format(self.base_url,
self.api_version,
self.project_id,
event_collection)
headers = utilities.headers(self.master_key)
response = self.fulfill(HTTPMethods.DELETE, url, params=params, headers=headers, timeout=self.post_timeout)
self._error_handling(response)
return True | python | def delete_events(self, event_collection, params):
url = "{0}/{1}/projects/{2}/events/{3}".format(self.base_url,
self.api_version,
self.project_id,
event_collection)
headers = utilities.headers(self.master_key)
response = self.fulfill(HTTPMethods.DELETE, url, params=params, headers=headers, timeout=self.post_timeout)
self._error_handling(response)
return True | [
"def",
"delete_events",
"(",
"self",
",",
"event_collection",
",",
"params",
")",
":",
"url",
"=",
"\"{0}/{1}/projects/{2}/events/{3}\"",
".",
"format",
"(",
"self",
".",
"base_url",
",",
"self",
".",
"api_version",
",",
"self",
".",
"project_id",
",",
"event_... | Deletes events via the Keen IO API. A master key must be set first.
:param event_collection: string, the event collection from which event are being deleted | [
"Deletes",
"events",
"via",
"the",
"Keen",
"IO",
"API",
".",
"A",
"master",
"key",
"must",
"be",
"set",
"first",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L210-L226 |
14,455 | keenlabs/KeenClient-Python | keen/api.py | KeenApi.get_collection | def get_collection(self, event_collection):
"""
Extracts info about a collection using the Keen IO API. A master key must be set first.
:param event_collection: the name of the collection to retrieve info for
"""
url = "{0}/{1}/projects/{2}/events/{3}".format(self.base_url, self.api_version,
self.project_id, event_collection)
headers = utilities.headers(self.read_key)
response = self.fulfill(HTTPMethods.GET, url, headers=headers, timeout=self.get_timeout)
self._error_handling(response)
return response.json() | python | def get_collection(self, event_collection):
url = "{0}/{1}/projects/{2}/events/{3}".format(self.base_url, self.api_version,
self.project_id, event_collection)
headers = utilities.headers(self.read_key)
response = self.fulfill(HTTPMethods.GET, url, headers=headers, timeout=self.get_timeout)
self._error_handling(response)
return response.json() | [
"def",
"get_collection",
"(",
"self",
",",
"event_collection",
")",
":",
"url",
"=",
"\"{0}/{1}/projects/{2}/events/{3}\"",
".",
"format",
"(",
"self",
".",
"base_url",
",",
"self",
".",
"api_version",
",",
"self",
".",
"project_id",
",",
"event_collection",
")"... | Extracts info about a collection using the Keen IO API. A master key must be set first.
:param event_collection: the name of the collection to retrieve info for | [
"Extracts",
"info",
"about",
"a",
"collection",
"using",
"the",
"Keen",
"IO",
"API",
".",
"A",
"master",
"key",
"must",
"be",
"set",
"first",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L229-L242 |
14,456 | keenlabs/KeenClient-Python | keen/api.py | KeenApi._update_access_key_pair | def _update_access_key_pair(self, access_key_id, key, val):
"""
Helper for updating access keys in a DRY fashion.
"""
# Get current state via HTTPS.
current_access_key = self.get_access_key(access_key_id)
# Copy and only change the single parameter.
payload_dict = KeenApi._build_access_key_dict(current_access_key)
payload_dict[key] = val
# Now just treat it like a full update.
return self.update_access_key_full(access_key_id, **payload_dict) | python | def _update_access_key_pair(self, access_key_id, key, val):
# Get current state via HTTPS.
current_access_key = self.get_access_key(access_key_id)
# Copy and only change the single parameter.
payload_dict = KeenApi._build_access_key_dict(current_access_key)
payload_dict[key] = val
# Now just treat it like a full update.
return self.update_access_key_full(access_key_id, **payload_dict) | [
"def",
"_update_access_key_pair",
"(",
"self",
",",
"access_key_id",
",",
"key",
",",
"val",
")",
":",
"# Get current state via HTTPS.",
"current_access_key",
"=",
"self",
".",
"get_access_key",
"(",
"access_key_id",
")",
"# Copy and only change the single parameter.",
"p... | Helper for updating access keys in a DRY fashion. | [
"Helper",
"for",
"updating",
"access",
"keys",
"in",
"a",
"DRY",
"fashion",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L328-L340 |
14,457 | keenlabs/KeenClient-Python | keen/api.py | KeenApi.add_access_key_permissions | def add_access_key_permissions(self, access_key_id, permissions):
"""
Adds to the existing list of permissions on this key with the contents of this list.
Will not remove any existing permissions or modify the remainder of the key.
:param access_key_id: the 'key' value of the access key to add permissions to
:param permissions: the new permissions to add to the existing list of permissions
"""
# Get current state via HTTPS.
current_access_key = self.get_access_key(access_key_id)
# Copy and only change the single parameter.
payload_dict = KeenApi._build_access_key_dict(current_access_key)
# Turn into sets to avoid duplicates.
old_permissions = set(payload_dict["permitted"])
new_permissions = set(permissions)
combined_permissions = old_permissions.union(new_permissions)
payload_dict["permitted"] = list(combined_permissions)
# Now just treat it like a full update.
return self.update_access_key_full(access_key_id, **payload_dict) | python | def add_access_key_permissions(self, access_key_id, permissions):
# Get current state via HTTPS.
current_access_key = self.get_access_key(access_key_id)
# Copy and only change the single parameter.
payload_dict = KeenApi._build_access_key_dict(current_access_key)
# Turn into sets to avoid duplicates.
old_permissions = set(payload_dict["permitted"])
new_permissions = set(permissions)
combined_permissions = old_permissions.union(new_permissions)
payload_dict["permitted"] = list(combined_permissions)
# Now just treat it like a full update.
return self.update_access_key_full(access_key_id, **payload_dict) | [
"def",
"add_access_key_permissions",
"(",
"self",
",",
"access_key_id",
",",
"permissions",
")",
":",
"# Get current state via HTTPS.",
"current_access_key",
"=",
"self",
".",
"get_access_key",
"(",
"access_key_id",
")",
"# Copy and only change the single parameter.",
"payloa... | Adds to the existing list of permissions on this key with the contents of this list.
Will not remove any existing permissions or modify the remainder of the key.
:param access_key_id: the 'key' value of the access key to add permissions to
:param permissions: the new permissions to add to the existing list of permissions | [
"Adds",
"to",
"the",
"existing",
"list",
"of",
"permissions",
"on",
"this",
"key",
"with",
"the",
"contents",
"of",
"this",
"list",
".",
"Will",
"not",
"remove",
"any",
"existing",
"permissions",
"or",
"modify",
"the",
"remainder",
"of",
"the",
"key",
"."
... | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L353-L374 |
14,458 | keenlabs/KeenClient-Python | keen/api.py | KeenApi.remove_access_key_permissions | def remove_access_key_permissions(self, access_key_id, permissions):
"""
Removes a list of permissions from the existing list of permissions.
Will not remove all existing permissions unless all such permissions are included
in this list. Not to be confused with key revocation.
See also: revoke_access_key()
:param access_key_id: the 'key' value of the access key to remove some permissions from
:param permissions: the permissions you wish to remove from this access key
"""
# Get current state via HTTPS.
current_access_key = self.get_access_key(access_key_id)
# Copy and only change the single parameter.
payload_dict = KeenApi._build_access_key_dict(current_access_key)
# Turn into sets to avoid duplicates.
old_permissions = set(payload_dict["permitted"])
removal_permissions = set(permissions)
reduced_permissions = old_permissions.difference(removal_permissions)
payload_dict["permitted"] = list(reduced_permissions)
# Now just treat it like a full update.
return self.update_access_key_full(access_key_id, **payload_dict) | python | def remove_access_key_permissions(self, access_key_id, permissions):
# Get current state via HTTPS.
current_access_key = self.get_access_key(access_key_id)
# Copy and only change the single parameter.
payload_dict = KeenApi._build_access_key_dict(current_access_key)
# Turn into sets to avoid duplicates.
old_permissions = set(payload_dict["permitted"])
removal_permissions = set(permissions)
reduced_permissions = old_permissions.difference(removal_permissions)
payload_dict["permitted"] = list(reduced_permissions)
# Now just treat it like a full update.
return self.update_access_key_full(access_key_id, **payload_dict) | [
"def",
"remove_access_key_permissions",
"(",
"self",
",",
"access_key_id",
",",
"permissions",
")",
":",
"# Get current state via HTTPS.",
"current_access_key",
"=",
"self",
".",
"get_access_key",
"(",
"access_key_id",
")",
"# Copy and only change the single parameter.",
"pay... | Removes a list of permissions from the existing list of permissions.
Will not remove all existing permissions unless all such permissions are included
in this list. Not to be confused with key revocation.
See also: revoke_access_key()
:param access_key_id: the 'key' value of the access key to remove some permissions from
:param permissions: the permissions you wish to remove from this access key | [
"Removes",
"a",
"list",
"of",
"permissions",
"from",
"the",
"existing",
"list",
"of",
"permissions",
".",
"Will",
"not",
"remove",
"all",
"existing",
"permissions",
"unless",
"all",
"such",
"permissions",
"are",
"included",
"in",
"this",
"list",
".",
"Not",
... | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L377-L401 |
14,459 | keenlabs/KeenClient-Python | keen/api.py | KeenApi._error_handling | def _error_handling(self, res):
"""
Helper function to do the error handling
:params res: the response from a request
"""
# making the error handling generic so if an status_code starting with 2 doesn't exist, we raise the error
if res.status_code // 100 != 2:
error = self._get_response_json(res)
raise exceptions.KeenApiError(error) | python | def _error_handling(self, res):
# making the error handling generic so if an status_code starting with 2 doesn't exist, we raise the error
if res.status_code // 100 != 2:
error = self._get_response_json(res)
raise exceptions.KeenApiError(error) | [
"def",
"_error_handling",
"(",
"self",
",",
"res",
")",
":",
"# making the error handling generic so if an status_code starting with 2 doesn't exist, we raise the error",
"if",
"res",
".",
"status_code",
"//",
"100",
"!=",
"2",
":",
"error",
"=",
"self",
".",
"_get_respon... | Helper function to do the error handling
:params res: the response from a request | [
"Helper",
"function",
"to",
"do",
"the",
"error",
"handling"
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L484-L494 |
14,460 | keenlabs/KeenClient-Python | keen/api.py | KeenApi._get_response_json | def _get_response_json(self, res):
"""
Helper function to extract the JSON body out of a response OR throw an exception.
:param res: the response from a request
:return: the JSON body OR throws an exception
"""
try:
error = res.json()
except ValueError:
error = {
"message": "The API did not respond with JSON, but: {0}".format(res.text[:1000]),
"error_code": "{0}".format(res.status_code)
}
return error | python | def _get_response_json(self, res):
try:
error = res.json()
except ValueError:
error = {
"message": "The API did not respond with JSON, but: {0}".format(res.text[:1000]),
"error_code": "{0}".format(res.status_code)
}
return error | [
"def",
"_get_response_json",
"(",
"self",
",",
"res",
")",
":",
"try",
":",
"error",
"=",
"res",
".",
"json",
"(",
")",
"except",
"ValueError",
":",
"error",
"=",
"{",
"\"message\"",
":",
"\"The API did not respond with JSON, but: {0}\"",
".",
"format",
"(",
... | Helper function to extract the JSON body out of a response OR throw an exception.
:param res: the response from a request
:return: the JSON body OR throws an exception | [
"Helper",
"function",
"to",
"extract",
"the",
"JSON",
"body",
"out",
"of",
"a",
"response",
"OR",
"throw",
"an",
"exception",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/api.py#L496-L511 |
14,461 | keenlabs/KeenClient-Python | keen/cached_datasets.py | CachedDatasetsInterface.all | def all(self):
""" Fetch all Cached Datasets for a Project. Read key must be set.
"""
return self._get_json(HTTPMethods.GET,
self._cached_datasets_url,
self._get_master_key()) | python | def all(self):
return self._get_json(HTTPMethods.GET,
self._cached_datasets_url,
self._get_master_key()) | [
"def",
"all",
"(",
"self",
")",
":",
"return",
"self",
".",
"_get_json",
"(",
"HTTPMethods",
".",
"GET",
",",
"self",
".",
"_cached_datasets_url",
",",
"self",
".",
"_get_master_key",
"(",
")",
")"
] | Fetch all Cached Datasets for a Project. Read key must be set. | [
"Fetch",
"all",
"Cached",
"Datasets",
"for",
"a",
"Project",
".",
"Read",
"key",
"must",
"be",
"set",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/cached_datasets.py#L16-L21 |
14,462 | keenlabs/KeenClient-Python | keen/cached_datasets.py | CachedDatasetsInterface.get | def get(self, dataset_name):
""" Fetch a single Cached Dataset for a Project. Read key must be set.
:param dataset_name: Name of Cached Dataset (not `display_name`)
"""
url = "{0}/{1}".format(self._cached_datasets_url, dataset_name)
return self._get_json(HTTPMethods.GET, url, self._get_read_key()) | python | def get(self, dataset_name):
url = "{0}/{1}".format(self._cached_datasets_url, dataset_name)
return self._get_json(HTTPMethods.GET, url, self._get_read_key()) | [
"def",
"get",
"(",
"self",
",",
"dataset_name",
")",
":",
"url",
"=",
"\"{0}/{1}\"",
".",
"format",
"(",
"self",
".",
"_cached_datasets_url",
",",
"dataset_name",
")",
"return",
"self",
".",
"_get_json",
"(",
"HTTPMethods",
".",
"GET",
",",
"url",
",",
"... | Fetch a single Cached Dataset for a Project. Read key must be set.
:param dataset_name: Name of Cached Dataset (not `display_name`) | [
"Fetch",
"a",
"single",
"Cached",
"Dataset",
"for",
"a",
"Project",
".",
"Read",
"key",
"must",
"be",
"set",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/cached_datasets.py#L24-L30 |
14,463 | keenlabs/KeenClient-Python | keen/cached_datasets.py | CachedDatasetsInterface.create | def create(self, dataset_name, query, index_by, display_name):
""" Create a Cached Dataset for a Project. Master key must be set.
"""
url = "{0}/{1}".format(self._cached_datasets_url, dataset_name)
payload = {
"query": query,
"index_by": index_by,
"display_name": display_name
}
return self._get_json(HTTPMethods.PUT, url, self._get_master_key(), json=payload) | python | def create(self, dataset_name, query, index_by, display_name):
url = "{0}/{1}".format(self._cached_datasets_url, dataset_name)
payload = {
"query": query,
"index_by": index_by,
"display_name": display_name
}
return self._get_json(HTTPMethods.PUT, url, self._get_master_key(), json=payload) | [
"def",
"create",
"(",
"self",
",",
"dataset_name",
",",
"query",
",",
"index_by",
",",
"display_name",
")",
":",
"url",
"=",
"\"{0}/{1}\"",
".",
"format",
"(",
"self",
".",
"_cached_datasets_url",
",",
"dataset_name",
")",
"payload",
"=",
"{",
"\"query\"",
... | Create a Cached Dataset for a Project. Master key must be set. | [
"Create",
"a",
"Cached",
"Dataset",
"for",
"a",
"Project",
".",
"Master",
"key",
"must",
"be",
"set",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/cached_datasets.py#L33-L42 |
14,464 | keenlabs/KeenClient-Python | keen/cached_datasets.py | CachedDatasetsInterface.results | def results(self, dataset_name, index_by, timeframe):
""" Retrieve results from a Cached Dataset. Read key must be set.
"""
url = "{0}/{1}/results".format(self._cached_datasets_url, dataset_name)
index_by = index_by if isinstance(index_by, str) else json.dumps(index_by)
timeframe = timeframe if isinstance(timeframe, str) else json.dumps(timeframe)
query_params = {
"index_by": index_by,
"timeframe": timeframe
}
return self._get_json(
HTTPMethods.GET, url, self._get_read_key(), params=query_params
) | python | def results(self, dataset_name, index_by, timeframe):
url = "{0}/{1}/results".format(self._cached_datasets_url, dataset_name)
index_by = index_by if isinstance(index_by, str) else json.dumps(index_by)
timeframe = timeframe if isinstance(timeframe, str) else json.dumps(timeframe)
query_params = {
"index_by": index_by,
"timeframe": timeframe
}
return self._get_json(
HTTPMethods.GET, url, self._get_read_key(), params=query_params
) | [
"def",
"results",
"(",
"self",
",",
"dataset_name",
",",
"index_by",
",",
"timeframe",
")",
":",
"url",
"=",
"\"{0}/{1}/results\"",
".",
"format",
"(",
"self",
".",
"_cached_datasets_url",
",",
"dataset_name",
")",
"index_by",
"=",
"index_by",
"if",
"isinstanc... | Retrieve results from a Cached Dataset. Read key must be set. | [
"Retrieve",
"results",
"from",
"a",
"Cached",
"Dataset",
".",
"Read",
"key",
"must",
"be",
"set",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/cached_datasets.py#L45-L60 |
14,465 | keenlabs/KeenClient-Python | keen/cached_datasets.py | CachedDatasetsInterface.delete | def delete(self, dataset_name):
""" Delete a Cached Dataset. Master Key must be set.
"""
url = "{0}/{1}".format(self._cached_datasets_url, dataset_name)
self._get_json(HTTPMethods.DELETE, url, self._get_master_key())
return True | python | def delete(self, dataset_name):
url = "{0}/{1}".format(self._cached_datasets_url, dataset_name)
self._get_json(HTTPMethods.DELETE, url, self._get_master_key())
return True | [
"def",
"delete",
"(",
"self",
",",
"dataset_name",
")",
":",
"url",
"=",
"\"{0}/{1}\"",
".",
"format",
"(",
"self",
".",
"_cached_datasets_url",
",",
"dataset_name",
")",
"self",
".",
"_get_json",
"(",
"HTTPMethods",
".",
"DELETE",
",",
"url",
",",
"self",... | Delete a Cached Dataset. Master Key must be set. | [
"Delete",
"a",
"Cached",
"Dataset",
".",
"Master",
"Key",
"must",
"be",
"set",
"."
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/cached_datasets.py#L63-L68 |
14,466 | keenlabs/KeenClient-Python | keen/utilities.py | switch.match | def match(self, *args):
"""Whether or not to enter a given case statement"""
self.fall = self.fall or not args
self.fall = self.fall or (self.value in args)
return self.fall | python | def match(self, *args):
self.fall = self.fall or not args
self.fall = self.fall or (self.value in args)
return self.fall | [
"def",
"match",
"(",
"self",
",",
"*",
"args",
")",
":",
"self",
".",
"fall",
"=",
"self",
".",
"fall",
"or",
"not",
"args",
"self",
".",
"fall",
"=",
"self",
".",
"fall",
"or",
"(",
"self",
".",
"value",
"in",
"args",
")",
"return",
"self",
".... | Whether or not to enter a given case statement | [
"Whether",
"or",
"not",
"to",
"enter",
"a",
"given",
"case",
"statement"
] | 266387c3376d1e000d117e17c45045ae3439d43f | https://github.com/keenlabs/KeenClient-Python/blob/266387c3376d1e000d117e17c45045ae3439d43f/keen/utilities.py#L50-L56 |
14,467 | quintusdias/glymur | glymur/config.py | glymurrc_fname | def glymurrc_fname():
"""Return the path to the configuration file.
Search order:
1) current working directory
2) environ var XDG_CONFIG_HOME
3) $HOME/.config/glymur/glymurrc
"""
# Current directory.
fname = os.path.join(os.getcwd(), 'glymurrc')
if os.path.exists(fname):
return fname
confdir = get_configdir()
if confdir is not None:
fname = os.path.join(confdir, 'glymurrc')
if os.path.exists(fname):
return fname
# didn't find a configuration file.
return None | python | def glymurrc_fname():
# Current directory.
fname = os.path.join(os.getcwd(), 'glymurrc')
if os.path.exists(fname):
return fname
confdir = get_configdir()
if confdir is not None:
fname = os.path.join(confdir, 'glymurrc')
if os.path.exists(fname):
return fname
# didn't find a configuration file.
return None | [
"def",
"glymurrc_fname",
"(",
")",
":",
"# Current directory.",
"fname",
"=",
"os",
".",
"path",
".",
"join",
"(",
"os",
".",
"getcwd",
"(",
")",
",",
"'glymurrc'",
")",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"fname",
")",
":",
"return",
"fname... | Return the path to the configuration file.
Search order:
1) current working directory
2) environ var XDG_CONFIG_HOME
3) $HOME/.config/glymur/glymurrc | [
"Return",
"the",
"path",
"to",
"the",
"configuration",
"file",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/config.py#L24-L45 |
14,468 | quintusdias/glymur | glymur/config.py | load_library_handle | def load_library_handle(libname, path):
"""Load the library, return the ctypes handle."""
if path is None or path in ['None', 'none']:
# Either could not find a library via ctypes or
# user-configuration-file, or we could not find it in any of the
# default locations, or possibly the user intentionally does not want
# one of the libraries to load.
return None
try:
if os.name == "nt":
opj_lib = ctypes.windll.LoadLibrary(path)
else:
opj_lib = ctypes.CDLL(path)
except (TypeError, OSError):
msg = 'The {libname} library at {path} could not be loaded.'
msg = msg.format(path=path, libname=libname)
warnings.warn(msg, UserWarning)
opj_lib = None
return opj_lib | python | def load_library_handle(libname, path):
if path is None or path in ['None', 'none']:
# Either could not find a library via ctypes or
# user-configuration-file, or we could not find it in any of the
# default locations, or possibly the user intentionally does not want
# one of the libraries to load.
return None
try:
if os.name == "nt":
opj_lib = ctypes.windll.LoadLibrary(path)
else:
opj_lib = ctypes.CDLL(path)
except (TypeError, OSError):
msg = 'The {libname} library at {path} could not be loaded.'
msg = msg.format(path=path, libname=libname)
warnings.warn(msg, UserWarning)
opj_lib = None
return opj_lib | [
"def",
"load_library_handle",
"(",
"libname",
",",
"path",
")",
":",
"if",
"path",
"is",
"None",
"or",
"path",
"in",
"[",
"'None'",
",",
"'none'",
"]",
":",
"# Either could not find a library via ctypes or",
"# user-configuration-file, or we could not find it in any of th... | Load the library, return the ctypes handle. | [
"Load",
"the",
"library",
"return",
"the",
"ctypes",
"handle",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/config.py#L86-L107 |
14,469 | quintusdias/glymur | glymur/config.py | read_config_file | def read_config_file(libname):
"""
Extract library locations from a configuration file.
Parameters
----------
libname : str
One of either 'openjp2' or 'openjpeg'
Returns
-------
path : None or str
None if no location is specified, otherwise a path to the library
"""
filename = glymurrc_fname()
if filename is None:
# There's no library file path to return in this case.
return None
# Read the configuration file for the library location.
parser = ConfigParser()
parser.read(filename)
try:
path = parser.get('library', libname)
except (NoOptionError, NoSectionError):
path = None
return path | python | def read_config_file(libname):
filename = glymurrc_fname()
if filename is None:
# There's no library file path to return in this case.
return None
# Read the configuration file for the library location.
parser = ConfigParser()
parser.read(filename)
try:
path = parser.get('library', libname)
except (NoOptionError, NoSectionError):
path = None
return path | [
"def",
"read_config_file",
"(",
"libname",
")",
":",
"filename",
"=",
"glymurrc_fname",
"(",
")",
"if",
"filename",
"is",
"None",
":",
"# There's no library file path to return in this case.",
"return",
"None",
"# Read the configuration file for the library location.",
"parse... | Extract library locations from a configuration file.
Parameters
----------
libname : str
One of either 'openjp2' or 'openjpeg'
Returns
-------
path : None or str
None if no location is specified, otherwise a path to the library | [
"Extract",
"library",
"locations",
"from",
"a",
"configuration",
"file",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/config.py#L110-L136 |
14,470 | quintusdias/glymur | glymur/config.py | glymur_config | def glymur_config():
"""
Try to ascertain locations of openjp2, openjpeg libraries.
Returns
-------
tuple
tuple of library handles
"""
handles = (load_openjpeg_library(x) for x in ['openjp2', 'openjpeg'])
handles = tuple(handles)
if all(handle is None for handle in handles):
msg = "Neither the openjp2 nor the openjpeg library could be loaded. "
warnings.warn(msg)
return handles | python | def glymur_config():
handles = (load_openjpeg_library(x) for x in ['openjp2', 'openjpeg'])
handles = tuple(handles)
if all(handle is None for handle in handles):
msg = "Neither the openjp2 nor the openjpeg library could be loaded. "
warnings.warn(msg)
return handles | [
"def",
"glymur_config",
"(",
")",
":",
"handles",
"=",
"(",
"load_openjpeg_library",
"(",
"x",
")",
"for",
"x",
"in",
"[",
"'openjp2'",
",",
"'openjpeg'",
"]",
")",
"handles",
"=",
"tuple",
"(",
"handles",
")",
"if",
"all",
"(",
"handle",
"is",
"None",... | Try to ascertain locations of openjp2, openjpeg libraries.
Returns
-------
tuple
tuple of library handles | [
"Try",
"to",
"ascertain",
"locations",
"of",
"openjp2",
"openjpeg",
"libraries",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/config.py#L139-L154 |
14,471 | quintusdias/glymur | glymur/config.py | get_configdir | def get_configdir():
"""Return string representing the configuration directory.
Default is $HOME/.config/glymur. You can override this with the
XDG_CONFIG_HOME environment variable.
"""
if 'XDG_CONFIG_HOME' in os.environ:
return os.path.join(os.environ['XDG_CONFIG_HOME'], 'glymur')
if 'HOME' in os.environ and os.name != 'nt':
# HOME is set by WinPython to something unusual, so we don't
# necessarily want that.
return os.path.join(os.environ['HOME'], '.config', 'glymur')
# Last stand. Should handle windows... others?
return os.path.join(os.path.expanduser('~'), 'glymur') | python | def get_configdir():
if 'XDG_CONFIG_HOME' in os.environ:
return os.path.join(os.environ['XDG_CONFIG_HOME'], 'glymur')
if 'HOME' in os.environ and os.name != 'nt':
# HOME is set by WinPython to something unusual, so we don't
# necessarily want that.
return os.path.join(os.environ['HOME'], '.config', 'glymur')
# Last stand. Should handle windows... others?
return os.path.join(os.path.expanduser('~'), 'glymur') | [
"def",
"get_configdir",
"(",
")",
":",
"if",
"'XDG_CONFIG_HOME'",
"in",
"os",
".",
"environ",
":",
"return",
"os",
".",
"path",
".",
"join",
"(",
"os",
".",
"environ",
"[",
"'XDG_CONFIG_HOME'",
"]",
",",
"'glymur'",
")",
"if",
"'HOME'",
"in",
"os",
"."... | Return string representing the configuration directory.
Default is $HOME/.config/glymur. You can override this with the
XDG_CONFIG_HOME environment variable. | [
"Return",
"string",
"representing",
"the",
"configuration",
"directory",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/config.py#L157-L172 |
14,472 | quintusdias/glymur | glymur/config.py | set_option | def set_option(key, value):
"""Set the value of the specified option.
Available options:
parse.full_codestream
print.xml
print.codestream
print.short
Parameters
----------
key : str
Name of a single option.
value :
New value of option.
Option Descriptions
-------------------
parse.full_codestream : bool
When False, only the codestream header is parsed for metadata. This
can results in faster JP2/JPX parsing. When True, the entire
codestream is parsed. [default: False]
print.codestream : bool
When False, the codestream segments are not printed. Otherwise the
segments are printed depending on the value of the
parse.full_codestream option. [default: True]
print.short : bool
When True, only the box ID, offset, and length are displayed. Useful
for displaying only the basic structure or skeleton of a JPEG 2000
file. [default: False]
print.xml : bool
When False, printing of the XML contents of any XML boxes or UUID XMP
boxes is suppressed. [default: True]
See also
--------
get_option
"""
if key not in _options.keys():
raise KeyError('{key} not valid.'.format(key=key))
_options[key] = value | python | def set_option(key, value):
if key not in _options.keys():
raise KeyError('{key} not valid.'.format(key=key))
_options[key] = value | [
"def",
"set_option",
"(",
"key",
",",
"value",
")",
":",
"if",
"key",
"not",
"in",
"_options",
".",
"keys",
"(",
")",
":",
"raise",
"KeyError",
"(",
"'{key} not valid.'",
".",
"format",
"(",
"key",
"=",
"key",
")",
")",
"_options",
"[",
"key",
"]",
... | Set the value of the specified option.
Available options:
parse.full_codestream
print.xml
print.codestream
print.short
Parameters
----------
key : str
Name of a single option.
value :
New value of option.
Option Descriptions
-------------------
parse.full_codestream : bool
When False, only the codestream header is parsed for metadata. This
can results in faster JP2/JPX parsing. When True, the entire
codestream is parsed. [default: False]
print.codestream : bool
When False, the codestream segments are not printed. Otherwise the
segments are printed depending on the value of the
parse.full_codestream option. [default: True]
print.short : bool
When True, only the box ID, offset, and length are displayed. Useful
for displaying only the basic structure or skeleton of a JPEG 2000
file. [default: False]
print.xml : bool
When False, printing of the XML contents of any XML boxes or UUID XMP
boxes is suppressed. [default: True]
See also
--------
get_option | [
"Set",
"the",
"value",
"of",
"the",
"specified",
"option",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/config.py#L184-L225 |
14,473 | quintusdias/glymur | glymur/config.py | reset_option | def reset_option(key):
"""
Reset one or more options to their default value.
Pass "all" as argument to reset all options.
Available options:
parse.full_codestream
print.xml
print.codestream
print.short
Parameter
---------
key : str
Name of a single option.
"""
global _options
if key == 'all':
_options = copy.deepcopy(_original_options)
else:
if key not in _options.keys():
raise KeyError('{key} not valid.'.format(key=key))
_options[key] = _original_options[key] | python | def reset_option(key):
global _options
if key == 'all':
_options = copy.deepcopy(_original_options)
else:
if key not in _options.keys():
raise KeyError('{key} not valid.'.format(key=key))
_options[key] = _original_options[key] | [
"def",
"reset_option",
"(",
"key",
")",
":",
"global",
"_options",
"if",
"key",
"==",
"'all'",
":",
"_options",
"=",
"copy",
".",
"deepcopy",
"(",
"_original_options",
")",
"else",
":",
"if",
"key",
"not",
"in",
"_options",
".",
"keys",
"(",
")",
":",
... | Reset one or more options to their default value.
Pass "all" as argument to reset all options.
Available options:
parse.full_codestream
print.xml
print.codestream
print.short
Parameter
---------
key : str
Name of a single option. | [
"Reset",
"one",
"or",
"more",
"options",
"to",
"their",
"default",
"value",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/config.py#L254-L278 |
14,474 | quintusdias/glymur | glymur/config.py | set_printoptions | def set_printoptions(**kwargs):
"""Set printing options.
These options determine the way JPEG 2000 boxes are displayed.
Parameters
----------
short : bool, optional
When True, only the box ID, offset, and length are displayed. Useful
for displaying only the basic structure or skeleton of a JPEG 2000
file.
xml : bool, optional
When False, printing of the XML contents of any XML boxes or UUID XMP
boxes is suppressed.
codestream : bool, optional
When False, the codestream segments are not printed. Otherwise the
segments are printed depending on how set_parseoptions has been used.
See also
--------
get_printoptions
Examples
--------
To put back the default options, you can use:
>>> import glymur
>>> glymur.set_printoptions(short=False, xml=True, codestream=True)
"""
warnings.warn('Use set_option instead of set_printoptions.',
DeprecationWarning)
for key, value in kwargs.items():
if key not in ['short', 'xml', 'codestream']:
raise KeyError('"{0}" not a valid keyword parameter.'.format(key))
set_option('print.' + key, value) | python | def set_printoptions(**kwargs):
warnings.warn('Use set_option instead of set_printoptions.',
DeprecationWarning)
for key, value in kwargs.items():
if key not in ['short', 'xml', 'codestream']:
raise KeyError('"{0}" not a valid keyword parameter.'.format(key))
set_option('print.' + key, value) | [
"def",
"set_printoptions",
"(",
"*",
"*",
"kwargs",
")",
":",
"warnings",
".",
"warn",
"(",
"'Use set_option instead of set_printoptions.'",
",",
"DeprecationWarning",
")",
"for",
"key",
",",
"value",
"in",
"kwargs",
".",
"items",
"(",
")",
":",
"if",
"key",
... | Set printing options.
These options determine the way JPEG 2000 boxes are displayed.
Parameters
----------
short : bool, optional
When True, only the box ID, offset, and length are displayed. Useful
for displaying only the basic structure or skeleton of a JPEG 2000
file.
xml : bool, optional
When False, printing of the XML contents of any XML boxes or UUID XMP
boxes is suppressed.
codestream : bool, optional
When False, the codestream segments are not printed. Otherwise the
segments are printed depending on how set_parseoptions has been used.
See also
--------
get_printoptions
Examples
--------
To put back the default options, you can use:
>>> import glymur
>>> glymur.set_printoptions(short=False, xml=True, codestream=True) | [
"Set",
"printing",
"options",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/config.py#L330-L364 |
14,475 | quintusdias/glymur | glymur/config.py | get_printoptions | def get_printoptions():
"""Return the current print options.
Returns
-------
dict
Dictionary of current print options with keys
- short : bool
- xml : bool
- codestream : bool
For a full description of these options, see `set_printoptions`.
See also
--------
set_printoptions
"""
warnings.warn('Use get_option instead of get_printoptions.',
DeprecationWarning)
d = {}
for key in ['short', 'xml', 'codestream']:
d[key] = _options['print.' + key]
return d | python | def get_printoptions():
warnings.warn('Use get_option instead of get_printoptions.',
DeprecationWarning)
d = {}
for key in ['short', 'xml', 'codestream']:
d[key] = _options['print.' + key]
return d | [
"def",
"get_printoptions",
"(",
")",
":",
"warnings",
".",
"warn",
"(",
"'Use get_option instead of get_printoptions.'",
",",
"DeprecationWarning",
")",
"d",
"=",
"{",
"}",
"for",
"key",
"in",
"[",
"'short'",
",",
"'xml'",
",",
"'codestream'",
"]",
":",
"d",
... | Return the current print options.
Returns
-------
dict
Dictionary of current print options with keys
- short : bool
- xml : bool
- codestream : bool
For a full description of these options, see `set_printoptions`.
See also
--------
set_printoptions | [
"Return",
"the",
"current",
"print",
"options",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/config.py#L367-L390 |
14,476 | quintusdias/glymur | glymur/command_line.py | main | def main():
"""
Entry point for console script jp2dump.
"""
kwargs = {'description': 'Print JPEG2000 metadata.',
'formatter_class': argparse.ArgumentDefaultsHelpFormatter}
parser = argparse.ArgumentParser(**kwargs)
parser.add_argument('-x', '--noxml',
help='suppress XML',
action='store_true')
parser.add_argument('-s', '--short',
help='only print box id, offset, and length',
action='store_true')
chelp = 'Level of codestream information. 0 suppresses all details, '
chelp += '1 prints the main header, 2 prints the full codestream.'
parser.add_argument('-c', '--codestream',
help=chelp,
metavar='LEVEL',
nargs=1,
type=int,
default=[1])
parser.add_argument('filename')
args = parser.parse_args()
if args.noxml:
set_option('print.xml', False)
if args.short:
set_option('print.short', True)
codestream_level = args.codestream[0]
if codestream_level not in [0, 1, 2]:
raise ValueError("Invalid level of codestream information specified.")
if codestream_level == 0:
set_option('print.codestream', False)
elif codestream_level == 2:
set_option('parse.full_codestream', True)
filename = args.filename
# JP2 metadata can be extensive, so don't print any warnings until we
# are done with the metadata.
with warnings.catch_warnings(record=True) as wctx:
jp2 = Jp2k(filename)
if jp2._codec_format == lib.openjp2.CODEC_J2K:
if codestream_level == 0:
print('File: {0}'.format(os.path.basename(filename)))
elif codestream_level == 1:
print(jp2)
elif codestream_level == 2:
print('File: {0}'.format(os.path.basename(filename)))
print(jp2.get_codestream(header_only=False))
else:
print(jp2)
# Now re-emit any suppressed warnings.
if len(wctx) > 0:
print("\n")
for warning in wctx:
print("{0}:{1}: {2}: {3}".format(warning.filename,
warning.lineno,
warning.category.__name__,
warning.message)) | python | def main():
kwargs = {'description': 'Print JPEG2000 metadata.',
'formatter_class': argparse.ArgumentDefaultsHelpFormatter}
parser = argparse.ArgumentParser(**kwargs)
parser.add_argument('-x', '--noxml',
help='suppress XML',
action='store_true')
parser.add_argument('-s', '--short',
help='only print box id, offset, and length',
action='store_true')
chelp = 'Level of codestream information. 0 suppresses all details, '
chelp += '1 prints the main header, 2 prints the full codestream.'
parser.add_argument('-c', '--codestream',
help=chelp,
metavar='LEVEL',
nargs=1,
type=int,
default=[1])
parser.add_argument('filename')
args = parser.parse_args()
if args.noxml:
set_option('print.xml', False)
if args.short:
set_option('print.short', True)
codestream_level = args.codestream[0]
if codestream_level not in [0, 1, 2]:
raise ValueError("Invalid level of codestream information specified.")
if codestream_level == 0:
set_option('print.codestream', False)
elif codestream_level == 2:
set_option('parse.full_codestream', True)
filename = args.filename
# JP2 metadata can be extensive, so don't print any warnings until we
# are done with the metadata.
with warnings.catch_warnings(record=True) as wctx:
jp2 = Jp2k(filename)
if jp2._codec_format == lib.openjp2.CODEC_J2K:
if codestream_level == 0:
print('File: {0}'.format(os.path.basename(filename)))
elif codestream_level == 1:
print(jp2)
elif codestream_level == 2:
print('File: {0}'.format(os.path.basename(filename)))
print(jp2.get_codestream(header_only=False))
else:
print(jp2)
# Now re-emit any suppressed warnings.
if len(wctx) > 0:
print("\n")
for warning in wctx:
print("{0}:{1}: {2}: {3}".format(warning.filename,
warning.lineno,
warning.category.__name__,
warning.message)) | [
"def",
"main",
"(",
")",
":",
"kwargs",
"=",
"{",
"'description'",
":",
"'Print JPEG2000 metadata.'",
",",
"'formatter_class'",
":",
"argparse",
".",
"ArgumentDefaultsHelpFormatter",
"}",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"*",
"*",
"kwargs",
... | Entry point for console script jp2dump. | [
"Entry",
"point",
"for",
"console",
"script",
"jp2dump",
"."
] | 8b8fb091130fff00f1028dc82219e69e3f9baf6d | https://github.com/quintusdias/glymur/blob/8b8fb091130fff00f1028dc82219e69e3f9baf6d/glymur/command_line.py#L13-L80 |
14,477 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient._set_token | def _set_token(self):
"""Set the Cerberus token based on auth type"""
try:
self.token = os.environ['CERBERUS_TOKEN']
if self.verbose:
print("Overriding Cerberus token with environment variable.", file=sys.stderr)
logger.info("Overriding Cerberus token with environment variable.")
return
except:
pass
if self.username:
ua = UserAuth(self.cerberus_url, self.username, self.password)
self.token = ua.get_token()
else:
awsa = AWSAuth(self.cerberus_url, region=self.region, aws_session=self.aws_session, verbose=self.verbose)
self.token = awsa.get_token() | python | def _set_token(self):
try:
self.token = os.environ['CERBERUS_TOKEN']
if self.verbose:
print("Overriding Cerberus token with environment variable.", file=sys.stderr)
logger.info("Overriding Cerberus token with environment variable.")
return
except:
pass
if self.username:
ua = UserAuth(self.cerberus_url, self.username, self.password)
self.token = ua.get_token()
else:
awsa = AWSAuth(self.cerberus_url, region=self.region, aws_session=self.aws_session, verbose=self.verbose)
self.token = awsa.get_token() | [
"def",
"_set_token",
"(",
"self",
")",
":",
"try",
":",
"self",
".",
"token",
"=",
"os",
".",
"environ",
"[",
"'CERBERUS_TOKEN'",
"]",
"if",
"self",
".",
"verbose",
":",
"print",
"(",
"\"Overriding Cerberus token with environment variable.\"",
",",
"file",
"="... | Set the Cerberus token based on auth type | [
"Set",
"the",
"Cerberus",
"token",
"based",
"on",
"auth",
"type"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L74-L89 |
14,478 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_role | def get_role(self, key):
"""Return id of named role."""
json_resp = self.get_roles()
for item in json_resp:
if key in item["name"]:
return item["id"]
raise CerberusClientException("Key '%s' not found" % key) | python | def get_role(self, key):
json_resp = self.get_roles()
for item in json_resp:
if key in item["name"]:
return item["id"]
raise CerberusClientException("Key '%s' not found" % key) | [
"def",
"get_role",
"(",
"self",
",",
"key",
")",
":",
"json_resp",
"=",
"self",
".",
"get_roles",
"(",
")",
"for",
"item",
"in",
"json_resp",
":",
"if",
"key",
"in",
"item",
"[",
"\"name\"",
"]",
":",
"return",
"item",
"[",
"\"id\"",
"]",
"raise",
... | Return id of named role. | [
"Return",
"id",
"of",
"named",
"role",
"."
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L106-L113 |
14,479 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_categories | def get_categories(self):
""" Return a list of categories that a safe deposit box can belong to"""
sdb_resp = get_with_retry(self.cerberus_url + '/v1/category',
headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp.json() | python | def get_categories(self):
sdb_resp = get_with_retry(self.cerberus_url + '/v1/category',
headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp.json() | [
"def",
"get_categories",
"(",
"self",
")",
":",
"sdb_resp",
"=",
"get_with_retry",
"(",
"self",
".",
"cerberus_url",
"+",
"'/v1/category'",
",",
"headers",
"=",
"self",
".",
"HEADERS",
")",
"throw_if_bad_response",
"(",
"sdb_resp",
")",
"return",
"sdb_resp",
"... | Return a list of categories that a safe deposit box can belong to | [
"Return",
"a",
"list",
"of",
"categories",
"that",
"a",
"safe",
"deposit",
"box",
"can",
"belong",
"to"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L124-L130 |
14,480 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.create_sdb | def create_sdb(self, name, category_id, owner, description="", user_group_permissions=None,
iam_principal_permissions=None):
"""Create a safe deposit box.
You need to refresh your token before the iam role is granted permission to the new safe deposit box.
Keyword arguments:
name (string) -- name of the safe deposit box
category_id (string) -- category id that determines where to store the sdb. (ex: shared, applications)
owner (string) -- AD group that owns the safe deposit box
description (string) -- Description of the safe deposit box
user_group_permissions (list) -- list of dictionaries containing the key name and maybe role_id
iam_principal_permissions (list) -- list of dictionaries containing the key name iam_principal_arn
and role_id
"""
# Do some sanity checking
if user_group_permissions is None:
user_group_permissions = []
if iam_principal_permissions is None:
iam_principal_permissions = []
if list != type(user_group_permissions):
raise(TypeError('Expected list, but got ' + str(type(user_group_permissions))))
if list != type(iam_principal_permissions):
raise(TypeError('Expected list, but got ' + str(type(iam_principal_permissions))))
temp_data = {
"name": name,
"description": description,
"category_id": category_id,
"owner": owner,
}
if len(user_group_permissions) > 0:
temp_data["user_group_permissions"] = user_group_permissions
if len(iam_principal_permissions) > 0:
temp_data["iam_principal_permissions"] = iam_principal_permissions
data = json.encoder.JSONEncoder().encode(temp_data)
sdb_resp = post_with_retry(self.cerberus_url + '/v2/safe-deposit-box', data=str(data), headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp.json() | python | def create_sdb(self, name, category_id, owner, description="", user_group_permissions=None,
iam_principal_permissions=None):
# Do some sanity checking
if user_group_permissions is None:
user_group_permissions = []
if iam_principal_permissions is None:
iam_principal_permissions = []
if list != type(user_group_permissions):
raise(TypeError('Expected list, but got ' + str(type(user_group_permissions))))
if list != type(iam_principal_permissions):
raise(TypeError('Expected list, but got ' + str(type(iam_principal_permissions))))
temp_data = {
"name": name,
"description": description,
"category_id": category_id,
"owner": owner,
}
if len(user_group_permissions) > 0:
temp_data["user_group_permissions"] = user_group_permissions
if len(iam_principal_permissions) > 0:
temp_data["iam_principal_permissions"] = iam_principal_permissions
data = json.encoder.JSONEncoder().encode(temp_data)
sdb_resp = post_with_retry(self.cerberus_url + '/v2/safe-deposit-box', data=str(data), headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp.json() | [
"def",
"create_sdb",
"(",
"self",
",",
"name",
",",
"category_id",
",",
"owner",
",",
"description",
"=",
"\"\"",
",",
"user_group_permissions",
"=",
"None",
",",
"iam_principal_permissions",
"=",
"None",
")",
":",
"# Do some sanity checking",
"if",
"user_group_pe... | Create a safe deposit box.
You need to refresh your token before the iam role is granted permission to the new safe deposit box.
Keyword arguments:
name (string) -- name of the safe deposit box
category_id (string) -- category id that determines where to store the sdb. (ex: shared, applications)
owner (string) -- AD group that owns the safe deposit box
description (string) -- Description of the safe deposit box
user_group_permissions (list) -- list of dictionaries containing the key name and maybe role_id
iam_principal_permissions (list) -- list of dictionaries containing the key name iam_principal_arn
and role_id | [
"Create",
"a",
"safe",
"deposit",
"box",
"."
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L132-L170 |
14,481 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.delete_sdb | def delete_sdb(self, sdb_id):
""" Delete a safe deposit box specified by id
Keyword arguments:
sdb_id -- this is the id of the safe deposit box, not the path."""
sdb_resp = delete_with_retry(self.cerberus_url + '/v2/safe-deposit-box/' + sdb_id,
headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp | python | def delete_sdb(self, sdb_id):
sdb_resp = delete_with_retry(self.cerberus_url + '/v2/safe-deposit-box/' + sdb_id,
headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp | [
"def",
"delete_sdb",
"(",
"self",
",",
"sdb_id",
")",
":",
"sdb_resp",
"=",
"delete_with_retry",
"(",
"self",
".",
"cerberus_url",
"+",
"'/v2/safe-deposit-box/'",
"+",
"sdb_id",
",",
"headers",
"=",
"self",
".",
"HEADERS",
")",
"throw_if_bad_response",
"(",
"s... | Delete a safe deposit box specified by id
Keyword arguments:
sdb_id -- this is the id of the safe deposit box, not the path. | [
"Delete",
"a",
"safe",
"deposit",
"box",
"specified",
"by",
"id"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L172-L180 |
14,482 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_sdb_path | def get_sdb_path(self, sdb):
"""Return the path for a SDB"""
sdb_id = self.get_sdb_id(sdb)
sdb_resp = get_with_retry(
self.cerberus_url + '/v1/safe-deposit-box/' + sdb_id + '/',
headers=self.HEADERS
)
throw_if_bad_response(sdb_resp)
return sdb_resp.json()['path'] | python | def get_sdb_path(self, sdb):
sdb_id = self.get_sdb_id(sdb)
sdb_resp = get_with_retry(
self.cerberus_url + '/v1/safe-deposit-box/' + sdb_id + '/',
headers=self.HEADERS
)
throw_if_bad_response(sdb_resp)
return sdb_resp.json()['path'] | [
"def",
"get_sdb_path",
"(",
"self",
",",
"sdb",
")",
":",
"sdb_id",
"=",
"self",
".",
"get_sdb_id",
"(",
"sdb",
")",
"sdb_resp",
"=",
"get_with_retry",
"(",
"self",
".",
"cerberus_url",
"+",
"'/v1/safe-deposit-box/'",
"+",
"sdb_id",
"+",
"'/'",
",",
"heade... | Return the path for a SDB | [
"Return",
"the",
"path",
"for",
"a",
"SDB"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L190-L200 |
14,483 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_sdb_keys | def get_sdb_keys(self, path):
"""Return the keys for a SDB, which are need for the full secure data path"""
list_resp = get_with_retry(
self.cerberus_url + '/v1/secret/' + path + '/?list=true',
headers=self.HEADERS
)
throw_if_bad_response(list_resp)
return list_resp.json()['data']['keys'] | python | def get_sdb_keys(self, path):
list_resp = get_with_retry(
self.cerberus_url + '/v1/secret/' + path + '/?list=true',
headers=self.HEADERS
)
throw_if_bad_response(list_resp)
return list_resp.json()['data']['keys'] | [
"def",
"get_sdb_keys",
"(",
"self",
",",
"path",
")",
":",
"list_resp",
"=",
"get_with_retry",
"(",
"self",
".",
"cerberus_url",
"+",
"'/v1/secret/'",
"+",
"path",
"+",
"'/?list=true'",
",",
"headers",
"=",
"self",
".",
"HEADERS",
")",
"throw_if_bad_response",... | Return the keys for a SDB, which are need for the full secure data path | [
"Return",
"the",
"keys",
"for",
"a",
"SDB",
"which",
"are",
"need",
"for",
"the",
"full",
"secure",
"data",
"path"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L202-L211 |
14,484 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_sdb_id | def get_sdb_id(self, sdb):
""" Return the ID for the given safe deposit box.
Keyword arguments:
sdb -- This is the name of the safe deposit box, not the path"""
json_resp = self.get_sdbs()
for r in json_resp:
if r['name'] == sdb:
return str(r['id'])
# If we haven't returned yet then we didn't find what we were
# looking for.
raise CerberusClientException("'%s' not found" % sdb) | python | def get_sdb_id(self, sdb):
json_resp = self.get_sdbs()
for r in json_resp:
if r['name'] == sdb:
return str(r['id'])
# If we haven't returned yet then we didn't find what we were
# looking for.
raise CerberusClientException("'%s' not found" % sdb) | [
"def",
"get_sdb_id",
"(",
"self",
",",
"sdb",
")",
":",
"json_resp",
"=",
"self",
".",
"get_sdbs",
"(",
")",
"for",
"r",
"in",
"json_resp",
":",
"if",
"r",
"[",
"'name'",
"]",
"==",
"sdb",
":",
"return",
"str",
"(",
"r",
"[",
"'id'",
"]",
")",
... | Return the ID for the given safe deposit box.
Keyword arguments:
sdb -- This is the name of the safe deposit box, not the path | [
"Return",
"the",
"ID",
"for",
"the",
"given",
"safe",
"deposit",
"box",
"."
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L213-L225 |
14,485 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_sdb_id_by_path | def get_sdb_id_by_path(self, sdb_path):
""" Given the path, return the ID for the given safe deposit box."""
json_resp = self.get_sdbs()
# Deal with the supplied path possibly missing an ending slash
path = self._add_slash(sdb_path)
for r in json_resp:
if r['path'] == path:
return str(r['id'])
# If we haven't returned yet then we didn't find what we were
# looking for.
raise CerberusClientException("'%s' not found" % sdb_path) | python | def get_sdb_id_by_path(self, sdb_path):
json_resp = self.get_sdbs()
# Deal with the supplied path possibly missing an ending slash
path = self._add_slash(sdb_path)
for r in json_resp:
if r['path'] == path:
return str(r['id'])
# If we haven't returned yet then we didn't find what we were
# looking for.
raise CerberusClientException("'%s' not found" % sdb_path) | [
"def",
"get_sdb_id_by_path",
"(",
"self",
",",
"sdb_path",
")",
":",
"json_resp",
"=",
"self",
".",
"get_sdbs",
"(",
")",
"# Deal with the supplied path possibly missing an ending slash",
"path",
"=",
"self",
".",
"_add_slash",
"(",
"sdb_path",
")",
"for",
"r",
"i... | Given the path, return the ID for the given safe deposit box. | [
"Given",
"the",
"path",
"return",
"the",
"ID",
"for",
"the",
"given",
"safe",
"deposit",
"box",
"."
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L227-L240 |
14,486 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_sdb_by_id | def get_sdb_by_id(self, sdb_id):
""" Return the details for the given safe deposit box id
Keyword arguments:
sdb_id -- this is the id of the safe deposit box, not the path.
"""
sdb_resp = get_with_retry(self.cerberus_url + '/v2/safe-deposit-box/' + sdb_id,
headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp.json() | python | def get_sdb_by_id(self, sdb_id):
sdb_resp = get_with_retry(self.cerberus_url + '/v2/safe-deposit-box/' + sdb_id,
headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp.json() | [
"def",
"get_sdb_by_id",
"(",
"self",
",",
"sdb_id",
")",
":",
"sdb_resp",
"=",
"get_with_retry",
"(",
"self",
".",
"cerberus_url",
"+",
"'/v2/safe-deposit-box/'",
"+",
"sdb_id",
",",
"headers",
"=",
"self",
".",
"HEADERS",
")",
"throw_if_bad_response",
"(",
"s... | Return the details for the given safe deposit box id
Keyword arguments:
sdb_id -- this is the id of the safe deposit box, not the path. | [
"Return",
"the",
"details",
"for",
"the",
"given",
"safe",
"deposit",
"box",
"id"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L242-L253 |
14,487 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_sdb_secret_version_paths | def get_sdb_secret_version_paths(self, sdb_id):
""" Get SDB secret version paths. This function takes the sdb_id """
sdb_resp = get_with_retry(str.join('', [self.cerberus_url, '/v1/sdb-secret-version-paths/', sdb_id]),
headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp.json() | python | def get_sdb_secret_version_paths(self, sdb_id):
sdb_resp = get_with_retry(str.join('', [self.cerberus_url, '/v1/sdb-secret-version-paths/', sdb_id]),
headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp.json() | [
"def",
"get_sdb_secret_version_paths",
"(",
"self",
",",
"sdb_id",
")",
":",
"sdb_resp",
"=",
"get_with_retry",
"(",
"str",
".",
"join",
"(",
"''",
",",
"[",
"self",
".",
"cerberus_url",
",",
"'/v1/sdb-secret-version-paths/'",
",",
"sdb_id",
"]",
")",
",",
"... | Get SDB secret version paths. This function takes the sdb_id | [
"Get",
"SDB",
"secret",
"version",
"paths",
".",
"This",
"function",
"takes",
"the",
"sdb_id"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L271-L278 |
14,488 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.list_sdbs | def list_sdbs(self):
""" Return sdbs by Name """
sdb_raw = self.get_sdbs()
sdbs = []
for s in sdb_raw:
sdbs.append(s['name'])
return sdbs | python | def list_sdbs(self):
sdb_raw = self.get_sdbs()
sdbs = []
for s in sdb_raw:
sdbs.append(s['name'])
return sdbs | [
"def",
"list_sdbs",
"(",
"self",
")",
":",
"sdb_raw",
"=",
"self",
".",
"get_sdbs",
"(",
")",
"sdbs",
"=",
"[",
"]",
"for",
"s",
"in",
"sdb_raw",
":",
"sdbs",
".",
"append",
"(",
"s",
"[",
"'name'",
"]",
")",
"return",
"sdbs"
] | Return sdbs by Name | [
"Return",
"sdbs",
"by",
"Name"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L284-L291 |
14,489 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.update_sdb | def update_sdb(self, sdb_id, owner=None, description=None, user_group_permissions=None,
iam_principal_permissions=None):
"""
Update a safe deposit box.
Keyword arguments:
owner (string) -- AD group that owns the safe deposit box
description (string) -- Description of the safe deposit box
user_group_permissions (list) -- list of dictionaries containing the key name and maybe role_id
iam_principal_permissions (list) -- list of dictionaries containing the key name iam_principal_arn
and role_id
"""
# Grab current data
old_data = self.get_sdb_by_id(sdb_id)
# Assemble information to update
temp_data = {}
keys = ('owner', 'description', 'iam_principal_permissions', 'user_group_permissions')
for k in keys:
if k in old_data:
temp_data[k] = old_data[k]
if owner is not None:
temp_data["owner"] = owner
if description is not None:
temp_data["description"] = description
if user_group_permissions is not None and len(user_group_permissions) > 0:
temp_data["user_group_permissions"] = user_group_permissions
if iam_principal_permissions is not None and len(iam_principal_permissions) > 0:
temp_data["iam_principal_permissions"] = iam_principal_permissions
data = json.encoder.JSONEncoder().encode(temp_data)
sdb_resp = put_with_retry(self.cerberus_url + '/v2/safe-deposit-box/' + sdb_id, data=str(data),
headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp.json() | python | def update_sdb(self, sdb_id, owner=None, description=None, user_group_permissions=None,
iam_principal_permissions=None):
# Grab current data
old_data = self.get_sdb_by_id(sdb_id)
# Assemble information to update
temp_data = {}
keys = ('owner', 'description', 'iam_principal_permissions', 'user_group_permissions')
for k in keys:
if k in old_data:
temp_data[k] = old_data[k]
if owner is not None:
temp_data["owner"] = owner
if description is not None:
temp_data["description"] = description
if user_group_permissions is not None and len(user_group_permissions) > 0:
temp_data["user_group_permissions"] = user_group_permissions
if iam_principal_permissions is not None and len(iam_principal_permissions) > 0:
temp_data["iam_principal_permissions"] = iam_principal_permissions
data = json.encoder.JSONEncoder().encode(temp_data)
sdb_resp = put_with_retry(self.cerberus_url + '/v2/safe-deposit-box/' + sdb_id, data=str(data),
headers=self.HEADERS)
throw_if_bad_response(sdb_resp)
return sdb_resp.json() | [
"def",
"update_sdb",
"(",
"self",
",",
"sdb_id",
",",
"owner",
"=",
"None",
",",
"description",
"=",
"None",
",",
"user_group_permissions",
"=",
"None",
",",
"iam_principal_permissions",
"=",
"None",
")",
":",
"# Grab current data",
"old_data",
"=",
"self",
".... | Update a safe deposit box.
Keyword arguments:
owner (string) -- AD group that owns the safe deposit box
description (string) -- Description of the safe deposit box
user_group_permissions (list) -- list of dictionaries containing the key name and maybe role_id
iam_principal_permissions (list) -- list of dictionaries containing the key name iam_principal_arn
and role_id | [
"Update",
"a",
"safe",
"deposit",
"box",
"."
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L293-L329 |
14,490 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.delete_file | def delete_file(self, secure_data_path):
"""Delete a file at the given secure data path"""
secret_resp = delete_with_retry(self.cerberus_url + '/v1/secure-file/' + secure_data_path,
headers=self.HEADERS)
throw_if_bad_response(secret_resp)
return secret_resp | python | def delete_file(self, secure_data_path):
secret_resp = delete_with_retry(self.cerberus_url + '/v1/secure-file/' + secure_data_path,
headers=self.HEADERS)
throw_if_bad_response(secret_resp)
return secret_resp | [
"def",
"delete_file",
"(",
"self",
",",
"secure_data_path",
")",
":",
"secret_resp",
"=",
"delete_with_retry",
"(",
"self",
".",
"cerberus_url",
"+",
"'/v1/secure-file/'",
"+",
"secure_data_path",
",",
"headers",
"=",
"self",
".",
"HEADERS",
")",
"throw_if_bad_res... | Delete a file at the given secure data path | [
"Delete",
"a",
"file",
"at",
"the",
"given",
"secure",
"data",
"path"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L332-L337 |
14,491 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_file_metadata | def get_file_metadata(self, secure_data_path, version=None):
"""Get just the metadata for a file, not the content"""
if not version:
version = "CURRENT"
payload = {'versionId': str(version)}
secret_resp = head_with_retry(str.join('', [self.cerberus_url, '/v1/secure-file/', secure_data_path]),
params=payload, headers=self.HEADERS)
throw_if_bad_response(secret_resp)
return secret_resp.headers | python | def get_file_metadata(self, secure_data_path, version=None):
if not version:
version = "CURRENT"
payload = {'versionId': str(version)}
secret_resp = head_with_retry(str.join('', [self.cerberus_url, '/v1/secure-file/', secure_data_path]),
params=payload, headers=self.HEADERS)
throw_if_bad_response(secret_resp)
return secret_resp.headers | [
"def",
"get_file_metadata",
"(",
"self",
",",
"secure_data_path",
",",
"version",
"=",
"None",
")",
":",
"if",
"not",
"version",
":",
"version",
"=",
"\"CURRENT\"",
"payload",
"=",
"{",
"'versionId'",
":",
"str",
"(",
"version",
")",
"}",
"secret_resp",
"=... | Get just the metadata for a file, not the content | [
"Get",
"just",
"the",
"metadata",
"for",
"a",
"file",
"not",
"the",
"content"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L339-L350 |
14,492 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient._parse_metadata_filename | def _parse_metadata_filename(self, metadata):
"""
Parse the header metadata to pull out the filename and then store it under the key 'filename'
"""
index = metadata['Content-Disposition'].index('=')+1
metadata['filename'] = metadata['Content-Disposition'][index:].replace('"', '')
return metadata | python | def _parse_metadata_filename(self, metadata):
index = metadata['Content-Disposition'].index('=')+1
metadata['filename'] = metadata['Content-Disposition'][index:].replace('"', '')
return metadata | [
"def",
"_parse_metadata_filename",
"(",
"self",
",",
"metadata",
")",
":",
"index",
"=",
"metadata",
"[",
"'Content-Disposition'",
"]",
".",
"index",
"(",
"'='",
")",
"+",
"1",
"metadata",
"[",
"'filename'",
"]",
"=",
"metadata",
"[",
"'Content-Disposition'",
... | Parse the header metadata to pull out the filename and then store it under the key 'filename' | [
"Parse",
"the",
"header",
"metadata",
"to",
"pull",
"out",
"the",
"filename",
"and",
"then",
"store",
"it",
"under",
"the",
"key",
"filename"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L371-L377 |
14,493 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_file_versions | def get_file_versions(self, secure_data_path, limit=None, offset=None):
"""
Get versions of a particular file
This is just a shim to get_secret_versions
secure_data_path -- full path to the file in the safety deposit box
limit -- Default(100), limits how many records to be returned from the api at once.
offset -- Default(0), used for pagination. Will request records from the given offset.
"""
return self.get_secret_versions(secure_data_path, limit, offset) | python | def get_file_versions(self, secure_data_path, limit=None, offset=None):
return self.get_secret_versions(secure_data_path, limit, offset) | [
"def",
"get_file_versions",
"(",
"self",
",",
"secure_data_path",
",",
"limit",
"=",
"None",
",",
"offset",
"=",
"None",
")",
":",
"return",
"self",
".",
"get_secret_versions",
"(",
"secure_data_path",
",",
"limit",
",",
"offset",
")"
] | Get versions of a particular file
This is just a shim to get_secret_versions
secure_data_path -- full path to the file in the safety deposit box
limit -- Default(100), limits how many records to be returned from the api at once.
offset -- Default(0), used for pagination. Will request records from the given offset. | [
"Get",
"versions",
"of",
"a",
"particular",
"file",
"This",
"is",
"just",
"a",
"shim",
"to",
"get_secret_versions"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L405-L415 |
14,494 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient._get_all_file_versions | def _get_all_file_versions(self, secure_data_path, limit=None):
"""
Convenience function that returns a generator yielding the contents of all versions of
a file and its version info
secure_data_path -- full path to the file in the safety deposit box
limit -- Default(100), limits how many records to be returned from the api at once.
"""
for secret in self._get_all_file_version_ids(secure_data_path, limit):
yield {'secret': self.get_file_data(secure_data_path, version=secret['id']),
'version': secret} | python | def _get_all_file_versions(self, secure_data_path, limit=None):
for secret in self._get_all_file_version_ids(secure_data_path, limit):
yield {'secret': self.get_file_data(secure_data_path, version=secret['id']),
'version': secret} | [
"def",
"_get_all_file_versions",
"(",
"self",
",",
"secure_data_path",
",",
"limit",
"=",
"None",
")",
":",
"for",
"secret",
"in",
"self",
".",
"_get_all_file_version_ids",
"(",
"secure_data_path",
",",
"limit",
")",
":",
"yield",
"{",
"'secret'",
":",
"self",... | Convenience function that returns a generator yielding the contents of all versions of
a file and its version info
secure_data_path -- full path to the file in the safety deposit box
limit -- Default(100), limits how many records to be returned from the api at once. | [
"Convenience",
"function",
"that",
"returns",
"a",
"generator",
"yielding",
"the",
"contents",
"of",
"all",
"versions",
"of",
"a",
"file",
"and",
"its",
"version",
"info"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L433-L443 |
14,495 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.list_files | def list_files(self, secure_data_path, limit=None, offset=None):
"""Return the list of files in the path. May need to be paginated"""
# Make sure that limit and offset are in range.
# Set the normal defaults
if not limit or limit <= 0:
limit = 100
if not offset or offset < 0:
offset = 0
payload = {'limit': str(limit), 'offset': str(offset)}
# Because of the addition of versionId and the way URLs are constructed, secure_data_path should
# always end in a '/'.
secure_data_path = self._add_slash(secure_data_path)
secret_resp = get_with_retry(self.cerberus_url + '/v1/secure-files/' + secure_data_path,
params=payload, headers=self.HEADERS)
throw_if_bad_response(secret_resp)
return secret_resp.json() | python | def list_files(self, secure_data_path, limit=None, offset=None):
# Make sure that limit and offset are in range.
# Set the normal defaults
if not limit or limit <= 0:
limit = 100
if not offset or offset < 0:
offset = 0
payload = {'limit': str(limit), 'offset': str(offset)}
# Because of the addition of versionId and the way URLs are constructed, secure_data_path should
# always end in a '/'.
secure_data_path = self._add_slash(secure_data_path)
secret_resp = get_with_retry(self.cerberus_url + '/v1/secure-files/' + secure_data_path,
params=payload, headers=self.HEADERS)
throw_if_bad_response(secret_resp)
return secret_resp.json() | [
"def",
"list_files",
"(",
"self",
",",
"secure_data_path",
",",
"limit",
"=",
"None",
",",
"offset",
"=",
"None",
")",
":",
"# Make sure that limit and offset are in range.",
"# Set the normal defaults",
"if",
"not",
"limit",
"or",
"limit",
"<=",
"0",
":",
"limit"... | Return the list of files in the path. May need to be paginated | [
"Return",
"the",
"list",
"of",
"files",
"in",
"the",
"path",
".",
"May",
"need",
"to",
"be",
"paginated"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L445-L464 |
14,496 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.put_file | def put_file(self, secure_data_path, filehandle, content_type=None):
"""
Upload a file to a secure data path provided
Keyword arguments:
secure_data_path -- full path in the safety deposit box that contains the file key to store things under
filehandle -- Pass an opened filehandle to the file you want to upload.
Make sure that the file was opened in binary mode, otherwise the size calculations
can be off for text files.
content_type -- Optional. Set the Mime type of the file you're uploading.
"""
# Parse out the filename from the path
filename = secure_data_path.rsplit('/', 1)
if content_type:
data = {'file-content': (filename, filehandle, content_type)}
else:
data = {'file-content': (filename, filehandle)}
headers = self.HEADERS.copy()
if 'Content-Type' in headers:
headers.__delitem__('Content-Type')
secret_resp = post_with_retry(self.cerberus_url + '/v1/secure-file/' + secure_data_path,
files=data, headers=headers)
throw_if_bad_response(secret_resp)
return secret_resp | python | def put_file(self, secure_data_path, filehandle, content_type=None):
# Parse out the filename from the path
filename = secure_data_path.rsplit('/', 1)
if content_type:
data = {'file-content': (filename, filehandle, content_type)}
else:
data = {'file-content': (filename, filehandle)}
headers = self.HEADERS.copy()
if 'Content-Type' in headers:
headers.__delitem__('Content-Type')
secret_resp = post_with_retry(self.cerberus_url + '/v1/secure-file/' + secure_data_path,
files=data, headers=headers)
throw_if_bad_response(secret_resp)
return secret_resp | [
"def",
"put_file",
"(",
"self",
",",
"secure_data_path",
",",
"filehandle",
",",
"content_type",
"=",
"None",
")",
":",
"# Parse out the filename from the path",
"filename",
"=",
"secure_data_path",
".",
"rsplit",
"(",
"'/'",
",",
"1",
")",
"if",
"content_type",
... | Upload a file to a secure data path provided
Keyword arguments:
secure_data_path -- full path in the safety deposit box that contains the file key to store things under
filehandle -- Pass an opened filehandle to the file you want to upload.
Make sure that the file was opened in binary mode, otherwise the size calculations
can be off for text files.
content_type -- Optional. Set the Mime type of the file you're uploading. | [
"Upload",
"a",
"file",
"to",
"a",
"secure",
"data",
"path",
"provided"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L466-L493 |
14,497 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.get_secret_versions | def get_secret_versions(self, secure_data_path, limit=None, offset=None):
"""
Get versions of a particular secret key
secure_data_path -- full path to the key in the safety deposit box
limit -- Default(100), limits how many records to be returned from the api at once.
offset -- Default(0), used for pagination. Will request records from the given offset.
"""
# Make sure that limit and offset are in range.
# Set the normal defaults
if not limit or limit <= 0:
limit = 100
if not offset or offset < 0:
offset = 0
payload = {'limit': str(limit), 'offset': str(offset)}
secret_resp = get_with_retry(str.join('', [self.cerberus_url, '/v1/secret-versions/', secure_data_path]),
params=payload, headers=self.HEADERS)
throw_if_bad_response(secret_resp)
return secret_resp.json() | python | def get_secret_versions(self, secure_data_path, limit=None, offset=None):
# Make sure that limit and offset are in range.
# Set the normal defaults
if not limit or limit <= 0:
limit = 100
if not offset or offset < 0:
offset = 0
payload = {'limit': str(limit), 'offset': str(offset)}
secret_resp = get_with_retry(str.join('', [self.cerberus_url, '/v1/secret-versions/', secure_data_path]),
params=payload, headers=self.HEADERS)
throw_if_bad_response(secret_resp)
return secret_resp.json() | [
"def",
"get_secret_versions",
"(",
"self",
",",
"secure_data_path",
",",
"limit",
"=",
"None",
",",
"offset",
"=",
"None",
")",
":",
"# Make sure that limit and offset are in range.",
"# Set the normal defaults",
"if",
"not",
"limit",
"or",
"limit",
"<=",
"0",
":",
... | Get versions of a particular secret key
secure_data_path -- full path to the key in the safety deposit box
limit -- Default(100), limits how many records to be returned from the api at once.
offset -- Default(0), used for pagination. Will request records from the given offset. | [
"Get",
"versions",
"of",
"a",
"particular",
"secret",
"key"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L564-L585 |
14,498 | Nike-Inc/cerberus-python-client | cerberus/client.py | CerberusClient.list_secrets | def list_secrets(self, secure_data_path):
"""Return json secrets based on the secure_data_path, this will list keys in a folder"""
# Because of the addition of versionId and the way URLs are constructed, secure_data_path should
# always end in a '/'.
secure_data_path = self._add_slash(secure_data_path)
secret_resp = get_with_retry(self.cerberus_url + '/v1/secret/' + secure_data_path + '?list=true',
headers=self.HEADERS)
throw_if_bad_response(secret_resp)
return secret_resp.json() | python | def list_secrets(self, secure_data_path):
# Because of the addition of versionId and the way URLs are constructed, secure_data_path should
# always end in a '/'.
secure_data_path = self._add_slash(secure_data_path)
secret_resp = get_with_retry(self.cerberus_url + '/v1/secret/' + secure_data_path + '?list=true',
headers=self.HEADERS)
throw_if_bad_response(secret_resp)
return secret_resp.json() | [
"def",
"list_secrets",
"(",
"self",
",",
"secure_data_path",
")",
":",
"# Because of the addition of versionId and the way URLs are constructed, secure_data_path should",
"# always end in a '/'.",
"secure_data_path",
"=",
"self",
".",
"_add_slash",
"(",
"secure_data_path",
")",
... | Return json secrets based on the secure_data_path, this will list keys in a folder | [
"Return",
"json",
"secrets",
"based",
"on",
"the",
"secure_data_path",
"this",
"will",
"list",
"keys",
"in",
"a",
"folder"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/client.py#L613-L622 |
14,499 | Nike-Inc/cerberus-python-client | cerberus/aws_auth.py | AWSAuth._get_v4_signed_headers | def _get_v4_signed_headers(self):
"""Returns V4 signed get-caller-identity request headers"""
if self.aws_session is None:
boto_session = session.Session()
creds = boto_session.get_credentials()
else:
creds = self.aws_session.get_credentials()
if creds is None:
raise CerberusClientException("Unable to locate AWS credentials")
readonly_credentials = creds.get_frozen_credentials()
# hardcode get-caller-identity request
data = OrderedDict((('Action','GetCallerIdentity'), ('Version', '2011-06-15')))
url = 'https://sts.{}.amazonaws.com/'.format(self.region)
request_object = awsrequest.AWSRequest(method='POST', url=url, data=data)
signer = auth.SigV4Auth(readonly_credentials, 'sts', self.region)
signer.add_auth(request_object)
return request_object.headers | python | def _get_v4_signed_headers(self):
if self.aws_session is None:
boto_session = session.Session()
creds = boto_session.get_credentials()
else:
creds = self.aws_session.get_credentials()
if creds is None:
raise CerberusClientException("Unable to locate AWS credentials")
readonly_credentials = creds.get_frozen_credentials()
# hardcode get-caller-identity request
data = OrderedDict((('Action','GetCallerIdentity'), ('Version', '2011-06-15')))
url = 'https://sts.{}.amazonaws.com/'.format(self.region)
request_object = awsrequest.AWSRequest(method='POST', url=url, data=data)
signer = auth.SigV4Auth(readonly_credentials, 'sts', self.region)
signer.add_auth(request_object)
return request_object.headers | [
"def",
"_get_v4_signed_headers",
"(",
"self",
")",
":",
"if",
"self",
".",
"aws_session",
"is",
"None",
":",
"boto_session",
"=",
"session",
".",
"Session",
"(",
")",
"creds",
"=",
"boto_session",
".",
"get_credentials",
"(",
")",
"else",
":",
"creds",
"="... | Returns V4 signed get-caller-identity request headers | [
"Returns",
"V4",
"signed",
"get",
"-",
"caller",
"-",
"identity",
"request",
"headers"
] | ef38356822e722fcb6a6ed4a1b38a5b493e753ae | https://github.com/Nike-Inc/cerberus-python-client/blob/ef38356822e722fcb6a6ed4a1b38a5b493e753ae/cerberus/aws_auth.py#L39-L57 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.