repo
stringlengths
7
55
path
stringlengths
4
223
func_name
stringlengths
1
134
original_string
stringlengths
75
104k
language
stringclasses
1 value
code
stringlengths
75
104k
code_tokens
listlengths
19
28.4k
docstring
stringlengths
1
46.9k
docstring_tokens
listlengths
1
1.97k
sha
stringlengths
40
40
url
stringlengths
87
315
partition
stringclasses
1 value
apache/incubator-mxnet
python/mxnet/module/base_module.py
BaseModule.bind
def bind(self, data_shapes, label_shapes=None, for_training=True, inputs_need_grad=False, force_rebind=False, shared_module=None, grad_req='write'): """Binds the symbols to construct executors. This is necessary before one can perform computation with the module. Parameters ---------- data_shapes : list of (str, tuple) or DataDesc objects Typically is ``data_iter.provide_data``. Can also be a list of (data name, data shape). label_shapes : list of (str, tuple) or DataDesc objects Typically is ``data_iter.provide_label``. Can also be a list of (label name, label shape). for_training : bool Default is ``True``. Whether the executors should be bind for training. inputs_need_grad : bool Default is ``False``. Whether the gradients to the input data need to be computed. Typically this is not needed. But this might be needed when implementing composition of modules. force_rebind : bool Default is ``False``. This function does nothing if the executors are already bound. But with this ``True``, the executors will be forced to rebind. shared_module : Module Default is ``None``. This is used in bucketing. When not ``None``, the shared module essentially corresponds to a different bucket -- a module with different symbol but with the same sets of parameters (e.g. unrolled RNNs with different lengths). grad_req : str, list of str, dict of str to str Requirement for gradient accumulation. Can be 'write', 'add', or 'null' (default to 'write'). Can be specified globally (str) or for each argument (list, dict). Examples -------- >>> # An example of binding symbols. >>> mod.bind(data_shapes=[('data', (1, 10, 10))]) >>> # Assume train_iter is already created. >>> mod.bind(data_shapes=train_iter.provide_data, label_shapes=train_iter.provide_label) """ raise NotImplementedError()
python
def bind(self, data_shapes, label_shapes=None, for_training=True, inputs_need_grad=False, force_rebind=False, shared_module=None, grad_req='write'): """Binds the symbols to construct executors. This is necessary before one can perform computation with the module. Parameters ---------- data_shapes : list of (str, tuple) or DataDesc objects Typically is ``data_iter.provide_data``. Can also be a list of (data name, data shape). label_shapes : list of (str, tuple) or DataDesc objects Typically is ``data_iter.provide_label``. Can also be a list of (label name, label shape). for_training : bool Default is ``True``. Whether the executors should be bind for training. inputs_need_grad : bool Default is ``False``. Whether the gradients to the input data need to be computed. Typically this is not needed. But this might be needed when implementing composition of modules. force_rebind : bool Default is ``False``. This function does nothing if the executors are already bound. But with this ``True``, the executors will be forced to rebind. shared_module : Module Default is ``None``. This is used in bucketing. When not ``None``, the shared module essentially corresponds to a different bucket -- a module with different symbol but with the same sets of parameters (e.g. unrolled RNNs with different lengths). grad_req : str, list of str, dict of str to str Requirement for gradient accumulation. Can be 'write', 'add', or 'null' (default to 'write'). Can be specified globally (str) or for each argument (list, dict). Examples -------- >>> # An example of binding symbols. >>> mod.bind(data_shapes=[('data', (1, 10, 10))]) >>> # Assume train_iter is already created. >>> mod.bind(data_shapes=train_iter.provide_data, label_shapes=train_iter.provide_label) """ raise NotImplementedError()
[ "def", "bind", "(", "self", ",", "data_shapes", ",", "label_shapes", "=", "None", ",", "for_training", "=", "True", ",", "inputs_need_grad", "=", "False", ",", "force_rebind", "=", "False", ",", "shared_module", "=", "None", ",", "grad_req", "=", "'write'", ")", ":", "raise", "NotImplementedError", "(", ")" ]
Binds the symbols to construct executors. This is necessary before one can perform computation with the module. Parameters ---------- data_shapes : list of (str, tuple) or DataDesc objects Typically is ``data_iter.provide_data``. Can also be a list of (data name, data shape). label_shapes : list of (str, tuple) or DataDesc objects Typically is ``data_iter.provide_label``. Can also be a list of (label name, label shape). for_training : bool Default is ``True``. Whether the executors should be bind for training. inputs_need_grad : bool Default is ``False``. Whether the gradients to the input data need to be computed. Typically this is not needed. But this might be needed when implementing composition of modules. force_rebind : bool Default is ``False``. This function does nothing if the executors are already bound. But with this ``True``, the executors will be forced to rebind. shared_module : Module Default is ``None``. This is used in bucketing. When not ``None``, the shared module essentially corresponds to a different bucket -- a module with different symbol but with the same sets of parameters (e.g. unrolled RNNs with different lengths). grad_req : str, list of str, dict of str to str Requirement for gradient accumulation. Can be 'write', 'add', or 'null' (default to 'write'). Can be specified globally (str) or for each argument (list, dict). Examples -------- >>> # An example of binding symbols. >>> mod.bind(data_shapes=[('data', (1, 10, 10))]) >>> # Assume train_iter is already created. >>> mod.bind(data_shapes=train_iter.provide_data, label_shapes=train_iter.provide_label)
[ "Binds", "the", "symbols", "to", "construct", "executors", ".", "This", "is", "necessary", "before", "one", "can", "perform", "computation", "with", "the", "module", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/module/base_module.py#L990-L1029
train
apache/incubator-mxnet
python/mxnet/libinfo.py
find_lib_path
def find_lib_path(): """Find MXNet dynamic library files. Returns ------- lib_path : list(string) List of all found path to the libraries. """ lib_from_env = os.environ.get('MXNET_LIBRARY_PATH') if lib_from_env: if os.path.isfile(lib_from_env): if not os.path.isabs(lib_from_env): logging.warning("MXNET_LIBRARY_PATH should be an absolute path, instead of: %s", lib_from_env) else: if os.name == 'nt': os.environ['PATH'] = os.environ['PATH'] + ';' + os.path.dirname(lib_from_env) return [lib_from_env] else: logging.warning("MXNET_LIBRARY_PATH '%s' doesn't exist", lib_from_env) curr_path = os.path.dirname(os.path.abspath(os.path.expanduser(__file__))) api_path = os.path.join(curr_path, '../../lib/') cmake_build_path = os.path.join(curr_path, '../../build/') dll_path = [curr_path, api_path, cmake_build_path] if os.name == 'nt': dll_path.append(os.path.join(curr_path, '../../build')) vs_configuration = 'Release' if platform.architecture()[0] == '64bit': dll_path.append(os.path.join(curr_path, '../../build', vs_configuration)) dll_path.append(os.path.join(curr_path, '../../windows/x64', vs_configuration)) else: dll_path.append(os.path.join(curr_path, '../../build', vs_configuration)) dll_path.append(os.path.join(curr_path, '../../windows', vs_configuration)) elif os.name == "posix" and os.environ.get('LD_LIBRARY_PATH', None): dll_path[0:0] = [p.strip() for p in os.environ['LD_LIBRARY_PATH'].split(":")] if os.name == 'nt': os.environ['PATH'] = os.path.dirname(__file__) + ';' + os.environ['PATH'] dll_path = [os.path.join(p, 'libmxnet.dll') for p in dll_path] elif platform.system() == 'Darwin': dll_path = [os.path.join(p, 'libmxnet.dylib') for p in dll_path] + \ [os.path.join(p, 'libmxnet.so') for p in dll_path] else: dll_path.append('../../../') dll_path = [os.path.join(p, 'libmxnet.so') for p in dll_path] lib_path = [p for p in dll_path if os.path.exists(p) and os.path.isfile(p)] if len(lib_path) == 0: raise RuntimeError('Cannot find the MXNet library.\n' + 'List of candidates:\n' + str('\n'.join(dll_path))) if os.name == 'nt': os.environ['PATH'] = os.environ['PATH'] + ';' + os.path.dirname(lib_path[0]) return lib_path
python
def find_lib_path(): """Find MXNet dynamic library files. Returns ------- lib_path : list(string) List of all found path to the libraries. """ lib_from_env = os.environ.get('MXNET_LIBRARY_PATH') if lib_from_env: if os.path.isfile(lib_from_env): if not os.path.isabs(lib_from_env): logging.warning("MXNET_LIBRARY_PATH should be an absolute path, instead of: %s", lib_from_env) else: if os.name == 'nt': os.environ['PATH'] = os.environ['PATH'] + ';' + os.path.dirname(lib_from_env) return [lib_from_env] else: logging.warning("MXNET_LIBRARY_PATH '%s' doesn't exist", lib_from_env) curr_path = os.path.dirname(os.path.abspath(os.path.expanduser(__file__))) api_path = os.path.join(curr_path, '../../lib/') cmake_build_path = os.path.join(curr_path, '../../build/') dll_path = [curr_path, api_path, cmake_build_path] if os.name == 'nt': dll_path.append(os.path.join(curr_path, '../../build')) vs_configuration = 'Release' if platform.architecture()[0] == '64bit': dll_path.append(os.path.join(curr_path, '../../build', vs_configuration)) dll_path.append(os.path.join(curr_path, '../../windows/x64', vs_configuration)) else: dll_path.append(os.path.join(curr_path, '../../build', vs_configuration)) dll_path.append(os.path.join(curr_path, '../../windows', vs_configuration)) elif os.name == "posix" and os.environ.get('LD_LIBRARY_PATH', None): dll_path[0:0] = [p.strip() for p in os.environ['LD_LIBRARY_PATH'].split(":")] if os.name == 'nt': os.environ['PATH'] = os.path.dirname(__file__) + ';' + os.environ['PATH'] dll_path = [os.path.join(p, 'libmxnet.dll') for p in dll_path] elif platform.system() == 'Darwin': dll_path = [os.path.join(p, 'libmxnet.dylib') for p in dll_path] + \ [os.path.join(p, 'libmxnet.so') for p in dll_path] else: dll_path.append('../../../') dll_path = [os.path.join(p, 'libmxnet.so') for p in dll_path] lib_path = [p for p in dll_path if os.path.exists(p) and os.path.isfile(p)] if len(lib_path) == 0: raise RuntimeError('Cannot find the MXNet library.\n' + 'List of candidates:\n' + str('\n'.join(dll_path))) if os.name == 'nt': os.environ['PATH'] = os.environ['PATH'] + ';' + os.path.dirname(lib_path[0]) return lib_path
[ "def", "find_lib_path", "(", ")", ":", "lib_from_env", "=", "os", ".", "environ", ".", "get", "(", "'MXNET_LIBRARY_PATH'", ")", "if", "lib_from_env", ":", "if", "os", ".", "path", ".", "isfile", "(", "lib_from_env", ")", ":", "if", "not", "os", ".", "path", ".", "isabs", "(", "lib_from_env", ")", ":", "logging", ".", "warning", "(", "\"MXNET_LIBRARY_PATH should be an absolute path, instead of: %s\"", ",", "lib_from_env", ")", "else", ":", "if", "os", ".", "name", "==", "'nt'", ":", "os", ".", "environ", "[", "'PATH'", "]", "=", "os", ".", "environ", "[", "'PATH'", "]", "+", "';'", "+", "os", ".", "path", ".", "dirname", "(", "lib_from_env", ")", "return", "[", "lib_from_env", "]", "else", ":", "logging", ".", "warning", "(", "\"MXNET_LIBRARY_PATH '%s' doesn't exist\"", ",", "lib_from_env", ")", "curr_path", "=", "os", ".", "path", ".", "dirname", "(", "os", ".", "path", ".", "abspath", "(", "os", ".", "path", ".", "expanduser", "(", "__file__", ")", ")", ")", "api_path", "=", "os", ".", "path", ".", "join", "(", "curr_path", ",", "'../../lib/'", ")", "cmake_build_path", "=", "os", ".", "path", ".", "join", "(", "curr_path", ",", "'../../build/'", ")", "dll_path", "=", "[", "curr_path", ",", "api_path", ",", "cmake_build_path", "]", "if", "os", ".", "name", "==", "'nt'", ":", "dll_path", ".", "append", "(", "os", ".", "path", ".", "join", "(", "curr_path", ",", "'../../build'", ")", ")", "vs_configuration", "=", "'Release'", "if", "platform", ".", "architecture", "(", ")", "[", "0", "]", "==", "'64bit'", ":", "dll_path", ".", "append", "(", "os", ".", "path", ".", "join", "(", "curr_path", ",", "'../../build'", ",", "vs_configuration", ")", ")", "dll_path", ".", "append", "(", "os", ".", "path", ".", "join", "(", "curr_path", ",", "'../../windows/x64'", ",", "vs_configuration", ")", ")", "else", ":", "dll_path", ".", "append", "(", "os", ".", "path", ".", "join", "(", "curr_path", ",", "'../../build'", ",", "vs_configuration", ")", ")", "dll_path", ".", "append", "(", "os", ".", "path", ".", "join", "(", "curr_path", ",", "'../../windows'", ",", "vs_configuration", ")", ")", "elif", "os", ".", "name", "==", "\"posix\"", "and", "os", ".", "environ", ".", "get", "(", "'LD_LIBRARY_PATH'", ",", "None", ")", ":", "dll_path", "[", "0", ":", "0", "]", "=", "[", "p", ".", "strip", "(", ")", "for", "p", "in", "os", ".", "environ", "[", "'LD_LIBRARY_PATH'", "]", ".", "split", "(", "\":\"", ")", "]", "if", "os", ".", "name", "==", "'nt'", ":", "os", ".", "environ", "[", "'PATH'", "]", "=", "os", ".", "path", ".", "dirname", "(", "__file__", ")", "+", "';'", "+", "os", ".", "environ", "[", "'PATH'", "]", "dll_path", "=", "[", "os", ".", "path", ".", "join", "(", "p", ",", "'libmxnet.dll'", ")", "for", "p", "in", "dll_path", "]", "elif", "platform", ".", "system", "(", ")", "==", "'Darwin'", ":", "dll_path", "=", "[", "os", ".", "path", ".", "join", "(", "p", ",", "'libmxnet.dylib'", ")", "for", "p", "in", "dll_path", "]", "+", "[", "os", ".", "path", ".", "join", "(", "p", ",", "'libmxnet.so'", ")", "for", "p", "in", "dll_path", "]", "else", ":", "dll_path", ".", "append", "(", "'../../../'", ")", "dll_path", "=", "[", "os", ".", "path", ".", "join", "(", "p", ",", "'libmxnet.so'", ")", "for", "p", "in", "dll_path", "]", "lib_path", "=", "[", "p", "for", "p", "in", "dll_path", "if", "os", ".", "path", ".", "exists", "(", "p", ")", "and", "os", ".", "path", ".", "isfile", "(", "p", ")", "]", "if", "len", "(", "lib_path", ")", "==", "0", ":", "raise", "RuntimeError", "(", "'Cannot find the MXNet library.\\n'", "+", "'List of candidates:\\n'", "+", "str", "(", "'\\n'", ".", "join", "(", "dll_path", ")", ")", ")", "if", "os", ".", "name", "==", "'nt'", ":", "os", ".", "environ", "[", "'PATH'", "]", "=", "os", ".", "environ", "[", "'PATH'", "]", "+", "';'", "+", "os", ".", "path", ".", "dirname", "(", "lib_path", "[", "0", "]", ")", "return", "lib_path" ]
Find MXNet dynamic library files. Returns ------- lib_path : list(string) List of all found path to the libraries.
[ "Find", "MXNet", "dynamic", "library", "files", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/libinfo.py#L26-L77
train
apache/incubator-mxnet
python/mxnet/libinfo.py
find_include_path
def find_include_path(): """Find MXNet included header files. Returns ------- incl_path : string Path to the header files. """ incl_from_env = os.environ.get('MXNET_INCLUDE_PATH') if incl_from_env: if os.path.isdir(incl_from_env): if not os.path.isabs(incl_from_env): logging.warning("MXNET_INCLUDE_PATH should be an absolute path, instead of: %s", incl_from_env) else: return incl_from_env else: logging.warning("MXNET_INCLUDE_PATH '%s' doesn't exist", incl_from_env) curr_path = os.path.dirname(os.path.abspath(os.path.expanduser(__file__))) # include path in pip package pip_incl_path = os.path.join(curr_path, 'include/') if os.path.isdir(pip_incl_path): return pip_incl_path else: # include path if build from source src_incl_path = os.path.join(curr_path, '../../include/') if os.path.isdir(src_incl_path): return src_incl_path else: raise RuntimeError('Cannot find the MXNet include path in either ' + pip_incl_path + ' or ' + src_incl_path + '\n')
python
def find_include_path(): """Find MXNet included header files. Returns ------- incl_path : string Path to the header files. """ incl_from_env = os.environ.get('MXNET_INCLUDE_PATH') if incl_from_env: if os.path.isdir(incl_from_env): if not os.path.isabs(incl_from_env): logging.warning("MXNET_INCLUDE_PATH should be an absolute path, instead of: %s", incl_from_env) else: return incl_from_env else: logging.warning("MXNET_INCLUDE_PATH '%s' doesn't exist", incl_from_env) curr_path = os.path.dirname(os.path.abspath(os.path.expanduser(__file__))) # include path in pip package pip_incl_path = os.path.join(curr_path, 'include/') if os.path.isdir(pip_incl_path): return pip_incl_path else: # include path if build from source src_incl_path = os.path.join(curr_path, '../../include/') if os.path.isdir(src_incl_path): return src_incl_path else: raise RuntimeError('Cannot find the MXNet include path in either ' + pip_incl_path + ' or ' + src_incl_path + '\n')
[ "def", "find_include_path", "(", ")", ":", "incl_from_env", "=", "os", ".", "environ", ".", "get", "(", "'MXNET_INCLUDE_PATH'", ")", "if", "incl_from_env", ":", "if", "os", ".", "path", ".", "isdir", "(", "incl_from_env", ")", ":", "if", "not", "os", ".", "path", ".", "isabs", "(", "incl_from_env", ")", ":", "logging", ".", "warning", "(", "\"MXNET_INCLUDE_PATH should be an absolute path, instead of: %s\"", ",", "incl_from_env", ")", "else", ":", "return", "incl_from_env", "else", ":", "logging", ".", "warning", "(", "\"MXNET_INCLUDE_PATH '%s' doesn't exist\"", ",", "incl_from_env", ")", "curr_path", "=", "os", ".", "path", ".", "dirname", "(", "os", ".", "path", ".", "abspath", "(", "os", ".", "path", ".", "expanduser", "(", "__file__", ")", ")", ")", "# include path in pip package", "pip_incl_path", "=", "os", ".", "path", ".", "join", "(", "curr_path", ",", "'include/'", ")", "if", "os", ".", "path", ".", "isdir", "(", "pip_incl_path", ")", ":", "return", "pip_incl_path", "else", ":", "# include path if build from source", "src_incl_path", "=", "os", ".", "path", ".", "join", "(", "curr_path", ",", "'../../include/'", ")", "if", "os", ".", "path", ".", "isdir", "(", "src_incl_path", ")", ":", "return", "src_incl_path", "else", ":", "raise", "RuntimeError", "(", "'Cannot find the MXNet include path in either '", "+", "pip_incl_path", "+", "' or '", "+", "src_incl_path", "+", "'\\n'", ")" ]
Find MXNet included header files. Returns ------- incl_path : string Path to the header files.
[ "Find", "MXNet", "included", "header", "files", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/libinfo.py#L79-L110
train
apache/incubator-mxnet
example/ctc/captcha_generator.py
CaptchaGen.image
def image(self, captcha_str): """Generate a greyscale captcha image representing number string Parameters ---------- captcha_str: str string a characters for captcha image Returns ------- numpy.ndarray Generated greyscale image in np.ndarray float type with values normalized to [0, 1] """ img = self.captcha.generate(captcha_str) img = np.fromstring(img.getvalue(), dtype='uint8') img = cv2.imdecode(img, cv2.IMREAD_GRAYSCALE) img = cv2.resize(img, (self.h, self.w)) img = img.transpose(1, 0) img = np.multiply(img, 1 / 255.0) return img
python
def image(self, captcha_str): """Generate a greyscale captcha image representing number string Parameters ---------- captcha_str: str string a characters for captcha image Returns ------- numpy.ndarray Generated greyscale image in np.ndarray float type with values normalized to [0, 1] """ img = self.captcha.generate(captcha_str) img = np.fromstring(img.getvalue(), dtype='uint8') img = cv2.imdecode(img, cv2.IMREAD_GRAYSCALE) img = cv2.resize(img, (self.h, self.w)) img = img.transpose(1, 0) img = np.multiply(img, 1 / 255.0) return img
[ "def", "image", "(", "self", ",", "captcha_str", ")", ":", "img", "=", "self", ".", "captcha", ".", "generate", "(", "captcha_str", ")", "img", "=", "np", ".", "fromstring", "(", "img", ".", "getvalue", "(", ")", ",", "dtype", "=", "'uint8'", ")", "img", "=", "cv2", ".", "imdecode", "(", "img", ",", "cv2", ".", "IMREAD_GRAYSCALE", ")", "img", "=", "cv2", ".", "resize", "(", "img", ",", "(", "self", ".", "h", ",", "self", ".", "w", ")", ")", "img", "=", "img", ".", "transpose", "(", "1", ",", "0", ")", "img", "=", "np", ".", "multiply", "(", "img", ",", "1", "/", "255.0", ")", "return", "img" ]
Generate a greyscale captcha image representing number string Parameters ---------- captcha_str: str string a characters for captcha image Returns ------- numpy.ndarray Generated greyscale image in np.ndarray float type with values normalized to [0, 1]
[ "Generate", "a", "greyscale", "captcha", "image", "representing", "number", "string" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ctc/captcha_generator.py#L48-L67
train
apache/incubator-mxnet
example/ctc/captcha_generator.py
DigitCaptcha.get_rand
def get_rand(num_digit_min, num_digit_max): """Generates a character string of digits. Number of digits are between self.num_digit_min and self.num_digit_max Returns ------- str """ buf = "" max_len = random.randint(num_digit_min, num_digit_max) for i in range(max_len): buf += str(random.randint(0, 9)) return buf
python
def get_rand(num_digit_min, num_digit_max): """Generates a character string of digits. Number of digits are between self.num_digit_min and self.num_digit_max Returns ------- str """ buf = "" max_len = random.randint(num_digit_min, num_digit_max) for i in range(max_len): buf += str(random.randint(0, 9)) return buf
[ "def", "get_rand", "(", "num_digit_min", ",", "num_digit_max", ")", ":", "buf", "=", "\"\"", "max_len", "=", "random", ".", "randint", "(", "num_digit_min", ",", "num_digit_max", ")", "for", "i", "in", "range", "(", "max_len", ")", ":", "buf", "+=", "str", "(", "random", ".", "randint", "(", "0", ",", "9", ")", ")", "return", "buf" ]
Generates a character string of digits. Number of digits are between self.num_digit_min and self.num_digit_max Returns ------- str
[ "Generates", "a", "character", "string", "of", "digits", ".", "Number", "of", "digits", "are", "between", "self", ".", "num_digit_min", "and", "self", ".", "num_digit_max", "Returns", "-------", "str" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ctc/captcha_generator.py#L113-L124
train
apache/incubator-mxnet
example/ctc/captcha_generator.py
DigitCaptcha._gen_sample
def _gen_sample(self): """Generate a random captcha image sample Returns ------- (numpy.ndarray, str) Tuple of image (numpy ndarray) and character string of digits used to generate the image """ num_str = self.get_rand(self.num_digit_min, self.num_digit_max) return self.captcha.image(num_str), num_str
python
def _gen_sample(self): """Generate a random captcha image sample Returns ------- (numpy.ndarray, str) Tuple of image (numpy ndarray) and character string of digits used to generate the image """ num_str = self.get_rand(self.num_digit_min, self.num_digit_max) return self.captcha.image(num_str), num_str
[ "def", "_gen_sample", "(", "self", ")", ":", "num_str", "=", "self", ".", "get_rand", "(", "self", ".", "num_digit_min", ",", "self", ".", "num_digit_max", ")", "return", "self", ".", "captcha", ".", "image", "(", "num_str", ")", ",", "num_str" ]
Generate a random captcha image sample Returns ------- (numpy.ndarray, str) Tuple of image (numpy ndarray) and character string of digits used to generate the image
[ "Generate", "a", "random", "captcha", "image", "sample", "Returns", "-------", "(", "numpy", ".", "ndarray", "str", ")", "Tuple", "of", "image", "(", "numpy", "ndarray", ")", "and", "character", "string", "of", "digits", "used", "to", "generate", "the", "image" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ctc/captcha_generator.py#L126-L134
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Optimizer.register
def register(klass): """Registers a new optimizer. Once an optimizer is registered, we can create an instance of this optimizer with `create_optimizer` later. Examples -------- >>> @mx.optimizer.Optimizer.register ... class MyOptimizer(mx.optimizer.Optimizer): ... pass >>> optim = mx.optimizer.Optimizer.create_optimizer('MyOptimizer') >>> print(type(optim)) <class '__main__.MyOptimizer'> """ assert(isinstance(klass, type)) name = klass.__name__.lower() if name in Optimizer.opt_registry: warnings.warn('WARNING: New optimizer %s.%s is overriding ' 'existing optimizer %s.%s' % (klass.__module__, klass.__name__, Optimizer.opt_registry[name].__module__, Optimizer.opt_registry[name].__name__)) Optimizer.opt_registry[name] = klass return klass
python
def register(klass): """Registers a new optimizer. Once an optimizer is registered, we can create an instance of this optimizer with `create_optimizer` later. Examples -------- >>> @mx.optimizer.Optimizer.register ... class MyOptimizer(mx.optimizer.Optimizer): ... pass >>> optim = mx.optimizer.Optimizer.create_optimizer('MyOptimizer') >>> print(type(optim)) <class '__main__.MyOptimizer'> """ assert(isinstance(klass, type)) name = klass.__name__.lower() if name in Optimizer.opt_registry: warnings.warn('WARNING: New optimizer %s.%s is overriding ' 'existing optimizer %s.%s' % (klass.__module__, klass.__name__, Optimizer.opt_registry[name].__module__, Optimizer.opt_registry[name].__name__)) Optimizer.opt_registry[name] = klass return klass
[ "def", "register", "(", "klass", ")", ":", "assert", "(", "isinstance", "(", "klass", ",", "type", ")", ")", "name", "=", "klass", ".", "__name__", ".", "lower", "(", ")", "if", "name", "in", "Optimizer", ".", "opt_registry", ":", "warnings", ".", "warn", "(", "'WARNING: New optimizer %s.%s is overriding '", "'existing optimizer %s.%s'", "%", "(", "klass", ".", "__module__", ",", "klass", ".", "__name__", ",", "Optimizer", ".", "opt_registry", "[", "name", "]", ".", "__module__", ",", "Optimizer", ".", "opt_registry", "[", "name", "]", ".", "__name__", ")", ")", "Optimizer", ".", "opt_registry", "[", "name", "]", "=", "klass", "return", "klass" ]
Registers a new optimizer. Once an optimizer is registered, we can create an instance of this optimizer with `create_optimizer` later. Examples -------- >>> @mx.optimizer.Optimizer.register ... class MyOptimizer(mx.optimizer.Optimizer): ... pass >>> optim = mx.optimizer.Optimizer.create_optimizer('MyOptimizer') >>> print(type(optim)) <class '__main__.MyOptimizer'>
[ "Registers", "a", "new", "optimizer", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L129-L154
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Optimizer.create_optimizer
def create_optimizer(name, **kwargs): """Instantiates an optimizer with a given name and kwargs. .. note:: We can use the alias `create` for ``Optimizer.create_optimizer``. Parameters ---------- name: str Name of the optimizer. Should be the name of a subclass of Optimizer. Case insensitive. kwargs: dict Parameters for the optimizer. Returns ------- Optimizer An instantiated optimizer. Examples -------- >>> sgd = mx.optimizer.Optimizer.create_optimizer('sgd') >>> type(sgd) <class 'mxnet.optimizer.SGD'> >>> adam = mx.optimizer.create('adam', learning_rate=.1) >>> type(adam) <class 'mxnet.optimizer.Adam'> """ if name.lower() in Optimizer.opt_registry: return Optimizer.opt_registry[name.lower()](**kwargs) else: raise ValueError('Cannot find optimizer %s' % name)
python
def create_optimizer(name, **kwargs): """Instantiates an optimizer with a given name and kwargs. .. note:: We can use the alias `create` for ``Optimizer.create_optimizer``. Parameters ---------- name: str Name of the optimizer. Should be the name of a subclass of Optimizer. Case insensitive. kwargs: dict Parameters for the optimizer. Returns ------- Optimizer An instantiated optimizer. Examples -------- >>> sgd = mx.optimizer.Optimizer.create_optimizer('sgd') >>> type(sgd) <class 'mxnet.optimizer.SGD'> >>> adam = mx.optimizer.create('adam', learning_rate=.1) >>> type(adam) <class 'mxnet.optimizer.Adam'> """ if name.lower() in Optimizer.opt_registry: return Optimizer.opt_registry[name.lower()](**kwargs) else: raise ValueError('Cannot find optimizer %s' % name)
[ "def", "create_optimizer", "(", "name", ",", "*", "*", "kwargs", ")", ":", "if", "name", ".", "lower", "(", ")", "in", "Optimizer", ".", "opt_registry", ":", "return", "Optimizer", ".", "opt_registry", "[", "name", ".", "lower", "(", ")", "]", "(", "*", "*", "kwargs", ")", "else", ":", "raise", "ValueError", "(", "'Cannot find optimizer %s'", "%", "name", ")" ]
Instantiates an optimizer with a given name and kwargs. .. note:: We can use the alias `create` for ``Optimizer.create_optimizer``. Parameters ---------- name: str Name of the optimizer. Should be the name of a subclass of Optimizer. Case insensitive. kwargs: dict Parameters for the optimizer. Returns ------- Optimizer An instantiated optimizer. Examples -------- >>> sgd = mx.optimizer.Optimizer.create_optimizer('sgd') >>> type(sgd) <class 'mxnet.optimizer.SGD'> >>> adam = mx.optimizer.create('adam', learning_rate=.1) >>> type(adam) <class 'mxnet.optimizer.Adam'>
[ "Instantiates", "an", "optimizer", "with", "a", "given", "name", "and", "kwargs", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L157-L188
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Optimizer.create_state_multi_precision
def create_state_multi_precision(self, index, weight): """Creates auxiliary state for a given weight, including FP32 high precision copy if original weight is FP16. This method is provided to perform automatic mixed precision training for optimizers that do not support it themselves. Parameters ---------- index : int An unique index to identify the weight. weight : NDArray The weight. Returns ------- state : any obj The state associated with the weight. """ weight_master_copy = None if self.multi_precision and weight.dtype == numpy.float16: weight_master_copy = weight.astype(numpy.float32) return (weight_master_copy,) + (self.create_state(index, weight_master_copy),) if weight.dtype == numpy.float16 and not self.multi_precision: warnings.warn("Accumulating with float16 in optimizer can lead to " "poor accuracy or slow convergence. " "Consider using multi_precision=True option of the " "optimizer") return self.create_state(index, weight)
python
def create_state_multi_precision(self, index, weight): """Creates auxiliary state for a given weight, including FP32 high precision copy if original weight is FP16. This method is provided to perform automatic mixed precision training for optimizers that do not support it themselves. Parameters ---------- index : int An unique index to identify the weight. weight : NDArray The weight. Returns ------- state : any obj The state associated with the weight. """ weight_master_copy = None if self.multi_precision and weight.dtype == numpy.float16: weight_master_copy = weight.astype(numpy.float32) return (weight_master_copy,) + (self.create_state(index, weight_master_copy),) if weight.dtype == numpy.float16 and not self.multi_precision: warnings.warn("Accumulating with float16 in optimizer can lead to " "poor accuracy or slow convergence. " "Consider using multi_precision=True option of the " "optimizer") return self.create_state(index, weight)
[ "def", "create_state_multi_precision", "(", "self", ",", "index", ",", "weight", ")", ":", "weight_master_copy", "=", "None", "if", "self", ".", "multi_precision", "and", "weight", ".", "dtype", "==", "numpy", ".", "float16", ":", "weight_master_copy", "=", "weight", ".", "astype", "(", "numpy", ".", "float32", ")", "return", "(", "weight_master_copy", ",", ")", "+", "(", "self", ".", "create_state", "(", "index", ",", "weight_master_copy", ")", ",", ")", "if", "weight", ".", "dtype", "==", "numpy", ".", "float16", "and", "not", "self", ".", "multi_precision", ":", "warnings", ".", "warn", "(", "\"Accumulating with float16 in optimizer can lead to \"", "\"poor accuracy or slow convergence. \"", "\"Consider using multi_precision=True option of the \"", "\"optimizer\"", ")", "return", "self", ".", "create_state", "(", "index", ",", "weight", ")" ]
Creates auxiliary state for a given weight, including FP32 high precision copy if original weight is FP16. This method is provided to perform automatic mixed precision training for optimizers that do not support it themselves. Parameters ---------- index : int An unique index to identify the weight. weight : NDArray The weight. Returns ------- state : any obj The state associated with the weight.
[ "Creates", "auxiliary", "state", "for", "a", "given", "weight", "including", "FP32", "high", "precision", "copy", "if", "original", "weight", "is", "FP16", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L218-L246
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Optimizer.update_multi_precision
def update_multi_precision(self, index, weight, grad, state): """Updates the given parameter using the corresponding gradient and state. Mixed precision version. Parameters ---------- index : int The unique index of the parameter into the individual learning rates and weight decays. Learning rates and weight decay may be set via `set_lr_mult()` and `set_wd_mult()`, respectively. weight : NDArray The parameter to be updated. grad : NDArray The gradient of the objective with respect to this parameter. state : any obj The state returned by `create_state()`. """ if self.multi_precision and weight.dtype == numpy.float16: # Wrapper for mixed precision weight_master_copy = state[0] original_state = state[1] grad32 = grad.astype(numpy.float32) self.update(index, weight_master_copy, grad32, original_state) cast(weight_master_copy, dtype=weight.dtype, out=weight) else: self.update(index, weight, grad, state)
python
def update_multi_precision(self, index, weight, grad, state): """Updates the given parameter using the corresponding gradient and state. Mixed precision version. Parameters ---------- index : int The unique index of the parameter into the individual learning rates and weight decays. Learning rates and weight decay may be set via `set_lr_mult()` and `set_wd_mult()`, respectively. weight : NDArray The parameter to be updated. grad : NDArray The gradient of the objective with respect to this parameter. state : any obj The state returned by `create_state()`. """ if self.multi_precision and weight.dtype == numpy.float16: # Wrapper for mixed precision weight_master_copy = state[0] original_state = state[1] grad32 = grad.astype(numpy.float32) self.update(index, weight_master_copy, grad32, original_state) cast(weight_master_copy, dtype=weight.dtype, out=weight) else: self.update(index, weight, grad, state)
[ "def", "update_multi_precision", "(", "self", ",", "index", ",", "weight", ",", "grad", ",", "state", ")", ":", "if", "self", ".", "multi_precision", "and", "weight", ".", "dtype", "==", "numpy", ".", "float16", ":", "# Wrapper for mixed precision", "weight_master_copy", "=", "state", "[", "0", "]", "original_state", "=", "state", "[", "1", "]", "grad32", "=", "grad", ".", "astype", "(", "numpy", ".", "float32", ")", "self", ".", "update", "(", "index", ",", "weight_master_copy", ",", "grad32", ",", "original_state", ")", "cast", "(", "weight_master_copy", ",", "dtype", "=", "weight", ".", "dtype", ",", "out", "=", "weight", ")", "else", ":", "self", ".", "update", "(", "index", ",", "weight", ",", "grad", ",", "state", ")" ]
Updates the given parameter using the corresponding gradient and state. Mixed precision version. Parameters ---------- index : int The unique index of the parameter into the individual learning rates and weight decays. Learning rates and weight decay may be set via `set_lr_mult()` and `set_wd_mult()`, respectively. weight : NDArray The parameter to be updated. grad : NDArray The gradient of the objective with respect to this parameter. state : any obj The state returned by `create_state()`.
[ "Updates", "the", "given", "parameter", "using", "the", "corresponding", "gradient", "and", "state", ".", "Mixed", "precision", "version", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L266-L291
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Optimizer.set_lr_mult
def set_lr_mult(self, args_lr_mult): """Sets an individual learning rate multiplier for each parameter. If you specify a learning rate multiplier for a parameter, then the learning rate for the parameter will be set as the product of the global learning rate `self.lr` and its multiplier. .. note:: The default learning rate multiplier of a `Variable` can be set with `lr_mult` argument in the constructor. Parameters ---------- args_lr_mult : dict of str/int to float For each of its key-value entries, the learning rate multipler for the parameter specified in the key will be set as the given value. You can specify the parameter with either its name or its index. If you use the name, you should pass `sym` in the constructor, and the name you specified in the key of `args_lr_mult` should match the name of the parameter in `sym`. If you use the index, it should correspond to the index of the parameter used in the `update` method. Specifying a parameter by its index is only supported for backward compatibility, and we recommend to use the name instead. """ self.lr_mult = {} if self.sym_info: attr, arg_names = self.sym_info for name in arg_names: if name in attr and '__lr_mult__' in attr[name]: self.lr_mult[name] = float(attr[name]['__lr_mult__']) self.lr_mult.update(args_lr_mult)
python
def set_lr_mult(self, args_lr_mult): """Sets an individual learning rate multiplier for each parameter. If you specify a learning rate multiplier for a parameter, then the learning rate for the parameter will be set as the product of the global learning rate `self.lr` and its multiplier. .. note:: The default learning rate multiplier of a `Variable` can be set with `lr_mult` argument in the constructor. Parameters ---------- args_lr_mult : dict of str/int to float For each of its key-value entries, the learning rate multipler for the parameter specified in the key will be set as the given value. You can specify the parameter with either its name or its index. If you use the name, you should pass `sym` in the constructor, and the name you specified in the key of `args_lr_mult` should match the name of the parameter in `sym`. If you use the index, it should correspond to the index of the parameter used in the `update` method. Specifying a parameter by its index is only supported for backward compatibility, and we recommend to use the name instead. """ self.lr_mult = {} if self.sym_info: attr, arg_names = self.sym_info for name in arg_names: if name in attr and '__lr_mult__' in attr[name]: self.lr_mult[name] = float(attr[name]['__lr_mult__']) self.lr_mult.update(args_lr_mult)
[ "def", "set_lr_mult", "(", "self", ",", "args_lr_mult", ")", ":", "self", ".", "lr_mult", "=", "{", "}", "if", "self", ".", "sym_info", ":", "attr", ",", "arg_names", "=", "self", ".", "sym_info", "for", "name", "in", "arg_names", ":", "if", "name", "in", "attr", "and", "'__lr_mult__'", "in", "attr", "[", "name", "]", ":", "self", ".", "lr_mult", "[", "name", "]", "=", "float", "(", "attr", "[", "name", "]", "[", "'__lr_mult__'", "]", ")", "self", ".", "lr_mult", ".", "update", "(", "args_lr_mult", ")" ]
Sets an individual learning rate multiplier for each parameter. If you specify a learning rate multiplier for a parameter, then the learning rate for the parameter will be set as the product of the global learning rate `self.lr` and its multiplier. .. note:: The default learning rate multiplier of a `Variable` can be set with `lr_mult` argument in the constructor. Parameters ---------- args_lr_mult : dict of str/int to float For each of its key-value entries, the learning rate multipler for the parameter specified in the key will be set as the given value. You can specify the parameter with either its name or its index. If you use the name, you should pass `sym` in the constructor, and the name you specified in the key of `args_lr_mult` should match the name of the parameter in `sym`. If you use the index, it should correspond to the index of the parameter used in the `update` method. Specifying a parameter by its index is only supported for backward compatibility, and we recommend to use the name instead.
[ "Sets", "an", "individual", "learning", "rate", "multiplier", "for", "each", "parameter", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L314-L345
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Optimizer.set_wd_mult
def set_wd_mult(self, args_wd_mult): """Sets an individual weight decay multiplier for each parameter. By default, if `param_idx2name` was provided in the constructor, the weight decay multipler is set as 0 for all parameters whose name don't end with ``_weight`` or ``_gamma``. .. note:: The default weight decay multiplier for a `Variable` can be set with its `wd_mult` argument in the constructor. Parameters ---------- args_wd_mult : dict of string/int to float For each of its key-value entries, the weight decay multipler for the parameter specified in the key will be set as the given value. You can specify the parameter with either its name or its index. If you use the name, you should pass `sym` in the constructor, and the name you specified in the key of `args_lr_mult` should match the name of the parameter in `sym`. If you use the index, it should correspond to the index of the parameter used in the `update` method. Specifying a parameter by its index is only supported for backward compatibility, and we recommend to use the name instead. """ self.wd_mult = {} for n in self.idx2name.values(): if not (n.endswith('_weight') or n.endswith('_gamma')): self.wd_mult[n] = 0.0 if self.sym_info: attr, arg_names = self.sym_info for name in arg_names: if name in attr and '__wd_mult__' in attr[name]: self.wd_mult[name] = float(attr[name]['__wd_mult__']) self.wd_mult.update(args_wd_mult)
python
def set_wd_mult(self, args_wd_mult): """Sets an individual weight decay multiplier for each parameter. By default, if `param_idx2name` was provided in the constructor, the weight decay multipler is set as 0 for all parameters whose name don't end with ``_weight`` or ``_gamma``. .. note:: The default weight decay multiplier for a `Variable` can be set with its `wd_mult` argument in the constructor. Parameters ---------- args_wd_mult : dict of string/int to float For each of its key-value entries, the weight decay multipler for the parameter specified in the key will be set as the given value. You can specify the parameter with either its name or its index. If you use the name, you should pass `sym` in the constructor, and the name you specified in the key of `args_lr_mult` should match the name of the parameter in `sym`. If you use the index, it should correspond to the index of the parameter used in the `update` method. Specifying a parameter by its index is only supported for backward compatibility, and we recommend to use the name instead. """ self.wd_mult = {} for n in self.idx2name.values(): if not (n.endswith('_weight') or n.endswith('_gamma')): self.wd_mult[n] = 0.0 if self.sym_info: attr, arg_names = self.sym_info for name in arg_names: if name in attr and '__wd_mult__' in attr[name]: self.wd_mult[name] = float(attr[name]['__wd_mult__']) self.wd_mult.update(args_wd_mult)
[ "def", "set_wd_mult", "(", "self", ",", "args_wd_mult", ")", ":", "self", ".", "wd_mult", "=", "{", "}", "for", "n", "in", "self", ".", "idx2name", ".", "values", "(", ")", ":", "if", "not", "(", "n", ".", "endswith", "(", "'_weight'", ")", "or", "n", ".", "endswith", "(", "'_gamma'", ")", ")", ":", "self", ".", "wd_mult", "[", "n", "]", "=", "0.0", "if", "self", ".", "sym_info", ":", "attr", ",", "arg_names", "=", "self", ".", "sym_info", "for", "name", "in", "arg_names", ":", "if", "name", "in", "attr", "and", "'__wd_mult__'", "in", "attr", "[", "name", "]", ":", "self", ".", "wd_mult", "[", "name", "]", "=", "float", "(", "attr", "[", "name", "]", "[", "'__wd_mult__'", "]", ")", "self", ".", "wd_mult", ".", "update", "(", "args_wd_mult", ")" ]
Sets an individual weight decay multiplier for each parameter. By default, if `param_idx2name` was provided in the constructor, the weight decay multipler is set as 0 for all parameters whose name don't end with ``_weight`` or ``_gamma``. .. note:: The default weight decay multiplier for a `Variable` can be set with its `wd_mult` argument in the constructor. Parameters ---------- args_wd_mult : dict of string/int to float For each of its key-value entries, the weight decay multipler for the parameter specified in the key will be set as the given value. You can specify the parameter with either its name or its index. If you use the name, you should pass `sym` in the constructor, and the name you specified in the key of `args_lr_mult` should match the name of the parameter in `sym`. If you use the index, it should correspond to the index of the parameter used in the `update` method. Specifying a parameter by its index is only supported for backward compatibility, and we recommend to use the name instead.
[ "Sets", "an", "individual", "weight", "decay", "multiplier", "for", "each", "parameter", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L347-L382
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Optimizer._set_current_context
def _set_current_context(self, device_id): """Sets the number of the currently handled device. Parameters ---------- device_id : int The number of current device. """ if device_id not in self._all_index_update_counts: self._all_index_update_counts[device_id] = {} self._index_update_count = self._all_index_update_counts[device_id]
python
def _set_current_context(self, device_id): """Sets the number of the currently handled device. Parameters ---------- device_id : int The number of current device. """ if device_id not in self._all_index_update_counts: self._all_index_update_counts[device_id] = {} self._index_update_count = self._all_index_update_counts[device_id]
[ "def", "_set_current_context", "(", "self", ",", "device_id", ")", ":", "if", "device_id", "not", "in", "self", ".", "_all_index_update_counts", ":", "self", ".", "_all_index_update_counts", "[", "device_id", "]", "=", "{", "}", "self", ".", "_index_update_count", "=", "self", ".", "_all_index_update_counts", "[", "device_id", "]" ]
Sets the number of the currently handled device. Parameters ---------- device_id : int The number of current device.
[ "Sets", "the", "number", "of", "the", "currently", "handled", "device", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L384-L394
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Optimizer._update_count
def _update_count(self, index): """Updates num_update. Parameters ---------- index : int or list of int The index to be updated. """ if not isinstance(index, (list, tuple)): index = [index] for idx in index: if idx not in self._index_update_count: self._index_update_count[idx] = self.begin_num_update self._index_update_count[idx] += 1 self.num_update = max(self._index_update_count[idx], self.num_update)
python
def _update_count(self, index): """Updates num_update. Parameters ---------- index : int or list of int The index to be updated. """ if not isinstance(index, (list, tuple)): index = [index] for idx in index: if idx not in self._index_update_count: self._index_update_count[idx] = self.begin_num_update self._index_update_count[idx] += 1 self.num_update = max(self._index_update_count[idx], self.num_update)
[ "def", "_update_count", "(", "self", ",", "index", ")", ":", "if", "not", "isinstance", "(", "index", ",", "(", "list", ",", "tuple", ")", ")", ":", "index", "=", "[", "index", "]", "for", "idx", "in", "index", ":", "if", "idx", "not", "in", "self", ".", "_index_update_count", ":", "self", ".", "_index_update_count", "[", "idx", "]", "=", "self", ".", "begin_num_update", "self", ".", "_index_update_count", "[", "idx", "]", "+=", "1", "self", ".", "num_update", "=", "max", "(", "self", ".", "_index_update_count", "[", "idx", "]", ",", "self", ".", "num_update", ")" ]
Updates num_update. Parameters ---------- index : int or list of int The index to be updated.
[ "Updates", "num_update", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L396-L410
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Optimizer._get_lrs
def _get_lrs(self, indices): """Gets the learning rates given the indices of the weights. Parameters ---------- indices : list of int Indices corresponding to weights. Returns ------- lrs : list of float Learning rates for those indices. """ if self.lr_scheduler is not None: lr = self.lr_scheduler(self.num_update) else: lr = self.lr lrs = [lr for _ in indices] for i, index in enumerate(indices): if index in self.param_dict: lrs[i] *= self.param_dict[index].lr_mult elif index in self.lr_mult: lrs[i] *= self.lr_mult[index] elif index in self.idx2name: lrs[i] *= self.lr_mult.get(self.idx2name[index], 1.0) return lrs
python
def _get_lrs(self, indices): """Gets the learning rates given the indices of the weights. Parameters ---------- indices : list of int Indices corresponding to weights. Returns ------- lrs : list of float Learning rates for those indices. """ if self.lr_scheduler is not None: lr = self.lr_scheduler(self.num_update) else: lr = self.lr lrs = [lr for _ in indices] for i, index in enumerate(indices): if index in self.param_dict: lrs[i] *= self.param_dict[index].lr_mult elif index in self.lr_mult: lrs[i] *= self.lr_mult[index] elif index in self.idx2name: lrs[i] *= self.lr_mult.get(self.idx2name[index], 1.0) return lrs
[ "def", "_get_lrs", "(", "self", ",", "indices", ")", ":", "if", "self", ".", "lr_scheduler", "is", "not", "None", ":", "lr", "=", "self", ".", "lr_scheduler", "(", "self", ".", "num_update", ")", "else", ":", "lr", "=", "self", ".", "lr", "lrs", "=", "[", "lr", "for", "_", "in", "indices", "]", "for", "i", ",", "index", "in", "enumerate", "(", "indices", ")", ":", "if", "index", "in", "self", ".", "param_dict", ":", "lrs", "[", "i", "]", "*=", "self", ".", "param_dict", "[", "index", "]", ".", "lr_mult", "elif", "index", "in", "self", ".", "lr_mult", ":", "lrs", "[", "i", "]", "*=", "self", ".", "lr_mult", "[", "index", "]", "elif", "index", "in", "self", ".", "idx2name", ":", "lrs", "[", "i", "]", "*=", "self", ".", "lr_mult", ".", "get", "(", "self", ".", "idx2name", "[", "index", "]", ",", "1.0", ")", "return", "lrs" ]
Gets the learning rates given the indices of the weights. Parameters ---------- indices : list of int Indices corresponding to weights. Returns ------- lrs : list of float Learning rates for those indices.
[ "Gets", "the", "learning", "rates", "given", "the", "indices", "of", "the", "weights", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L412-L438
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Optimizer._get_wds
def _get_wds(self, indices): """Gets weight decays for indices. Returns 0 for non-weights if the name of weights are provided for `__init__`. Parameters ---------- indices : list of int Indices of weights. Returns ------- wds : list of float Weight decays for those indices. """ wds = [self.wd for _ in indices] for i, index in enumerate(indices): if index in self.param_dict: wds[i] *= self.param_dict[index].wd_mult elif index in self.wd_mult: wds[i] *= self.wd_mult[index] elif index in self.idx2name: wds[i] *= self.wd_mult.get(self.idx2name[index], 1.0) return wds
python
def _get_wds(self, indices): """Gets weight decays for indices. Returns 0 for non-weights if the name of weights are provided for `__init__`. Parameters ---------- indices : list of int Indices of weights. Returns ------- wds : list of float Weight decays for those indices. """ wds = [self.wd for _ in indices] for i, index in enumerate(indices): if index in self.param_dict: wds[i] *= self.param_dict[index].wd_mult elif index in self.wd_mult: wds[i] *= self.wd_mult[index] elif index in self.idx2name: wds[i] *= self.wd_mult.get(self.idx2name[index], 1.0) return wds
[ "def", "_get_wds", "(", "self", ",", "indices", ")", ":", "wds", "=", "[", "self", ".", "wd", "for", "_", "in", "indices", "]", "for", "i", ",", "index", "in", "enumerate", "(", "indices", ")", ":", "if", "index", "in", "self", ".", "param_dict", ":", "wds", "[", "i", "]", "*=", "self", ".", "param_dict", "[", "index", "]", ".", "wd_mult", "elif", "index", "in", "self", ".", "wd_mult", ":", "wds", "[", "i", "]", "*=", "self", ".", "wd_mult", "[", "index", "]", "elif", "index", "in", "self", ".", "idx2name", ":", "wds", "[", "i", "]", "*=", "self", ".", "wd_mult", ".", "get", "(", "self", ".", "idx2name", "[", "index", "]", ",", "1.0", ")", "return", "wds" ]
Gets weight decays for indices. Returns 0 for non-weights if the name of weights are provided for `__init__`. Parameters ---------- indices : list of int Indices of weights. Returns ------- wds : list of float Weight decays for those indices.
[ "Gets", "weight", "decays", "for", "indices", ".", "Returns", "0", "for", "non", "-", "weights", "if", "the", "name", "of", "weights", "are", "provided", "for", "__init__", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L455-L477
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Updater.sync_state_context
def sync_state_context(self, state, context): """sync state context.""" if isinstance(state, NDArray): return state.as_in_context(context) elif isinstance(state, (tuple, list)): synced_state = (self.sync_state_context(i, context) for i in state) if isinstance(state, tuple): return tuple(synced_state) else: return list(synced_state) else: return state
python
def sync_state_context(self, state, context): """sync state context.""" if isinstance(state, NDArray): return state.as_in_context(context) elif isinstance(state, (tuple, list)): synced_state = (self.sync_state_context(i, context) for i in state) if isinstance(state, tuple): return tuple(synced_state) else: return list(synced_state) else: return state
[ "def", "sync_state_context", "(", "self", ",", "state", ",", "context", ")", ":", "if", "isinstance", "(", "state", ",", "NDArray", ")", ":", "return", "state", ".", "as_in_context", "(", "context", ")", "elif", "isinstance", "(", "state", ",", "(", "tuple", ",", "list", ")", ")", ":", "synced_state", "=", "(", "self", ".", "sync_state_context", "(", "i", ",", "context", ")", "for", "i", "in", "state", ")", "if", "isinstance", "(", "state", ",", "tuple", ")", ":", "return", "tuple", "(", "synced_state", ")", "else", ":", "return", "list", "(", "synced_state", ")", "else", ":", "return", "state" ]
sync state context.
[ "sync", "state", "context", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L1679-L1690
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Updater.set_states
def set_states(self, states): """Sets updater states.""" states = pickle.loads(states) if isinstance(states, tuple) and len(states) == 2: self.states, self.optimizer = states else: self.states = states self.states_synced = dict.fromkeys(self.states.keys(), False)
python
def set_states(self, states): """Sets updater states.""" states = pickle.loads(states) if isinstance(states, tuple) and len(states) == 2: self.states, self.optimizer = states else: self.states = states self.states_synced = dict.fromkeys(self.states.keys(), False)
[ "def", "set_states", "(", "self", ",", "states", ")", ":", "states", "=", "pickle", ".", "loads", "(", "states", ")", "if", "isinstance", "(", "states", ",", "tuple", ")", "and", "len", "(", "states", ")", "==", "2", ":", "self", ".", "states", ",", "self", ".", "optimizer", "=", "states", "else", ":", "self", ".", "states", "=", "states", "self", ".", "states_synced", "=", "dict", ".", "fromkeys", "(", "self", ".", "states", ".", "keys", "(", ")", ",", "False", ")" ]
Sets updater states.
[ "Sets", "updater", "states", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L1692-L1699
train
apache/incubator-mxnet
python/mxnet/optimizer/optimizer.py
Updater.get_states
def get_states(self, dump_optimizer=False): """Gets updater states. Parameters ---------- dump_optimizer : bool, default False Whether to also save the optimizer itself. This would also save optimizer information such as learning rate and weight decay schedules. """ return pickle.dumps((self.states, self.optimizer) if dump_optimizer else self.states)
python
def get_states(self, dump_optimizer=False): """Gets updater states. Parameters ---------- dump_optimizer : bool, default False Whether to also save the optimizer itself. This would also save optimizer information such as learning rate and weight decay schedules. """ return pickle.dumps((self.states, self.optimizer) if dump_optimizer else self.states)
[ "def", "get_states", "(", "self", ",", "dump_optimizer", "=", "False", ")", ":", "return", "pickle", ".", "dumps", "(", "(", "self", ".", "states", ",", "self", ".", "optimizer", ")", "if", "dump_optimizer", "else", "self", ".", "states", ")" ]
Gets updater states. Parameters ---------- dump_optimizer : bool, default False Whether to also save the optimizer itself. This would also save optimizer information such as learning rate and weight decay schedules.
[ "Gets", "updater", "states", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/optimizer/optimizer.py#L1701-L1710
train
apache/incubator-mxnet
example/gluon/lipnet/utils/preprocess_data.py
preprocess
def preprocess(from_idx, to_idx, _params): """ Preprocess: Convert a video into the mouth images """ source_exts = '*.mpg' src_path = _params['src_path'] tgt_path = _params['tgt_path'] face_predictor_path = './shape_predictor_68_face_landmarks.dat' succ = set() fail = set() for idx in range(from_idx, to_idx): s_id = 's' + str(idx) + '/' source_path = src_path + '/' + s_id target_path = tgt_path + '/' + s_id fail_cnt = 0 for filepath in find_files(source_path, source_exts): print("Processing: {}".format(filepath)) filepath_wo_ext = os.path.splitext(filepath)[0].split('/')[-2:] target_dir = os.path.join(tgt_path, '/'.join(filepath_wo_ext)) if os.path.exists(target_dir): continue try: video = Video(vtype='face', \ face_predictor_path=face_predictor_path).from_video(filepath) mkdir_p(target_dir) i = 0 if video.mouth[0] is None: continue for frame in video.mouth: io.imsave(os.path.join(target_dir, "mouth_{0:03d}.png".format(i)), frame) i += 1 except ValueError as error: print(error) fail_cnt += 1 if fail_cnt == 0: succ.add(idx) else: fail.add(idx) return (succ, fail)
python
def preprocess(from_idx, to_idx, _params): """ Preprocess: Convert a video into the mouth images """ source_exts = '*.mpg' src_path = _params['src_path'] tgt_path = _params['tgt_path'] face_predictor_path = './shape_predictor_68_face_landmarks.dat' succ = set() fail = set() for idx in range(from_idx, to_idx): s_id = 's' + str(idx) + '/' source_path = src_path + '/' + s_id target_path = tgt_path + '/' + s_id fail_cnt = 0 for filepath in find_files(source_path, source_exts): print("Processing: {}".format(filepath)) filepath_wo_ext = os.path.splitext(filepath)[0].split('/')[-2:] target_dir = os.path.join(tgt_path, '/'.join(filepath_wo_ext)) if os.path.exists(target_dir): continue try: video = Video(vtype='face', \ face_predictor_path=face_predictor_path).from_video(filepath) mkdir_p(target_dir) i = 0 if video.mouth[0] is None: continue for frame in video.mouth: io.imsave(os.path.join(target_dir, "mouth_{0:03d}.png".format(i)), frame) i += 1 except ValueError as error: print(error) fail_cnt += 1 if fail_cnt == 0: succ.add(idx) else: fail.add(idx) return (succ, fail)
[ "def", "preprocess", "(", "from_idx", ",", "to_idx", ",", "_params", ")", ":", "source_exts", "=", "'*.mpg'", "src_path", "=", "_params", "[", "'src_path'", "]", "tgt_path", "=", "_params", "[", "'tgt_path'", "]", "face_predictor_path", "=", "'./shape_predictor_68_face_landmarks.dat'", "succ", "=", "set", "(", ")", "fail", "=", "set", "(", ")", "for", "idx", "in", "range", "(", "from_idx", ",", "to_idx", ")", ":", "s_id", "=", "'s'", "+", "str", "(", "idx", ")", "+", "'/'", "source_path", "=", "src_path", "+", "'/'", "+", "s_id", "target_path", "=", "tgt_path", "+", "'/'", "+", "s_id", "fail_cnt", "=", "0", "for", "filepath", "in", "find_files", "(", "source_path", ",", "source_exts", ")", ":", "print", "(", "\"Processing: {}\"", ".", "format", "(", "filepath", ")", ")", "filepath_wo_ext", "=", "os", ".", "path", ".", "splitext", "(", "filepath", ")", "[", "0", "]", ".", "split", "(", "'/'", ")", "[", "-", "2", ":", "]", "target_dir", "=", "os", ".", "path", ".", "join", "(", "tgt_path", ",", "'/'", ".", "join", "(", "filepath_wo_ext", ")", ")", "if", "os", ".", "path", ".", "exists", "(", "target_dir", ")", ":", "continue", "try", ":", "video", "=", "Video", "(", "vtype", "=", "'face'", ",", "face_predictor_path", "=", "face_predictor_path", ")", ".", "from_video", "(", "filepath", ")", "mkdir_p", "(", "target_dir", ")", "i", "=", "0", "if", "video", ".", "mouth", "[", "0", "]", "is", "None", ":", "continue", "for", "frame", "in", "video", ".", "mouth", ":", "io", ".", "imsave", "(", "os", ".", "path", ".", "join", "(", "target_dir", ",", "\"mouth_{0:03d}.png\"", ".", "format", "(", "i", ")", ")", ",", "frame", ")", "i", "+=", "1", "except", "ValueError", "as", "error", ":", "print", "(", "error", ")", "fail_cnt", "+=", "1", "if", "fail_cnt", "==", "0", ":", "succ", ".", "add", "(", "idx", ")", "else", ":", "fail", ".", "add", "(", "idx", ")", "return", "(", "succ", ",", "fail", ")" ]
Preprocess: Convert a video into the mouth images
[ "Preprocess", ":", "Convert", "a", "video", "into", "the", "mouth", "images" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/utils/preprocess_data.py#L202-L243
train
apache/incubator-mxnet
example/gluon/lipnet/utils/preprocess_data.py
Video.from_frames
def from_frames(self, path): """ Read from frames """ frames_path = sorted([os.path.join(path, x) for x in os.listdir(path)]) frames = [ndimage.imread(frame_path) for frame_path in frames_path] self.handle_type(frames) return self
python
def from_frames(self, path): """ Read from frames """ frames_path = sorted([os.path.join(path, x) for x in os.listdir(path)]) frames = [ndimage.imread(frame_path) for frame_path in frames_path] self.handle_type(frames) return self
[ "def", "from_frames", "(", "self", ",", "path", ")", ":", "frames_path", "=", "sorted", "(", "[", "os", ".", "path", ".", "join", "(", "path", ",", "x", ")", "for", "x", "in", "os", ".", "listdir", "(", "path", ")", "]", ")", "frames", "=", "[", "ndimage", ".", "imread", "(", "frame_path", ")", "for", "frame_path", "in", "frames_path", "]", "self", ".", "handle_type", "(", "frames", ")", "return", "self" ]
Read from frames
[ "Read", "from", "frames" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/utils/preprocess_data.py#L71-L78
train
apache/incubator-mxnet
example/gluon/lipnet/utils/preprocess_data.py
Video.from_video
def from_video(self, path): """ Read from videos """ frames = self.get_video_frames(path) self.handle_type(frames) return self
python
def from_video(self, path): """ Read from videos """ frames = self.get_video_frames(path) self.handle_type(frames) return self
[ "def", "from_video", "(", "self", ",", "path", ")", ":", "frames", "=", "self", ".", "get_video_frames", "(", "path", ")", "self", ".", "handle_type", "(", "frames", ")", "return", "self" ]
Read from videos
[ "Read", "from", "videos" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/utils/preprocess_data.py#L80-L86
train
apache/incubator-mxnet
example/gluon/lipnet/utils/preprocess_data.py
Video.handle_type
def handle_type(self, frames): """ Config video types """ if self.vtype == 'mouth': self.process_frames_mouth(frames) elif self.vtype == 'face': self.process_frames_face(frames) else: raise Exception('Video type not found')
python
def handle_type(self, frames): """ Config video types """ if self.vtype == 'mouth': self.process_frames_mouth(frames) elif self.vtype == 'face': self.process_frames_face(frames) else: raise Exception('Video type not found')
[ "def", "handle_type", "(", "self", ",", "frames", ")", ":", "if", "self", ".", "vtype", "==", "'mouth'", ":", "self", ".", "process_frames_mouth", "(", "frames", ")", "elif", "self", ".", "vtype", "==", "'face'", ":", "self", ".", "process_frames_face", "(", "frames", ")", "else", ":", "raise", "Exception", "(", "'Video type not found'", ")" ]
Config video types
[ "Config", "video", "types" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/utils/preprocess_data.py#L95-L104
train
apache/incubator-mxnet
example/gluon/lipnet/utils/preprocess_data.py
Video.process_frames_face
def process_frames_face(self, frames): """ Preprocess from frames using face detector """ detector = dlib.get_frontal_face_detector() predictor = dlib.shape_predictor(self.face_predictor_path) mouth_frames = self.get_frames_mouth(detector, predictor, frames) self.face = np.array(frames) self.mouth = np.array(mouth_frames) if mouth_frames[0] is not None: self.set_data(mouth_frames)
python
def process_frames_face(self, frames): """ Preprocess from frames using face detector """ detector = dlib.get_frontal_face_detector() predictor = dlib.shape_predictor(self.face_predictor_path) mouth_frames = self.get_frames_mouth(detector, predictor, frames) self.face = np.array(frames) self.mouth = np.array(mouth_frames) if mouth_frames[0] is not None: self.set_data(mouth_frames)
[ "def", "process_frames_face", "(", "self", ",", "frames", ")", ":", "detector", "=", "dlib", ".", "get_frontal_face_detector", "(", ")", "predictor", "=", "dlib", ".", "shape_predictor", "(", "self", ".", "face_predictor_path", ")", "mouth_frames", "=", "self", ".", "get_frames_mouth", "(", "detector", ",", "predictor", ",", "frames", ")", "self", ".", "face", "=", "np", ".", "array", "(", "frames", ")", "self", ".", "mouth", "=", "np", ".", "array", "(", "mouth_frames", ")", "if", "mouth_frames", "[", "0", "]", "is", "not", "None", ":", "self", ".", "set_data", "(", "mouth_frames", ")" ]
Preprocess from frames using face detector
[ "Preprocess", "from", "frames", "using", "face", "detector" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/utils/preprocess_data.py#L106-L116
train
apache/incubator-mxnet
example/gluon/lipnet/utils/preprocess_data.py
Video.process_frames_mouth
def process_frames_mouth(self, frames): """ Preprocess from frames using mouth detector """ self.face = np.array(frames) self.mouth = np.array(frames) self.set_data(frames)
python
def process_frames_mouth(self, frames): """ Preprocess from frames using mouth detector """ self.face = np.array(frames) self.mouth = np.array(frames) self.set_data(frames)
[ "def", "process_frames_mouth", "(", "self", ",", "frames", ")", ":", "self", ".", "face", "=", "np", ".", "array", "(", "frames", ")", "self", ".", "mouth", "=", "np", ".", "array", "(", "frames", ")", "self", ".", "set_data", "(", "frames", ")" ]
Preprocess from frames using mouth detector
[ "Preprocess", "from", "frames", "using", "mouth", "detector" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/utils/preprocess_data.py#L118-L124
train
apache/incubator-mxnet
example/gluon/lipnet/utils/preprocess_data.py
Video.get_frames_mouth
def get_frames_mouth(self, detector, predictor, frames): """ Get frames using mouth crop """ mouth_width = 100 mouth_height = 50 horizontal_pad = 0.19 normalize_ratio = None mouth_frames = [] for frame in frames: dets = detector(frame, 1) shape = None for det in dets: shape = predictor(frame, det) i = -1 if shape is None: # Detector doesn't detect face, just return None return [None] mouth_points = [] for part in shape.parts(): i += 1 if i < 48: # Only take mouth region continue mouth_points.append((part.x, part.y)) np_mouth_points = np.array(mouth_points) mouth_centroid = np.mean(np_mouth_points[:, -2:], axis=0) if normalize_ratio is None: mouth_left = np.min(np_mouth_points[:, :-1]) * (1.0 - horizontal_pad) mouth_right = np.max(np_mouth_points[:, :-1]) * (1.0 + horizontal_pad) normalize_ratio = mouth_width / float(mouth_right - mouth_left) new_img_shape = (int(frame.shape[0] * normalize_ratio), int(frame.shape[1] * normalize_ratio)) resized_img = imresize(frame, new_img_shape) mouth_centroid_norm = mouth_centroid * normalize_ratio mouth_l = int(mouth_centroid_norm[0] - mouth_width / 2) mouth_r = int(mouth_centroid_norm[0] + mouth_width / 2) mouth_t = int(mouth_centroid_norm[1] - mouth_height / 2) mouth_b = int(mouth_centroid_norm[1] + mouth_height / 2) mouth_crop_image = resized_img[mouth_t:mouth_b, mouth_l:mouth_r] mouth_frames.append(mouth_crop_image) return mouth_frames
python
def get_frames_mouth(self, detector, predictor, frames): """ Get frames using mouth crop """ mouth_width = 100 mouth_height = 50 horizontal_pad = 0.19 normalize_ratio = None mouth_frames = [] for frame in frames: dets = detector(frame, 1) shape = None for det in dets: shape = predictor(frame, det) i = -1 if shape is None: # Detector doesn't detect face, just return None return [None] mouth_points = [] for part in shape.parts(): i += 1 if i < 48: # Only take mouth region continue mouth_points.append((part.x, part.y)) np_mouth_points = np.array(mouth_points) mouth_centroid = np.mean(np_mouth_points[:, -2:], axis=0) if normalize_ratio is None: mouth_left = np.min(np_mouth_points[:, :-1]) * (1.0 - horizontal_pad) mouth_right = np.max(np_mouth_points[:, :-1]) * (1.0 + horizontal_pad) normalize_ratio = mouth_width / float(mouth_right - mouth_left) new_img_shape = (int(frame.shape[0] * normalize_ratio), int(frame.shape[1] * normalize_ratio)) resized_img = imresize(frame, new_img_shape) mouth_centroid_norm = mouth_centroid * normalize_ratio mouth_l = int(mouth_centroid_norm[0] - mouth_width / 2) mouth_r = int(mouth_centroid_norm[0] + mouth_width / 2) mouth_t = int(mouth_centroid_norm[1] - mouth_height / 2) mouth_b = int(mouth_centroid_norm[1] + mouth_height / 2) mouth_crop_image = resized_img[mouth_t:mouth_b, mouth_l:mouth_r] mouth_frames.append(mouth_crop_image) return mouth_frames
[ "def", "get_frames_mouth", "(", "self", ",", "detector", ",", "predictor", ",", "frames", ")", ":", "mouth_width", "=", "100", "mouth_height", "=", "50", "horizontal_pad", "=", "0.19", "normalize_ratio", "=", "None", "mouth_frames", "=", "[", "]", "for", "frame", "in", "frames", ":", "dets", "=", "detector", "(", "frame", ",", "1", ")", "shape", "=", "None", "for", "det", "in", "dets", ":", "shape", "=", "predictor", "(", "frame", ",", "det", ")", "i", "=", "-", "1", "if", "shape", "is", "None", ":", "# Detector doesn't detect face, just return None", "return", "[", "None", "]", "mouth_points", "=", "[", "]", "for", "part", "in", "shape", ".", "parts", "(", ")", ":", "i", "+=", "1", "if", "i", "<", "48", ":", "# Only take mouth region", "continue", "mouth_points", ".", "append", "(", "(", "part", ".", "x", ",", "part", ".", "y", ")", ")", "np_mouth_points", "=", "np", ".", "array", "(", "mouth_points", ")", "mouth_centroid", "=", "np", ".", "mean", "(", "np_mouth_points", "[", ":", ",", "-", "2", ":", "]", ",", "axis", "=", "0", ")", "if", "normalize_ratio", "is", "None", ":", "mouth_left", "=", "np", ".", "min", "(", "np_mouth_points", "[", ":", ",", ":", "-", "1", "]", ")", "*", "(", "1.0", "-", "horizontal_pad", ")", "mouth_right", "=", "np", ".", "max", "(", "np_mouth_points", "[", ":", ",", ":", "-", "1", "]", ")", "*", "(", "1.0", "+", "horizontal_pad", ")", "normalize_ratio", "=", "mouth_width", "/", "float", "(", "mouth_right", "-", "mouth_left", ")", "new_img_shape", "=", "(", "int", "(", "frame", ".", "shape", "[", "0", "]", "*", "normalize_ratio", ")", ",", "int", "(", "frame", ".", "shape", "[", "1", "]", "*", "normalize_ratio", ")", ")", "resized_img", "=", "imresize", "(", "frame", ",", "new_img_shape", ")", "mouth_centroid_norm", "=", "mouth_centroid", "*", "normalize_ratio", "mouth_l", "=", "int", "(", "mouth_centroid_norm", "[", "0", "]", "-", "mouth_width", "/", "2", ")", "mouth_r", "=", "int", "(", "mouth_centroid_norm", "[", "0", "]", "+", "mouth_width", "/", "2", ")", "mouth_t", "=", "int", "(", "mouth_centroid_norm", "[", "1", "]", "-", "mouth_height", "/", "2", ")", "mouth_b", "=", "int", "(", "mouth_centroid_norm", "[", "1", "]", "+", "mouth_height", "/", "2", ")", "mouth_crop_image", "=", "resized_img", "[", "mouth_t", ":", "mouth_b", ",", "mouth_l", ":", "mouth_r", "]", "mouth_frames", ".", "append", "(", "mouth_crop_image", ")", "return", "mouth_frames" ]
Get frames using mouth crop
[ "Get", "frames", "using", "mouth", "crop" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/utils/preprocess_data.py#L126-L173
train
apache/incubator-mxnet
example/gluon/lipnet/utils/preprocess_data.py
Video.get_video_frames
def get_video_frames(self, path): """ Get video frames """ videogen = skvideo.io.vreader(path) frames = np.array([frame for frame in videogen]) return frames
python
def get_video_frames(self, path): """ Get video frames """ videogen = skvideo.io.vreader(path) frames = np.array([frame for frame in videogen]) return frames
[ "def", "get_video_frames", "(", "self", ",", "path", ")", ":", "videogen", "=", "skvideo", ".", "io", ".", "vreader", "(", "path", ")", "frames", "=", "np", ".", "array", "(", "[", "frame", "for", "frame", "in", "videogen", "]", ")", "return", "frames" ]
Get video frames
[ "Get", "video", "frames" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/utils/preprocess_data.py#L175-L181
train
apache/incubator-mxnet
example/gluon/lipnet/utils/preprocess_data.py
Video.set_data
def set_data(self, frames): """ Prepare the input of model """ data_frames = [] for frame in frames: #frame H x W x C frame = frame.swapaxes(0, 1) # swap width and height to form format W x H x C if len(frame.shape) < 3: frame = np.array([frame]).swapaxes(0, 2).swapaxes(0, 1) # Add grayscale channel data_frames.append(frame) frames_n = len(data_frames) data_frames = np.array(data_frames) # T x W x H x C data_frames = np.rollaxis(data_frames, 3) # C x T x W x H data_frames = data_frames.swapaxes(2, 3) # C x T x H x W = NCDHW self.data = data_frames self.length = frames_n
python
def set_data(self, frames): """ Prepare the input of model """ data_frames = [] for frame in frames: #frame H x W x C frame = frame.swapaxes(0, 1) # swap width and height to form format W x H x C if len(frame.shape) < 3: frame = np.array([frame]).swapaxes(0, 2).swapaxes(0, 1) # Add grayscale channel data_frames.append(frame) frames_n = len(data_frames) data_frames = np.array(data_frames) # T x W x H x C data_frames = np.rollaxis(data_frames, 3) # C x T x W x H data_frames = data_frames.swapaxes(2, 3) # C x T x H x W = NCDHW self.data = data_frames self.length = frames_n
[ "def", "set_data", "(", "self", ",", "frames", ")", ":", "data_frames", "=", "[", "]", "for", "frame", "in", "frames", ":", "#frame H x W x C", "frame", "=", "frame", ".", "swapaxes", "(", "0", ",", "1", ")", "# swap width and height to form format W x H x C", "if", "len", "(", "frame", ".", "shape", ")", "<", "3", ":", "frame", "=", "np", ".", "array", "(", "[", "frame", "]", ")", ".", "swapaxes", "(", "0", ",", "2", ")", ".", "swapaxes", "(", "0", ",", "1", ")", "# Add grayscale channel", "data_frames", ".", "append", "(", "frame", ")", "frames_n", "=", "len", "(", "data_frames", ")", "data_frames", "=", "np", ".", "array", "(", "data_frames", ")", "# T x W x H x C", "data_frames", "=", "np", ".", "rollaxis", "(", "data_frames", ",", "3", ")", "# C x T x W x H", "data_frames", "=", "data_frames", ".", "swapaxes", "(", "2", ",", "3", ")", "# C x T x H x W = NCDHW", "self", ".", "data", "=", "data_frames", "self", ".", "length", "=", "frames_n" ]
Prepare the input of model
[ "Prepare", "the", "input", "of", "model" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/utils/preprocess_data.py#L183-L200
train
apache/incubator-mxnet
example/speech_recognition/stt_io_bucketingiter.py
BucketSTTIter.reset
def reset(self): """Resets the iterator to the beginning of the data.""" self.curr_idx = 0 random.shuffle(self.idx) for buck in self.data: np.random.shuffle(buck)
python
def reset(self): """Resets the iterator to the beginning of the data.""" self.curr_idx = 0 random.shuffle(self.idx) for buck in self.data: np.random.shuffle(buck)
[ "def", "reset", "(", "self", ")", ":", "self", ".", "curr_idx", "=", "0", "random", ".", "shuffle", "(", "self", ".", "idx", ")", "for", "buck", "in", "self", ".", "data", ":", "np", ".", "random", ".", "shuffle", "(", "buck", ")" ]
Resets the iterator to the beginning of the data.
[ "Resets", "the", "iterator", "to", "the", "beginning", "of", "the", "data", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/speech_recognition/stt_io_bucketingiter.py#L125-L130
train
apache/incubator-mxnet
example/speech_recognition/stt_io_bucketingiter.py
BucketSTTIter.next
def next(self): """Returns the next batch of data.""" if self.curr_idx == len(self.idx): raise StopIteration i, j = self.idx[self.curr_idx] self.curr_idx += 1 audio_paths = [] texts = [] for duration, audio_path, text in self.data[i][j:j+self.batch_size]: audio_paths.append(audio_path) texts.append(text) if self.is_first_epoch: data_set = self.datagen.prepare_minibatch(audio_paths, texts, overwrite=True, is_bi_graphemes=self.is_bi_graphemes, seq_length=self.buckets[i], save_feature_as_csvfile=self.save_feature_as_csvfile) else: data_set = self.datagen.prepare_minibatch(audio_paths, texts, overwrite=False, is_bi_graphemes=self.is_bi_graphemes, seq_length=self.buckets[i], save_feature_as_csvfile=self.save_feature_as_csvfile) data_all = [mx.nd.array(data_set['x'])] + self.init_state_arrays label_all = [mx.nd.array(data_set['y'])] self.label = label_all provide_data = [('data', (self.batch_size, self.buckets[i], self.width * self.height))] + self.init_states return mx.io.DataBatch(data_all, label_all, pad=0, bucket_key=self.buckets[i], provide_data=provide_data, provide_label=self.provide_label)
python
def next(self): """Returns the next batch of data.""" if self.curr_idx == len(self.idx): raise StopIteration i, j = self.idx[self.curr_idx] self.curr_idx += 1 audio_paths = [] texts = [] for duration, audio_path, text in self.data[i][j:j+self.batch_size]: audio_paths.append(audio_path) texts.append(text) if self.is_first_epoch: data_set = self.datagen.prepare_minibatch(audio_paths, texts, overwrite=True, is_bi_graphemes=self.is_bi_graphemes, seq_length=self.buckets[i], save_feature_as_csvfile=self.save_feature_as_csvfile) else: data_set = self.datagen.prepare_minibatch(audio_paths, texts, overwrite=False, is_bi_graphemes=self.is_bi_graphemes, seq_length=self.buckets[i], save_feature_as_csvfile=self.save_feature_as_csvfile) data_all = [mx.nd.array(data_set['x'])] + self.init_state_arrays label_all = [mx.nd.array(data_set['y'])] self.label = label_all provide_data = [('data', (self.batch_size, self.buckets[i], self.width * self.height))] + self.init_states return mx.io.DataBatch(data_all, label_all, pad=0, bucket_key=self.buckets[i], provide_data=provide_data, provide_label=self.provide_label)
[ "def", "next", "(", "self", ")", ":", "if", "self", ".", "curr_idx", "==", "len", "(", "self", ".", "idx", ")", ":", "raise", "StopIteration", "i", ",", "j", "=", "self", ".", "idx", "[", "self", ".", "curr_idx", "]", "self", ".", "curr_idx", "+=", "1", "audio_paths", "=", "[", "]", "texts", "=", "[", "]", "for", "duration", ",", "audio_path", ",", "text", "in", "self", ".", "data", "[", "i", "]", "[", "j", ":", "j", "+", "self", ".", "batch_size", "]", ":", "audio_paths", ".", "append", "(", "audio_path", ")", "texts", ".", "append", "(", "text", ")", "if", "self", ".", "is_first_epoch", ":", "data_set", "=", "self", ".", "datagen", ".", "prepare_minibatch", "(", "audio_paths", ",", "texts", ",", "overwrite", "=", "True", ",", "is_bi_graphemes", "=", "self", ".", "is_bi_graphemes", ",", "seq_length", "=", "self", ".", "buckets", "[", "i", "]", ",", "save_feature_as_csvfile", "=", "self", ".", "save_feature_as_csvfile", ")", "else", ":", "data_set", "=", "self", ".", "datagen", ".", "prepare_minibatch", "(", "audio_paths", ",", "texts", ",", "overwrite", "=", "False", ",", "is_bi_graphemes", "=", "self", ".", "is_bi_graphemes", ",", "seq_length", "=", "self", ".", "buckets", "[", "i", "]", ",", "save_feature_as_csvfile", "=", "self", ".", "save_feature_as_csvfile", ")", "data_all", "=", "[", "mx", ".", "nd", ".", "array", "(", "data_set", "[", "'x'", "]", ")", "]", "+", "self", ".", "init_state_arrays", "label_all", "=", "[", "mx", ".", "nd", ".", "array", "(", "data_set", "[", "'y'", "]", ")", "]", "self", ".", "label", "=", "label_all", "provide_data", "=", "[", "(", "'data'", ",", "(", "self", ".", "batch_size", ",", "self", ".", "buckets", "[", "i", "]", ",", "self", ".", "width", "*", "self", ".", "height", ")", ")", "]", "+", "self", ".", "init_states", "return", "mx", ".", "io", ".", "DataBatch", "(", "data_all", ",", "label_all", ",", "pad", "=", "0", ",", "bucket_key", "=", "self", ".", "buckets", "[", "i", "]", ",", "provide_data", "=", "provide_data", ",", "provide_label", "=", "self", ".", "provide_label", ")" ]
Returns the next batch of data.
[ "Returns", "the", "next", "batch", "of", "data", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/speech_recognition/stt_io_bucketingiter.py#L132-L165
train
apache/incubator-mxnet
example/gluon/style_transfer/utils.py
subtract_imagenet_mean_preprocess_batch
def subtract_imagenet_mean_preprocess_batch(batch): """Subtract ImageNet mean pixel-wise from a BGR image.""" batch = F.swapaxes(batch,0, 1) (r, g, b) = F.split(batch, num_outputs=3, axis=0) r = r - 123.680 g = g - 116.779 b = b - 103.939 batch = F.concat(b, g, r, dim=0) batch = F.swapaxes(batch,0, 1) return batch
python
def subtract_imagenet_mean_preprocess_batch(batch): """Subtract ImageNet mean pixel-wise from a BGR image.""" batch = F.swapaxes(batch,0, 1) (r, g, b) = F.split(batch, num_outputs=3, axis=0) r = r - 123.680 g = g - 116.779 b = b - 103.939 batch = F.concat(b, g, r, dim=0) batch = F.swapaxes(batch,0, 1) return batch
[ "def", "subtract_imagenet_mean_preprocess_batch", "(", "batch", ")", ":", "batch", "=", "F", ".", "swapaxes", "(", "batch", ",", "0", ",", "1", ")", "(", "r", ",", "g", ",", "b", ")", "=", "F", ".", "split", "(", "batch", ",", "num_outputs", "=", "3", ",", "axis", "=", "0", ")", "r", "=", "r", "-", "123.680", "g", "=", "g", "-", "116.779", "b", "=", "b", "-", "103.939", "batch", "=", "F", ".", "concat", "(", "b", ",", "g", ",", "r", ",", "dim", "=", "0", ")", "batch", "=", "F", ".", "swapaxes", "(", "batch", ",", "0", ",", "1", ")", "return", "batch" ]
Subtract ImageNet mean pixel-wise from a BGR image.
[ "Subtract", "ImageNet", "mean", "pixel", "-", "wise", "from", "a", "BGR", "image", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/style_transfer/utils.py#L69-L78
train
apache/incubator-mxnet
example/gluon/style_transfer/utils.py
imagenet_clamp_batch
def imagenet_clamp_batch(batch, low, high): """ Not necessary in practice """ F.clip(batch[:,0,:,:],low-123.680, high-123.680) F.clip(batch[:,1,:,:],low-116.779, high-116.779) F.clip(batch[:,2,:,:],low-103.939, high-103.939)
python
def imagenet_clamp_batch(batch, low, high): """ Not necessary in practice """ F.clip(batch[:,0,:,:],low-123.680, high-123.680) F.clip(batch[:,1,:,:],low-116.779, high-116.779) F.clip(batch[:,2,:,:],low-103.939, high-103.939)
[ "def", "imagenet_clamp_batch", "(", "batch", ",", "low", ",", "high", ")", ":", "F", ".", "clip", "(", "batch", "[", ":", ",", "0", ",", ":", ",", ":", "]", ",", "low", "-", "123.680", ",", "high", "-", "123.680", ")", "F", ".", "clip", "(", "batch", "[", ":", ",", "1", ",", ":", ",", ":", "]", ",", "low", "-", "116.779", ",", "high", "-", "116.779", ")", "F", ".", "clip", "(", "batch", "[", ":", ",", "2", ",", ":", ",", ":", "]", ",", "low", "-", "103.939", ",", "high", "-", "103.939", ")" ]
Not necessary in practice
[ "Not", "necessary", "in", "practice" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/style_transfer/utils.py#L95-L99
train
apache/incubator-mxnet
example/svrg_module/api_usage_example/example_inference.py
create_network
def create_network(batch_size, update_freq): """Create a linear regression network for performing SVRG optimization. :return: an instance of mx.io.NDArrayIter :return: an instance of mx.mod.svrgmodule for performing SVRG optimization """ head = '%(asctime)-15s %(message)s' logging.basicConfig(level=logging.INFO, format=head) data = np.random.randint(1, 5, [1000, 2]) #Test_Train data split n_train = int(data.shape[0] * 0.8) weights = np.array([1.0, 2.0]) label = data.dot(weights) di = mx.io.NDArrayIter(data[:n_train, :], label[:n_train], batch_size=batch_size, shuffle=True, label_name='lin_reg_label') val_iter = mx.io.NDArrayIter(data[n_train:, :], label[n_train:], batch_size=batch_size) X = mx.sym.Variable('data') Y = mx.symbol.Variable('lin_reg_label') fully_connected_layer = mx.sym.FullyConnected(data=X, name='fc1', num_hidden=1) lro = mx.sym.LinearRegressionOutput(data=fully_connected_layer, label=Y, name="lro") mod = SVRGModule( symbol=lro, data_names=['data'], label_names=['lin_reg_label'], update_freq=update_freq, logger=logging) return di, val_iter, mod
python
def create_network(batch_size, update_freq): """Create a linear regression network for performing SVRG optimization. :return: an instance of mx.io.NDArrayIter :return: an instance of mx.mod.svrgmodule for performing SVRG optimization """ head = '%(asctime)-15s %(message)s' logging.basicConfig(level=logging.INFO, format=head) data = np.random.randint(1, 5, [1000, 2]) #Test_Train data split n_train = int(data.shape[0] * 0.8) weights = np.array([1.0, 2.0]) label = data.dot(weights) di = mx.io.NDArrayIter(data[:n_train, :], label[:n_train], batch_size=batch_size, shuffle=True, label_name='lin_reg_label') val_iter = mx.io.NDArrayIter(data[n_train:, :], label[n_train:], batch_size=batch_size) X = mx.sym.Variable('data') Y = mx.symbol.Variable('lin_reg_label') fully_connected_layer = mx.sym.FullyConnected(data=X, name='fc1', num_hidden=1) lro = mx.sym.LinearRegressionOutput(data=fully_connected_layer, label=Y, name="lro") mod = SVRGModule( symbol=lro, data_names=['data'], label_names=['lin_reg_label'], update_freq=update_freq, logger=logging) return di, val_iter, mod
[ "def", "create_network", "(", "batch_size", ",", "update_freq", ")", ":", "head", "=", "'%(asctime)-15s %(message)s'", "logging", ".", "basicConfig", "(", "level", "=", "logging", ".", "INFO", ",", "format", "=", "head", ")", "data", "=", "np", ".", "random", ".", "randint", "(", "1", ",", "5", ",", "[", "1000", ",", "2", "]", ")", "#Test_Train data split", "n_train", "=", "int", "(", "data", ".", "shape", "[", "0", "]", "*", "0.8", ")", "weights", "=", "np", ".", "array", "(", "[", "1.0", ",", "2.0", "]", ")", "label", "=", "data", ".", "dot", "(", "weights", ")", "di", "=", "mx", ".", "io", ".", "NDArrayIter", "(", "data", "[", ":", "n_train", ",", ":", "]", ",", "label", "[", ":", "n_train", "]", ",", "batch_size", "=", "batch_size", ",", "shuffle", "=", "True", ",", "label_name", "=", "'lin_reg_label'", ")", "val_iter", "=", "mx", ".", "io", ".", "NDArrayIter", "(", "data", "[", "n_train", ":", ",", ":", "]", ",", "label", "[", "n_train", ":", "]", ",", "batch_size", "=", "batch_size", ")", "X", "=", "mx", ".", "sym", ".", "Variable", "(", "'data'", ")", "Y", "=", "mx", ".", "symbol", ".", "Variable", "(", "'lin_reg_label'", ")", "fully_connected_layer", "=", "mx", ".", "sym", ".", "FullyConnected", "(", "data", "=", "X", ",", "name", "=", "'fc1'", ",", "num_hidden", "=", "1", ")", "lro", "=", "mx", ".", "sym", ".", "LinearRegressionOutput", "(", "data", "=", "fully_connected_layer", ",", "label", "=", "Y", ",", "name", "=", "\"lro\"", ")", "mod", "=", "SVRGModule", "(", "symbol", "=", "lro", ",", "data_names", "=", "[", "'data'", "]", ",", "label_names", "=", "[", "'lin_reg_label'", "]", ",", "update_freq", "=", "update_freq", ",", "logger", "=", "logging", ")", "return", "di", ",", "val_iter", ",", "mod" ]
Create a linear regression network for performing SVRG optimization. :return: an instance of mx.io.NDArrayIter :return: an instance of mx.mod.svrgmodule for performing SVRG optimization
[ "Create", "a", "linear", "regression", "network", "for", "performing", "SVRG", "optimization", ".", ":", "return", ":", "an", "instance", "of", "mx", ".", "io", ".", "NDArrayIter", ":", "return", ":", "an", "instance", "of", "mx", ".", "mod", ".", "svrgmodule", "for", "performing", "SVRG", "optimization" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/svrg_module/api_usage_example/example_inference.py#L64-L91
train
apache/incubator-mxnet
example/gluon/audio/urban_sounds/train.py
evaluate_accuracy
def evaluate_accuracy(data_iterator, net): """Function to evaluate accuracy of any data iterator passed to it as an argument""" acc = mx.metric.Accuracy() for data, label in data_iterator: output = net(data) predictions = nd.argmax(output, axis=1) predictions = predictions.reshape((-1, 1)) acc.update(preds=predictions, labels=label) return acc.get()[1]
python
def evaluate_accuracy(data_iterator, net): """Function to evaluate accuracy of any data iterator passed to it as an argument""" acc = mx.metric.Accuracy() for data, label in data_iterator: output = net(data) predictions = nd.argmax(output, axis=1) predictions = predictions.reshape((-1, 1)) acc.update(preds=predictions, labels=label) return acc.get()[1]
[ "def", "evaluate_accuracy", "(", "data_iterator", ",", "net", ")", ":", "acc", "=", "mx", ".", "metric", ".", "Accuracy", "(", ")", "for", "data", ",", "label", "in", "data_iterator", ":", "output", "=", "net", "(", "data", ")", "predictions", "=", "nd", ".", "argmax", "(", "output", ",", "axis", "=", "1", ")", "predictions", "=", "predictions", ".", "reshape", "(", "(", "-", "1", ",", "1", ")", ")", "acc", ".", "update", "(", "preds", "=", "predictions", ",", "labels", "=", "label", ")", "return", "acc", ".", "get", "(", ")", "[", "1", "]" ]
Function to evaluate accuracy of any data iterator passed to it as an argument
[ "Function", "to", "evaluate", "accuracy", "of", "any", "data", "iterator", "passed", "to", "it", "as", "an", "argument" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/audio/urban_sounds/train.py#L29-L37
train
apache/incubator-mxnet
example/gluon/audio/urban_sounds/train.py
train
def train(train_dir=None, train_csv=None, epochs=30, batch_size=32): """Function responsible for running the training the model.""" if not train_dir or not os.path.exists(train_dir) or not train_csv: warnings.warn("No train directory could be found ") return # Make a dataset from the local folder containing Audio data print("\nMaking an Audio Dataset...\n") tick = time.time() aud_dataset = AudioFolderDataset(train_dir, train_csv=train_csv, file_format='.wav', skip_header=True) tock = time.time() print("Loading the dataset took ", (tock-tick), " seconds.") print("\n=======================================\n") print("Number of output classes = ", len(aud_dataset.synsets)) print("\nThe labels are : \n") print(aud_dataset.synsets) # Get the model to train net = model.get_net(len(aud_dataset.synsets)) print("\nNeural Network = \n") print(net) print("\nModel - Neural Network Generated!\n") print("=======================================\n") #Define the loss - Softmax CE Loss softmax_loss = gluon.loss.SoftmaxCELoss(from_logits=False, sparse_label=True) print("Loss function initialized!\n") print("=======================================\n") #Define the trainer with the optimizer trainer = gluon.Trainer(net.collect_params(), 'adadelta') print("Optimizer - Trainer function initialized!\n") print("=======================================\n") print("Loading the dataset to the Gluon's OOTB Dataloader...") #Getting the data loader out of the AudioDataset and passing the transform from transforms import MFCC aud_transform = MFCC() tick = time.time() audio_train_loader = gluon.data.DataLoader(aud_dataset.transform_first(aud_transform), batch_size=32, shuffle=True) tock = time.time() print("Time taken to load data and apply transform here is ", (tock-tick), " seconds.") print("=======================================\n") print("Starting the training....\n") # Training loop tick = time.time() batch_size = batch_size num_examples = len(aud_dataset) for epoch in range(epochs): cumulative_loss = 0 for data, label in audio_train_loader: with autograd.record(): output = net(data) loss = softmax_loss(output, label) loss.backward() trainer.step(batch_size) cumulative_loss += mx.nd.sum(loss).asscalar() if epoch%5 == 0: train_accuracy = evaluate_accuracy(audio_train_loader, net) print("Epoch {}. Loss: {} Train accuracy : {} ".format(epoch, cumulative_loss/num_examples, train_accuracy)) print("\n------------------------------\n") train_accuracy = evaluate_accuracy(audio_train_loader, net) tock = time.time() print("\nFinal training accuracy: ", train_accuracy) print("Training the sound classification for ", epochs, " epochs, MLP model took ", (tock-tick), " seconds") print("====================== END ======================\n") print("Trying to save the model parameters here...") net.save_parameters("./net.params") print("Saved the model parameters in current directory.")
python
def train(train_dir=None, train_csv=None, epochs=30, batch_size=32): """Function responsible for running the training the model.""" if not train_dir or not os.path.exists(train_dir) or not train_csv: warnings.warn("No train directory could be found ") return # Make a dataset from the local folder containing Audio data print("\nMaking an Audio Dataset...\n") tick = time.time() aud_dataset = AudioFolderDataset(train_dir, train_csv=train_csv, file_format='.wav', skip_header=True) tock = time.time() print("Loading the dataset took ", (tock-tick), " seconds.") print("\n=======================================\n") print("Number of output classes = ", len(aud_dataset.synsets)) print("\nThe labels are : \n") print(aud_dataset.synsets) # Get the model to train net = model.get_net(len(aud_dataset.synsets)) print("\nNeural Network = \n") print(net) print("\nModel - Neural Network Generated!\n") print("=======================================\n") #Define the loss - Softmax CE Loss softmax_loss = gluon.loss.SoftmaxCELoss(from_logits=False, sparse_label=True) print("Loss function initialized!\n") print("=======================================\n") #Define the trainer with the optimizer trainer = gluon.Trainer(net.collect_params(), 'adadelta') print("Optimizer - Trainer function initialized!\n") print("=======================================\n") print("Loading the dataset to the Gluon's OOTB Dataloader...") #Getting the data loader out of the AudioDataset and passing the transform from transforms import MFCC aud_transform = MFCC() tick = time.time() audio_train_loader = gluon.data.DataLoader(aud_dataset.transform_first(aud_transform), batch_size=32, shuffle=True) tock = time.time() print("Time taken to load data and apply transform here is ", (tock-tick), " seconds.") print("=======================================\n") print("Starting the training....\n") # Training loop tick = time.time() batch_size = batch_size num_examples = len(aud_dataset) for epoch in range(epochs): cumulative_loss = 0 for data, label in audio_train_loader: with autograd.record(): output = net(data) loss = softmax_loss(output, label) loss.backward() trainer.step(batch_size) cumulative_loss += mx.nd.sum(loss).asscalar() if epoch%5 == 0: train_accuracy = evaluate_accuracy(audio_train_loader, net) print("Epoch {}. Loss: {} Train accuracy : {} ".format(epoch, cumulative_loss/num_examples, train_accuracy)) print("\n------------------------------\n") train_accuracy = evaluate_accuracy(audio_train_loader, net) tock = time.time() print("\nFinal training accuracy: ", train_accuracy) print("Training the sound classification for ", epochs, " epochs, MLP model took ", (tock-tick), " seconds") print("====================== END ======================\n") print("Trying to save the model parameters here...") net.save_parameters("./net.params") print("Saved the model parameters in current directory.")
[ "def", "train", "(", "train_dir", "=", "None", ",", "train_csv", "=", "None", ",", "epochs", "=", "30", ",", "batch_size", "=", "32", ")", ":", "if", "not", "train_dir", "or", "not", "os", ".", "path", ".", "exists", "(", "train_dir", ")", "or", "not", "train_csv", ":", "warnings", ".", "warn", "(", "\"No train directory could be found \"", ")", "return", "# Make a dataset from the local folder containing Audio data", "print", "(", "\"\\nMaking an Audio Dataset...\\n\"", ")", "tick", "=", "time", ".", "time", "(", ")", "aud_dataset", "=", "AudioFolderDataset", "(", "train_dir", ",", "train_csv", "=", "train_csv", ",", "file_format", "=", "'.wav'", ",", "skip_header", "=", "True", ")", "tock", "=", "time", ".", "time", "(", ")", "print", "(", "\"Loading the dataset took \"", ",", "(", "tock", "-", "tick", ")", ",", "\" seconds.\"", ")", "print", "(", "\"\\n=======================================\\n\"", ")", "print", "(", "\"Number of output classes = \"", ",", "len", "(", "aud_dataset", ".", "synsets", ")", ")", "print", "(", "\"\\nThe labels are : \\n\"", ")", "print", "(", "aud_dataset", ".", "synsets", ")", "# Get the model to train", "net", "=", "model", ".", "get_net", "(", "len", "(", "aud_dataset", ".", "synsets", ")", ")", "print", "(", "\"\\nNeural Network = \\n\"", ")", "print", "(", "net", ")", "print", "(", "\"\\nModel - Neural Network Generated!\\n\"", ")", "print", "(", "\"=======================================\\n\"", ")", "#Define the loss - Softmax CE Loss", "softmax_loss", "=", "gluon", ".", "loss", ".", "SoftmaxCELoss", "(", "from_logits", "=", "False", ",", "sparse_label", "=", "True", ")", "print", "(", "\"Loss function initialized!\\n\"", ")", "print", "(", "\"=======================================\\n\"", ")", "#Define the trainer with the optimizer", "trainer", "=", "gluon", ".", "Trainer", "(", "net", ".", "collect_params", "(", ")", ",", "'adadelta'", ")", "print", "(", "\"Optimizer - Trainer function initialized!\\n\"", ")", "print", "(", "\"=======================================\\n\"", ")", "print", "(", "\"Loading the dataset to the Gluon's OOTB Dataloader...\"", ")", "#Getting the data loader out of the AudioDataset and passing the transform", "from", "transforms", "import", "MFCC", "aud_transform", "=", "MFCC", "(", ")", "tick", "=", "time", ".", "time", "(", ")", "audio_train_loader", "=", "gluon", ".", "data", ".", "DataLoader", "(", "aud_dataset", ".", "transform_first", "(", "aud_transform", ")", ",", "batch_size", "=", "32", ",", "shuffle", "=", "True", ")", "tock", "=", "time", ".", "time", "(", ")", "print", "(", "\"Time taken to load data and apply transform here is \"", ",", "(", "tock", "-", "tick", ")", ",", "\" seconds.\"", ")", "print", "(", "\"=======================================\\n\"", ")", "print", "(", "\"Starting the training....\\n\"", ")", "# Training loop", "tick", "=", "time", ".", "time", "(", ")", "batch_size", "=", "batch_size", "num_examples", "=", "len", "(", "aud_dataset", ")", "for", "epoch", "in", "range", "(", "epochs", ")", ":", "cumulative_loss", "=", "0", "for", "data", ",", "label", "in", "audio_train_loader", ":", "with", "autograd", ".", "record", "(", ")", ":", "output", "=", "net", "(", "data", ")", "loss", "=", "softmax_loss", "(", "output", ",", "label", ")", "loss", ".", "backward", "(", ")", "trainer", ".", "step", "(", "batch_size", ")", "cumulative_loss", "+=", "mx", ".", "nd", ".", "sum", "(", "loss", ")", ".", "asscalar", "(", ")", "if", "epoch", "%", "5", "==", "0", ":", "train_accuracy", "=", "evaluate_accuracy", "(", "audio_train_loader", ",", "net", ")", "print", "(", "\"Epoch {}. Loss: {} Train accuracy : {} \"", ".", "format", "(", "epoch", ",", "cumulative_loss", "/", "num_examples", ",", "train_accuracy", ")", ")", "print", "(", "\"\\n------------------------------\\n\"", ")", "train_accuracy", "=", "evaluate_accuracy", "(", "audio_train_loader", ",", "net", ")", "tock", "=", "time", ".", "time", "(", ")", "print", "(", "\"\\nFinal training accuracy: \"", ",", "train_accuracy", ")", "print", "(", "\"Training the sound classification for \"", ",", "epochs", ",", "\" epochs, MLP model took \"", ",", "(", "tock", "-", "tick", ")", ",", "\" seconds\"", ")", "print", "(", "\"====================== END ======================\\n\"", ")", "print", "(", "\"Trying to save the model parameters here...\"", ")", "net", ".", "save_parameters", "(", "\"./net.params\"", ")", "print", "(", "\"Saved the model parameters in current directory.\"", ")" ]
Function responsible for running the training the model.
[ "Function", "responsible", "for", "running", "the", "training", "the", "model", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/audio/urban_sounds/train.py#L40-L117
train
apache/incubator-mxnet
python/mxnet/engine.py
set_bulk_size
def set_bulk_size(size): """Set size limit on bulk execution. Bulk execution bundles many operators to run together. This can improve performance when running a lot of small operators sequentially. Parameters ---------- size : int Maximum number of operators that can be bundled in a bulk. Returns ------- int Previous bulk size. """ prev = ctypes.c_int() check_call(_LIB.MXEngineSetBulkSize( ctypes.c_int(size), ctypes.byref(prev))) return prev.value
python
def set_bulk_size(size): """Set size limit on bulk execution. Bulk execution bundles many operators to run together. This can improve performance when running a lot of small operators sequentially. Parameters ---------- size : int Maximum number of operators that can be bundled in a bulk. Returns ------- int Previous bulk size. """ prev = ctypes.c_int() check_call(_LIB.MXEngineSetBulkSize( ctypes.c_int(size), ctypes.byref(prev))) return prev.value
[ "def", "set_bulk_size", "(", "size", ")", ":", "prev", "=", "ctypes", ".", "c_int", "(", ")", "check_call", "(", "_LIB", ".", "MXEngineSetBulkSize", "(", "ctypes", ".", "c_int", "(", "size", ")", ",", "ctypes", ".", "byref", "(", "prev", ")", ")", ")", "return", "prev", ".", "value" ]
Set size limit on bulk execution. Bulk execution bundles many operators to run together. This can improve performance when running a lot of small operators sequentially. Parameters ---------- size : int Maximum number of operators that can be bundled in a bulk. Returns ------- int Previous bulk size.
[ "Set", "size", "limit", "on", "bulk", "execution", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/engine.py#L26-L46
train
apache/incubator-mxnet
example/gluon/lipnet/BeamSearch.py
applyLM
def applyLM(parentBeam, childBeam, classes, lm): """ calculate LM score of child beam by taking score from parent beam and bigram probability of last two chars """ if lm and not childBeam.lmApplied: c1 = classes[parentBeam.labeling[-1] if parentBeam.labeling else classes.index(' ')] # first char c2 = classes[childBeam.labeling[-1]] # second char lmFactor = 0.01 # influence of language model bigramProb = lm.getCharBigram(c1, c2) ** lmFactor # probability of seeing first and second char next to each other childBeam.prText = parentBeam.prText * bigramProb # probability of char sequence childBeam.lmApplied = True
python
def applyLM(parentBeam, childBeam, classes, lm): """ calculate LM score of child beam by taking score from parent beam and bigram probability of last two chars """ if lm and not childBeam.lmApplied: c1 = classes[parentBeam.labeling[-1] if parentBeam.labeling else classes.index(' ')] # first char c2 = classes[childBeam.labeling[-1]] # second char lmFactor = 0.01 # influence of language model bigramProb = lm.getCharBigram(c1, c2) ** lmFactor # probability of seeing first and second char next to each other childBeam.prText = parentBeam.prText * bigramProb # probability of char sequence childBeam.lmApplied = True
[ "def", "applyLM", "(", "parentBeam", ",", "childBeam", ",", "classes", ",", "lm", ")", ":", "if", "lm", "and", "not", "childBeam", ".", "lmApplied", ":", "c1", "=", "classes", "[", "parentBeam", ".", "labeling", "[", "-", "1", "]", "if", "parentBeam", ".", "labeling", "else", "classes", ".", "index", "(", "' '", ")", "]", "# first char", "c2", "=", "classes", "[", "childBeam", ".", "labeling", "[", "-", "1", "]", "]", "# second char", "lmFactor", "=", "0.01", "# influence of language model", "bigramProb", "=", "lm", ".", "getCharBigram", "(", "c1", ",", "c2", ")", "**", "lmFactor", "# probability of seeing first and second char next to each other", "childBeam", ".", "prText", "=", "parentBeam", ".", "prText", "*", "bigramProb", "# probability of char sequence", "childBeam", ".", "lmApplied", "=", "True" ]
calculate LM score of child beam by taking score from parent beam and bigram probability of last two chars
[ "calculate", "LM", "score", "of", "child", "beam", "by", "taking", "score", "from", "parent", "beam", "and", "bigram", "probability", "of", "last", "two", "chars" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/BeamSearch.py#L64-L74
train
apache/incubator-mxnet
example/gluon/lipnet/BeamSearch.py
addBeam
def addBeam(beamState, labeling): """ add beam if it does not yet exist """ if labeling not in beamState.entries: beamState.entries[labeling] = BeamEntry()
python
def addBeam(beamState, labeling): """ add beam if it does not yet exist """ if labeling not in beamState.entries: beamState.entries[labeling] = BeamEntry()
[ "def", "addBeam", "(", "beamState", ",", "labeling", ")", ":", "if", "labeling", "not", "in", "beamState", ".", "entries", ":", "beamState", ".", "entries", "[", "labeling", "]", "=", "BeamEntry", "(", ")" ]
add beam if it does not yet exist
[ "add", "beam", "if", "it", "does", "not", "yet", "exist" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/BeamSearch.py#L76-L81
train
apache/incubator-mxnet
example/gluon/lipnet/BeamSearch.py
ctcBeamSearch
def ctcBeamSearch(mat, classes, lm, k, beamWidth): """ beam search as described by the paper of Hwang et al. and the paper of Graves et al. """ blankIdx = len(classes) maxT, maxC = mat.shape # initialise beam state last = BeamState() labeling = () last.entries[labeling] = BeamEntry() last.entries[labeling].prBlank = 1 last.entries[labeling].prTotal = 1 # go over all time-steps for t in range(maxT): curr = BeamState() # get beam-labelings of best beams bestLabelings = last.sort()[0:beamWidth] # go over best beams for labeling in bestLabelings: # probability of paths ending with a non-blank prNonBlank = 0 # in case of non-empty beam if labeling: # probability of paths with repeated last char at the end try: prNonBlank = last.entries[labeling].prNonBlank * mat[t, labeling[-1]] except FloatingPointError: prNonBlank = 0 # probability of paths ending with a blank prBlank = (last.entries[labeling].prTotal) * mat[t, blankIdx] # add beam at current time-step if needed addBeam(curr, labeling) # fill in data curr.entries[labeling].labeling = labeling curr.entries[labeling].prNonBlank += prNonBlank curr.entries[labeling].prBlank += prBlank curr.entries[labeling].prTotal += prBlank + prNonBlank curr.entries[labeling].prText = last.entries[labeling].prText # beam-labeling not changed, therefore also LM score unchanged from curr.entries[labeling].lmApplied = True # LM already applied at previous time-step for this beam-labeling # extend current beam-labeling for c in range(maxC - 1): # add new char to current beam-labeling newLabeling = labeling + (c,) # if new labeling contains duplicate char at the end, only consider paths ending with a blank if labeling and labeling[-1] == c: prNonBlank = mat[t, c] * last.entries[labeling].prBlank else: prNonBlank = mat[t, c] * last.entries[labeling].prTotal # add beam at current time-step if needed addBeam(curr, newLabeling) # fill in data curr.entries[newLabeling].labeling = newLabeling curr.entries[newLabeling].prNonBlank += prNonBlank curr.entries[newLabeling].prTotal += prNonBlank # apply LM applyLM(curr.entries[labeling], curr.entries[newLabeling], classes, lm) # set new beam state last = curr # normalise LM scores according to beam-labeling-length last.norm() # sort by probability bestLabelings = last.sort()[:k] # get most probable labeling output = [] for bestLabeling in bestLabelings: # map labels to chars res = '' for l in bestLabeling: res += classes[l] output.append(res) return output
python
def ctcBeamSearch(mat, classes, lm, k, beamWidth): """ beam search as described by the paper of Hwang et al. and the paper of Graves et al. """ blankIdx = len(classes) maxT, maxC = mat.shape # initialise beam state last = BeamState() labeling = () last.entries[labeling] = BeamEntry() last.entries[labeling].prBlank = 1 last.entries[labeling].prTotal = 1 # go over all time-steps for t in range(maxT): curr = BeamState() # get beam-labelings of best beams bestLabelings = last.sort()[0:beamWidth] # go over best beams for labeling in bestLabelings: # probability of paths ending with a non-blank prNonBlank = 0 # in case of non-empty beam if labeling: # probability of paths with repeated last char at the end try: prNonBlank = last.entries[labeling].prNonBlank * mat[t, labeling[-1]] except FloatingPointError: prNonBlank = 0 # probability of paths ending with a blank prBlank = (last.entries[labeling].prTotal) * mat[t, blankIdx] # add beam at current time-step if needed addBeam(curr, labeling) # fill in data curr.entries[labeling].labeling = labeling curr.entries[labeling].prNonBlank += prNonBlank curr.entries[labeling].prBlank += prBlank curr.entries[labeling].prTotal += prBlank + prNonBlank curr.entries[labeling].prText = last.entries[labeling].prText # beam-labeling not changed, therefore also LM score unchanged from curr.entries[labeling].lmApplied = True # LM already applied at previous time-step for this beam-labeling # extend current beam-labeling for c in range(maxC - 1): # add new char to current beam-labeling newLabeling = labeling + (c,) # if new labeling contains duplicate char at the end, only consider paths ending with a blank if labeling and labeling[-1] == c: prNonBlank = mat[t, c] * last.entries[labeling].prBlank else: prNonBlank = mat[t, c] * last.entries[labeling].prTotal # add beam at current time-step if needed addBeam(curr, newLabeling) # fill in data curr.entries[newLabeling].labeling = newLabeling curr.entries[newLabeling].prNonBlank += prNonBlank curr.entries[newLabeling].prTotal += prNonBlank # apply LM applyLM(curr.entries[labeling], curr.entries[newLabeling], classes, lm) # set new beam state last = curr # normalise LM scores according to beam-labeling-length last.norm() # sort by probability bestLabelings = last.sort()[:k] # get most probable labeling output = [] for bestLabeling in bestLabelings: # map labels to chars res = '' for l in bestLabeling: res += classes[l] output.append(res) return output
[ "def", "ctcBeamSearch", "(", "mat", ",", "classes", ",", "lm", ",", "k", ",", "beamWidth", ")", ":", "blankIdx", "=", "len", "(", "classes", ")", "maxT", ",", "maxC", "=", "mat", ".", "shape", "# initialise beam state", "last", "=", "BeamState", "(", ")", "labeling", "=", "(", ")", "last", ".", "entries", "[", "labeling", "]", "=", "BeamEntry", "(", ")", "last", ".", "entries", "[", "labeling", "]", ".", "prBlank", "=", "1", "last", ".", "entries", "[", "labeling", "]", ".", "prTotal", "=", "1", "# go over all time-steps", "for", "t", "in", "range", "(", "maxT", ")", ":", "curr", "=", "BeamState", "(", ")", "# get beam-labelings of best beams", "bestLabelings", "=", "last", ".", "sort", "(", ")", "[", "0", ":", "beamWidth", "]", "# go over best beams", "for", "labeling", "in", "bestLabelings", ":", "# probability of paths ending with a non-blank", "prNonBlank", "=", "0", "# in case of non-empty beam", "if", "labeling", ":", "# probability of paths with repeated last char at the end", "try", ":", "prNonBlank", "=", "last", ".", "entries", "[", "labeling", "]", ".", "prNonBlank", "*", "mat", "[", "t", ",", "labeling", "[", "-", "1", "]", "]", "except", "FloatingPointError", ":", "prNonBlank", "=", "0", "# probability of paths ending with a blank", "prBlank", "=", "(", "last", ".", "entries", "[", "labeling", "]", ".", "prTotal", ")", "*", "mat", "[", "t", ",", "blankIdx", "]", "# add beam at current time-step if needed", "addBeam", "(", "curr", ",", "labeling", ")", "# fill in data", "curr", ".", "entries", "[", "labeling", "]", ".", "labeling", "=", "labeling", "curr", ".", "entries", "[", "labeling", "]", ".", "prNonBlank", "+=", "prNonBlank", "curr", ".", "entries", "[", "labeling", "]", ".", "prBlank", "+=", "prBlank", "curr", ".", "entries", "[", "labeling", "]", ".", "prTotal", "+=", "prBlank", "+", "prNonBlank", "curr", ".", "entries", "[", "labeling", "]", ".", "prText", "=", "last", ".", "entries", "[", "labeling", "]", ".", "prText", "# beam-labeling not changed, therefore also LM score unchanged from", "curr", ".", "entries", "[", "labeling", "]", ".", "lmApplied", "=", "True", "# LM already applied at previous time-step for this beam-labeling", "# extend current beam-labeling", "for", "c", "in", "range", "(", "maxC", "-", "1", ")", ":", "# add new char to current beam-labeling", "newLabeling", "=", "labeling", "+", "(", "c", ",", ")", "# if new labeling contains duplicate char at the end, only consider paths ending with a blank", "if", "labeling", "and", "labeling", "[", "-", "1", "]", "==", "c", ":", "prNonBlank", "=", "mat", "[", "t", ",", "c", "]", "*", "last", ".", "entries", "[", "labeling", "]", ".", "prBlank", "else", ":", "prNonBlank", "=", "mat", "[", "t", ",", "c", "]", "*", "last", ".", "entries", "[", "labeling", "]", ".", "prTotal", "# add beam at current time-step if needed", "addBeam", "(", "curr", ",", "newLabeling", ")", "# fill in data", "curr", ".", "entries", "[", "newLabeling", "]", ".", "labeling", "=", "newLabeling", "curr", ".", "entries", "[", "newLabeling", "]", ".", "prNonBlank", "+=", "prNonBlank", "curr", ".", "entries", "[", "newLabeling", "]", ".", "prTotal", "+=", "prNonBlank", "# apply LM", "applyLM", "(", "curr", ".", "entries", "[", "labeling", "]", ",", "curr", ".", "entries", "[", "newLabeling", "]", ",", "classes", ",", "lm", ")", "# set new beam state", "last", "=", "curr", "# normalise LM scores according to beam-labeling-length", "last", ".", "norm", "(", ")", "# sort by probability", "bestLabelings", "=", "last", ".", "sort", "(", ")", "[", ":", "k", "]", "# get most probable labeling", "output", "=", "[", "]", "for", "bestLabeling", "in", "bestLabelings", ":", "# map labels to chars", "res", "=", "''", "for", "l", "in", "bestLabeling", ":", "res", "+=", "classes", "[", "l", "]", "output", ".", "append", "(", "res", ")", "return", "output" ]
beam search as described by the paper of Hwang et al. and the paper of Graves et al.
[ "beam", "search", "as", "described", "by", "the", "paper", "of", "Hwang", "et", "al", ".", "and", "the", "paper", "of", "Graves", "et", "al", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/BeamSearch.py#L83-L170
train
apache/incubator-mxnet
example/gluon/lipnet/BeamSearch.py
BeamState.norm
def norm(self): """ length-normalise LM score """ for (k, _) in self.entries.items(): labelingLen = len(self.entries[k].labeling) self.entries[k].prText = self.entries[k].prText ** (1.0 / (labelingLen if labelingLen else 1.0))
python
def norm(self): """ length-normalise LM score """ for (k, _) in self.entries.items(): labelingLen = len(self.entries[k].labeling) self.entries[k].prText = self.entries[k].prText ** (1.0 / (labelingLen if labelingLen else 1.0))
[ "def", "norm", "(", "self", ")", ":", "for", "(", "k", ",", "_", ")", "in", "self", ".", "entries", ".", "items", "(", ")", ":", "labelingLen", "=", "len", "(", "self", ".", "entries", "[", "k", "]", ".", "labeling", ")", "self", ".", "entries", "[", "k", "]", ".", "prText", "=", "self", ".", "entries", "[", "k", "]", ".", "prText", "**", "(", "1.0", "/", "(", "labelingLen", "if", "labelingLen", "else", "1.0", ")", ")" ]
length-normalise LM score
[ "length", "-", "normalise", "LM", "score" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/BeamSearch.py#L48-L54
train
apache/incubator-mxnet
example/gluon/lipnet/BeamSearch.py
BeamState.sort
def sort(self): """ return beam-labelings, sorted by probability """ beams = [v for (_, v) in self.entries.items()] sortedBeams = sorted(beams, reverse=True, key=lambda x: x.prTotal*x.prText) return [x.labeling for x in sortedBeams]
python
def sort(self): """ return beam-labelings, sorted by probability """ beams = [v for (_, v) in self.entries.items()] sortedBeams = sorted(beams, reverse=True, key=lambda x: x.prTotal*x.prText) return [x.labeling for x in sortedBeams]
[ "def", "sort", "(", "self", ")", ":", "beams", "=", "[", "v", "for", "(", "_", ",", "v", ")", "in", "self", ".", "entries", ".", "items", "(", ")", "]", "sortedBeams", "=", "sorted", "(", "beams", ",", "reverse", "=", "True", ",", "key", "=", "lambda", "x", ":", "x", ".", "prTotal", "*", "x", ".", "prText", ")", "return", "[", "x", ".", "labeling", "for", "x", "in", "sortedBeams", "]" ]
return beam-labelings, sorted by probability
[ "return", "beam", "-", "labelings", "sorted", "by", "probability" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/lipnet/BeamSearch.py#L56-L62
train
apache/incubator-mxnet
example/image-classification/symbols/lenet.py
get_loc
def get_loc(data, attr={'lr_mult':'0.01'}): """ the localisation network in lenet-stn, it will increase acc about more than 1%, when num-epoch >=15 """ loc = mx.symbol.Convolution(data=data, num_filter=30, kernel=(5, 5), stride=(2,2)) loc = mx.symbol.Activation(data = loc, act_type='relu') loc = mx.symbol.Pooling(data=loc, kernel=(2, 2), stride=(2, 2), pool_type='max') loc = mx.symbol.Convolution(data=loc, num_filter=60, kernel=(3, 3), stride=(1,1), pad=(1, 1)) loc = mx.symbol.Activation(data = loc, act_type='relu') loc = mx.symbol.Pooling(data=loc, global_pool=True, kernel=(2, 2), pool_type='avg') loc = mx.symbol.Flatten(data=loc) loc = mx.symbol.FullyConnected(data=loc, num_hidden=6, name="stn_loc", attr=attr) return loc
python
def get_loc(data, attr={'lr_mult':'0.01'}): """ the localisation network in lenet-stn, it will increase acc about more than 1%, when num-epoch >=15 """ loc = mx.symbol.Convolution(data=data, num_filter=30, kernel=(5, 5), stride=(2,2)) loc = mx.symbol.Activation(data = loc, act_type='relu') loc = mx.symbol.Pooling(data=loc, kernel=(2, 2), stride=(2, 2), pool_type='max') loc = mx.symbol.Convolution(data=loc, num_filter=60, kernel=(3, 3), stride=(1,1), pad=(1, 1)) loc = mx.symbol.Activation(data = loc, act_type='relu') loc = mx.symbol.Pooling(data=loc, global_pool=True, kernel=(2, 2), pool_type='avg') loc = mx.symbol.Flatten(data=loc) loc = mx.symbol.FullyConnected(data=loc, num_hidden=6, name="stn_loc", attr=attr) return loc
[ "def", "get_loc", "(", "data", ",", "attr", "=", "{", "'lr_mult'", ":", "'0.01'", "}", ")", ":", "loc", "=", "mx", ".", "symbol", ".", "Convolution", "(", "data", "=", "data", ",", "num_filter", "=", "30", ",", "kernel", "=", "(", "5", ",", "5", ")", ",", "stride", "=", "(", "2", ",", "2", ")", ")", "loc", "=", "mx", ".", "symbol", ".", "Activation", "(", "data", "=", "loc", ",", "act_type", "=", "'relu'", ")", "loc", "=", "mx", ".", "symbol", ".", "Pooling", "(", "data", "=", "loc", ",", "kernel", "=", "(", "2", ",", "2", ")", ",", "stride", "=", "(", "2", ",", "2", ")", ",", "pool_type", "=", "'max'", ")", "loc", "=", "mx", ".", "symbol", ".", "Convolution", "(", "data", "=", "loc", ",", "num_filter", "=", "60", ",", "kernel", "=", "(", "3", ",", "3", ")", ",", "stride", "=", "(", "1", ",", "1", ")", ",", "pad", "=", "(", "1", ",", "1", ")", ")", "loc", "=", "mx", ".", "symbol", ".", "Activation", "(", "data", "=", "loc", ",", "act_type", "=", "'relu'", ")", "loc", "=", "mx", ".", "symbol", ".", "Pooling", "(", "data", "=", "loc", ",", "global_pool", "=", "True", ",", "kernel", "=", "(", "2", ",", "2", ")", ",", "pool_type", "=", "'avg'", ")", "loc", "=", "mx", ".", "symbol", ".", "Flatten", "(", "data", "=", "loc", ")", "loc", "=", "mx", ".", "symbol", ".", "FullyConnected", "(", "data", "=", "loc", ",", "num_hidden", "=", "6", ",", "name", "=", "\"stn_loc\"", ",", "attr", "=", "attr", ")", "return", "loc" ]
the localisation network in lenet-stn, it will increase acc about more than 1%, when num-epoch >=15
[ "the", "localisation", "network", "in", "lenet", "-", "stn", "it", "will", "increase", "acc", "about", "more", "than", "1%", "when", "num", "-", "epoch", ">", "=", "15" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/image-classification/symbols/lenet.py#L25-L38
train
apache/incubator-mxnet
example/ssd/demo.py
get_detector
def get_detector(net, prefix, epoch, data_shape, mean_pixels, ctx, num_class, nms_thresh=0.5, force_nms=True, nms_topk=400): """ wrapper for initialize a detector Parameters: ---------- net : str test network name prefix : str load model prefix epoch : int load model epoch data_shape : int resize image shape mean_pixels : tuple (float, float, float) mean pixel values (R, G, B) ctx : mx.ctx running context, mx.cpu() or mx.gpu(?) num_class : int number of classes nms_thresh : float non-maximum suppression threshold force_nms : bool force suppress different categories """ if net is not None: if isinstance(data_shape, tuple): data_shape = data_shape[0] net = get_symbol(net, data_shape, num_classes=num_class, nms_thresh=nms_thresh, force_nms=force_nms, nms_topk=nms_topk) detector = Detector(net, prefix, epoch, data_shape, mean_pixels, ctx=ctx) return detector
python
def get_detector(net, prefix, epoch, data_shape, mean_pixels, ctx, num_class, nms_thresh=0.5, force_nms=True, nms_topk=400): """ wrapper for initialize a detector Parameters: ---------- net : str test network name prefix : str load model prefix epoch : int load model epoch data_shape : int resize image shape mean_pixels : tuple (float, float, float) mean pixel values (R, G, B) ctx : mx.ctx running context, mx.cpu() or mx.gpu(?) num_class : int number of classes nms_thresh : float non-maximum suppression threshold force_nms : bool force suppress different categories """ if net is not None: if isinstance(data_shape, tuple): data_shape = data_shape[0] net = get_symbol(net, data_shape, num_classes=num_class, nms_thresh=nms_thresh, force_nms=force_nms, nms_topk=nms_topk) detector = Detector(net, prefix, epoch, data_shape, mean_pixels, ctx=ctx) return detector
[ "def", "get_detector", "(", "net", ",", "prefix", ",", "epoch", ",", "data_shape", ",", "mean_pixels", ",", "ctx", ",", "num_class", ",", "nms_thresh", "=", "0.5", ",", "force_nms", "=", "True", ",", "nms_topk", "=", "400", ")", ":", "if", "net", "is", "not", "None", ":", "if", "isinstance", "(", "data_shape", ",", "tuple", ")", ":", "data_shape", "=", "data_shape", "[", "0", "]", "net", "=", "get_symbol", "(", "net", ",", "data_shape", ",", "num_classes", "=", "num_class", ",", "nms_thresh", "=", "nms_thresh", ",", "force_nms", "=", "force_nms", ",", "nms_topk", "=", "nms_topk", ")", "detector", "=", "Detector", "(", "net", ",", "prefix", ",", "epoch", ",", "data_shape", ",", "mean_pixels", ",", "ctx", "=", "ctx", ")", "return", "detector" ]
wrapper for initialize a detector Parameters: ---------- net : str test network name prefix : str load model prefix epoch : int load model epoch data_shape : int resize image shape mean_pixels : tuple (float, float, float) mean pixel values (R, G, B) ctx : mx.ctx running context, mx.cpu() or mx.gpu(?) num_class : int number of classes nms_thresh : float non-maximum suppression threshold force_nms : bool force suppress different categories
[ "wrapper", "for", "initialize", "a", "detector" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ssd/demo.py#L32-L64
train
apache/incubator-mxnet
example/ssd/demo.py
parse_class_names
def parse_class_names(class_names): """ parse # classes and class_names if applicable """ if len(class_names) > 0: if os.path.isfile(class_names): # try to open it to read class names with open(class_names, 'r') as f: class_names = [l.strip() for l in f.readlines()] else: class_names = [c.strip() for c in class_names.split(',')] for name in class_names: assert len(name) > 0 else: raise RuntimeError("No valid class_name provided...") return class_names
python
def parse_class_names(class_names): """ parse # classes and class_names if applicable """ if len(class_names) > 0: if os.path.isfile(class_names): # try to open it to read class names with open(class_names, 'r') as f: class_names = [l.strip() for l in f.readlines()] else: class_names = [c.strip() for c in class_names.split(',')] for name in class_names: assert len(name) > 0 else: raise RuntimeError("No valid class_name provided...") return class_names
[ "def", "parse_class_names", "(", "class_names", ")", ":", "if", "len", "(", "class_names", ")", ">", "0", ":", "if", "os", ".", "path", ".", "isfile", "(", "class_names", ")", ":", "# try to open it to read class names", "with", "open", "(", "class_names", ",", "'r'", ")", "as", "f", ":", "class_names", "=", "[", "l", ".", "strip", "(", ")", "for", "l", "in", "f", ".", "readlines", "(", ")", "]", "else", ":", "class_names", "=", "[", "c", ".", "strip", "(", ")", "for", "c", "in", "class_names", ".", "split", "(", "','", ")", "]", "for", "name", "in", "class_names", ":", "assert", "len", "(", "name", ")", ">", "0", "else", ":", "raise", "RuntimeError", "(", "\"No valid class_name provided...\"", ")", "return", "class_names" ]
parse # classes and class_names if applicable
[ "parse", "#", "classes", "and", "class_names", "if", "applicable" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ssd/demo.py#L117-L130
train
apache/incubator-mxnet
example/ssd/demo.py
parse_data_shape
def parse_data_shape(data_shape_str): """Parse string to tuple or int""" ds = data_shape_str.strip().split(',') if len(ds) == 1: data_shape = (int(ds[0]), int(ds[0])) elif len(ds) == 2: data_shape = (int(ds[0]), int(ds[1])) else: raise ValueError("Unexpected data_shape: %s", data_shape_str) return data_shape
python
def parse_data_shape(data_shape_str): """Parse string to tuple or int""" ds = data_shape_str.strip().split(',') if len(ds) == 1: data_shape = (int(ds[0]), int(ds[0])) elif len(ds) == 2: data_shape = (int(ds[0]), int(ds[1])) else: raise ValueError("Unexpected data_shape: %s", data_shape_str) return data_shape
[ "def", "parse_data_shape", "(", "data_shape_str", ")", ":", "ds", "=", "data_shape_str", ".", "strip", "(", ")", ".", "split", "(", "','", ")", "if", "len", "(", "ds", ")", "==", "1", ":", "data_shape", "=", "(", "int", "(", "ds", "[", "0", "]", ")", ",", "int", "(", "ds", "[", "0", "]", ")", ")", "elif", "len", "(", "ds", ")", "==", "2", ":", "data_shape", "=", "(", "int", "(", "ds", "[", "0", "]", ")", ",", "int", "(", "ds", "[", "1", "]", ")", ")", "else", ":", "raise", "ValueError", "(", "\"Unexpected data_shape: %s\"", ",", "data_shape_str", ")", "return", "data_shape" ]
Parse string to tuple or int
[ "Parse", "string", "to", "tuple", "or", "int" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ssd/demo.py#L141-L150
train
apache/incubator-mxnet
example/kaggle-ndsb2/Train.py
get_lenet
def get_lenet(): """ A lenet style net, takes difference of each frame as input. """ source = mx.sym.Variable("data") source = (source - 128) * (1.0/128) frames = mx.sym.SliceChannel(source, num_outputs=30) diffs = [frames[i+1] - frames[i] for i in range(29)] source = mx.sym.Concat(*diffs) net = mx.sym.Convolution(source, kernel=(5, 5), num_filter=40) net = mx.sym.BatchNorm(net, fix_gamma=True) net = mx.sym.Activation(net, act_type="relu") net = mx.sym.Pooling(net, pool_type="max", kernel=(2,2), stride=(2,2)) net = mx.sym.Convolution(net, kernel=(3, 3), num_filter=40) net = mx.sym.BatchNorm(net, fix_gamma=True) net = mx.sym.Activation(net, act_type="relu") net = mx.sym.Pooling(net, pool_type="max", kernel=(2,2), stride=(2,2)) # first fullc flatten = mx.symbol.Flatten(net) flatten = mx.symbol.Dropout(flatten) fc1 = mx.symbol.FullyConnected(data=flatten, num_hidden=600) # Name the final layer as softmax so it auto matches the naming of data iterator # Otherwise we can also change the provide_data in the data iter return mx.symbol.LogisticRegressionOutput(data=fc1, name='softmax')
python
def get_lenet(): """ A lenet style net, takes difference of each frame as input. """ source = mx.sym.Variable("data") source = (source - 128) * (1.0/128) frames = mx.sym.SliceChannel(source, num_outputs=30) diffs = [frames[i+1] - frames[i] for i in range(29)] source = mx.sym.Concat(*diffs) net = mx.sym.Convolution(source, kernel=(5, 5), num_filter=40) net = mx.sym.BatchNorm(net, fix_gamma=True) net = mx.sym.Activation(net, act_type="relu") net = mx.sym.Pooling(net, pool_type="max", kernel=(2,2), stride=(2,2)) net = mx.sym.Convolution(net, kernel=(3, 3), num_filter=40) net = mx.sym.BatchNorm(net, fix_gamma=True) net = mx.sym.Activation(net, act_type="relu") net = mx.sym.Pooling(net, pool_type="max", kernel=(2,2), stride=(2,2)) # first fullc flatten = mx.symbol.Flatten(net) flatten = mx.symbol.Dropout(flatten) fc1 = mx.symbol.FullyConnected(data=flatten, num_hidden=600) # Name the final layer as softmax so it auto matches the naming of data iterator # Otherwise we can also change the provide_data in the data iter return mx.symbol.LogisticRegressionOutput(data=fc1, name='softmax')
[ "def", "get_lenet", "(", ")", ":", "source", "=", "mx", ".", "sym", ".", "Variable", "(", "\"data\"", ")", "source", "=", "(", "source", "-", "128", ")", "*", "(", "1.0", "/", "128", ")", "frames", "=", "mx", ".", "sym", ".", "SliceChannel", "(", "source", ",", "num_outputs", "=", "30", ")", "diffs", "=", "[", "frames", "[", "i", "+", "1", "]", "-", "frames", "[", "i", "]", "for", "i", "in", "range", "(", "29", ")", "]", "source", "=", "mx", ".", "sym", ".", "Concat", "(", "*", "diffs", ")", "net", "=", "mx", ".", "sym", ".", "Convolution", "(", "source", ",", "kernel", "=", "(", "5", ",", "5", ")", ",", "num_filter", "=", "40", ")", "net", "=", "mx", ".", "sym", ".", "BatchNorm", "(", "net", ",", "fix_gamma", "=", "True", ")", "net", "=", "mx", ".", "sym", ".", "Activation", "(", "net", ",", "act_type", "=", "\"relu\"", ")", "net", "=", "mx", ".", "sym", ".", "Pooling", "(", "net", ",", "pool_type", "=", "\"max\"", ",", "kernel", "=", "(", "2", ",", "2", ")", ",", "stride", "=", "(", "2", ",", "2", ")", ")", "net", "=", "mx", ".", "sym", ".", "Convolution", "(", "net", ",", "kernel", "=", "(", "3", ",", "3", ")", ",", "num_filter", "=", "40", ")", "net", "=", "mx", ".", "sym", ".", "BatchNorm", "(", "net", ",", "fix_gamma", "=", "True", ")", "net", "=", "mx", ".", "sym", ".", "Activation", "(", "net", ",", "act_type", "=", "\"relu\"", ")", "net", "=", "mx", ".", "sym", ".", "Pooling", "(", "net", ",", "pool_type", "=", "\"max\"", ",", "kernel", "=", "(", "2", ",", "2", ")", ",", "stride", "=", "(", "2", ",", "2", ")", ")", "# first fullc", "flatten", "=", "mx", ".", "symbol", ".", "Flatten", "(", "net", ")", "flatten", "=", "mx", ".", "symbol", ".", "Dropout", "(", "flatten", ")", "fc1", "=", "mx", ".", "symbol", ".", "FullyConnected", "(", "data", "=", "flatten", ",", "num_hidden", "=", "600", ")", "# Name the final layer as softmax so it auto matches the naming of data iterator", "# Otherwise we can also change the provide_data in the data iter", "return", "mx", ".", "symbol", ".", "LogisticRegressionOutput", "(", "data", "=", "fc1", ",", "name", "=", "'softmax'", ")" ]
A lenet style net, takes difference of each frame as input.
[ "A", "lenet", "style", "net", "takes", "difference", "of", "each", "frame", "as", "input", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/kaggle-ndsb2/Train.py#L33-L55
train
apache/incubator-mxnet
example/kaggle-ndsb2/Train.py
CRPS
def CRPS(label, pred): """ Custom evaluation metric on CRPS. """ for i in range(pred.shape[0]): for j in range(pred.shape[1] - 1): if pred[i, j] > pred[i, j + 1]: pred[i, j + 1] = pred[i, j] return np.sum(np.square(label - pred)) / label.size
python
def CRPS(label, pred): """ Custom evaluation metric on CRPS. """ for i in range(pred.shape[0]): for j in range(pred.shape[1] - 1): if pred[i, j] > pred[i, j + 1]: pred[i, j + 1] = pred[i, j] return np.sum(np.square(label - pred)) / label.size
[ "def", "CRPS", "(", "label", ",", "pred", ")", ":", "for", "i", "in", "range", "(", "pred", ".", "shape", "[", "0", "]", ")", ":", "for", "j", "in", "range", "(", "pred", ".", "shape", "[", "1", "]", "-", "1", ")", ":", "if", "pred", "[", "i", ",", "j", "]", ">", "pred", "[", "i", ",", "j", "+", "1", "]", ":", "pred", "[", "i", ",", "j", "+", "1", "]", "=", "pred", "[", "i", ",", "j", "]", "return", "np", ".", "sum", "(", "np", ".", "square", "(", "label", "-", "pred", ")", ")", "/", "label", ".", "size" ]
Custom evaluation metric on CRPS.
[ "Custom", "evaluation", "metric", "on", "CRPS", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/kaggle-ndsb2/Train.py#L57-L64
train
apache/incubator-mxnet
example/kaggle-ndsb2/Train.py
encode_label
def encode_label(label_data): """Run encoding to encode the label into the CDF target. """ systole = label_data[:, 1] diastole = label_data[:, 2] systole_encode = np.array([ (x < np.arange(600)) for x in systole ], dtype=np.uint8) diastole_encode = np.array([ (x < np.arange(600)) for x in diastole ], dtype=np.uint8) return systole_encode, diastole_encode
python
def encode_label(label_data): """Run encoding to encode the label into the CDF target. """ systole = label_data[:, 1] diastole = label_data[:, 2] systole_encode = np.array([ (x < np.arange(600)) for x in systole ], dtype=np.uint8) diastole_encode = np.array([ (x < np.arange(600)) for x in diastole ], dtype=np.uint8) return systole_encode, diastole_encode
[ "def", "encode_label", "(", "label_data", ")", ":", "systole", "=", "label_data", "[", ":", ",", "1", "]", "diastole", "=", "label_data", "[", ":", ",", "2", "]", "systole_encode", "=", "np", ".", "array", "(", "[", "(", "x", "<", "np", ".", "arange", "(", "600", ")", ")", "for", "x", "in", "systole", "]", ",", "dtype", "=", "np", ".", "uint8", ")", "diastole_encode", "=", "np", ".", "array", "(", "[", "(", "x", "<", "np", ".", "arange", "(", "600", ")", ")", "for", "x", "in", "diastole", "]", ",", "dtype", "=", "np", ".", "uint8", ")", "return", "systole_encode", ",", "diastole_encode" ]
Run encoding to encode the label into the CDF target.
[ "Run", "encoding", "to", "encode", "the", "label", "into", "the", "CDF", "target", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/kaggle-ndsb2/Train.py#L69-L80
train
apache/incubator-mxnet
example/rcnn/symimdb/coco.py
coco._load_annotation
def _load_annotation(self, _coco, coco_ind_to_class_ind, index): """ coco ann: [u'segmentation', u'area', u'iscrowd', u'image_id', u'bbox', u'category_id', u'id'] iscrowd: crowd instances are handled by marking their overlaps with all categories to -1 and later excluded in training bbox: [x1, y1, w, h] :param index: coco image id :return: roidb entry """ im_ann = _coco.loadImgs(index)[0] filename = self._image_file_tmpl.format(im_ann['file_name']) width = im_ann['width'] height = im_ann['height'] annIds = _coco.getAnnIds(imgIds=index, iscrowd=None) objs = _coco.loadAnns(annIds) # sanitize bboxes valid_objs = [] for obj in objs: x, y, w, h = obj['bbox'] x1 = np.max((0, x)) y1 = np.max((0, y)) x2 = np.min((width - 1, x1 + np.max((0, w - 1)))) y2 = np.min((height - 1, y1 + np.max((0, h - 1)))) if obj['area'] > 0 and x2 >= x1 and y2 >= y1: obj['clean_bbox'] = [x1, y1, x2, y2] valid_objs.append(obj) objs = valid_objs num_objs = len(objs) boxes = np.zeros((num_objs, 4), dtype=np.uint16) gt_classes = np.zeros((num_objs,), dtype=np.int32) for ix, obj in enumerate(objs): cls = coco_ind_to_class_ind[obj['category_id']] boxes[ix, :] = obj['clean_bbox'] gt_classes[ix] = cls roi_rec = {'index': index, 'image': filename, 'height': height, 'width': width, 'boxes': boxes, 'gt_classes': gt_classes, 'flipped': False} return roi_rec
python
def _load_annotation(self, _coco, coco_ind_to_class_ind, index): """ coco ann: [u'segmentation', u'area', u'iscrowd', u'image_id', u'bbox', u'category_id', u'id'] iscrowd: crowd instances are handled by marking their overlaps with all categories to -1 and later excluded in training bbox: [x1, y1, w, h] :param index: coco image id :return: roidb entry """ im_ann = _coco.loadImgs(index)[0] filename = self._image_file_tmpl.format(im_ann['file_name']) width = im_ann['width'] height = im_ann['height'] annIds = _coco.getAnnIds(imgIds=index, iscrowd=None) objs = _coco.loadAnns(annIds) # sanitize bboxes valid_objs = [] for obj in objs: x, y, w, h = obj['bbox'] x1 = np.max((0, x)) y1 = np.max((0, y)) x2 = np.min((width - 1, x1 + np.max((0, w - 1)))) y2 = np.min((height - 1, y1 + np.max((0, h - 1)))) if obj['area'] > 0 and x2 >= x1 and y2 >= y1: obj['clean_bbox'] = [x1, y1, x2, y2] valid_objs.append(obj) objs = valid_objs num_objs = len(objs) boxes = np.zeros((num_objs, 4), dtype=np.uint16) gt_classes = np.zeros((num_objs,), dtype=np.int32) for ix, obj in enumerate(objs): cls = coco_ind_to_class_ind[obj['category_id']] boxes[ix, :] = obj['clean_bbox'] gt_classes[ix] = cls roi_rec = {'index': index, 'image': filename, 'height': height, 'width': width, 'boxes': boxes, 'gt_classes': gt_classes, 'flipped': False} return roi_rec
[ "def", "_load_annotation", "(", "self", ",", "_coco", ",", "coco_ind_to_class_ind", ",", "index", ")", ":", "im_ann", "=", "_coco", ".", "loadImgs", "(", "index", ")", "[", "0", "]", "filename", "=", "self", ".", "_image_file_tmpl", ".", "format", "(", "im_ann", "[", "'file_name'", "]", ")", "width", "=", "im_ann", "[", "'width'", "]", "height", "=", "im_ann", "[", "'height'", "]", "annIds", "=", "_coco", ".", "getAnnIds", "(", "imgIds", "=", "index", ",", "iscrowd", "=", "None", ")", "objs", "=", "_coco", ".", "loadAnns", "(", "annIds", ")", "# sanitize bboxes", "valid_objs", "=", "[", "]", "for", "obj", "in", "objs", ":", "x", ",", "y", ",", "w", ",", "h", "=", "obj", "[", "'bbox'", "]", "x1", "=", "np", ".", "max", "(", "(", "0", ",", "x", ")", ")", "y1", "=", "np", ".", "max", "(", "(", "0", ",", "y", ")", ")", "x2", "=", "np", ".", "min", "(", "(", "width", "-", "1", ",", "x1", "+", "np", ".", "max", "(", "(", "0", ",", "w", "-", "1", ")", ")", ")", ")", "y2", "=", "np", ".", "min", "(", "(", "height", "-", "1", ",", "y1", "+", "np", ".", "max", "(", "(", "0", ",", "h", "-", "1", ")", ")", ")", ")", "if", "obj", "[", "'area'", "]", ">", "0", "and", "x2", ">=", "x1", "and", "y2", ">=", "y1", ":", "obj", "[", "'clean_bbox'", "]", "=", "[", "x1", ",", "y1", ",", "x2", ",", "y2", "]", "valid_objs", ".", "append", "(", "obj", ")", "objs", "=", "valid_objs", "num_objs", "=", "len", "(", "objs", ")", "boxes", "=", "np", ".", "zeros", "(", "(", "num_objs", ",", "4", ")", ",", "dtype", "=", "np", ".", "uint16", ")", "gt_classes", "=", "np", ".", "zeros", "(", "(", "num_objs", ",", ")", ",", "dtype", "=", "np", ".", "int32", ")", "for", "ix", ",", "obj", "in", "enumerate", "(", "objs", ")", ":", "cls", "=", "coco_ind_to_class_ind", "[", "obj", "[", "'category_id'", "]", "]", "boxes", "[", "ix", ",", ":", "]", "=", "obj", "[", "'clean_bbox'", "]", "gt_classes", "[", "ix", "]", "=", "cls", "roi_rec", "=", "{", "'index'", ":", "index", ",", "'image'", ":", "filename", ",", "'height'", ":", "height", ",", "'width'", ":", "width", ",", "'boxes'", ":", "boxes", ",", "'gt_classes'", ":", "gt_classes", ",", "'flipped'", ":", "False", "}", "return", "roi_rec" ]
coco ann: [u'segmentation', u'area', u'iscrowd', u'image_id', u'bbox', u'category_id', u'id'] iscrowd: crowd instances are handled by marking their overlaps with all categories to -1 and later excluded in training bbox: [x1, y1, w, h] :param index: coco image id :return: roidb entry
[ "coco", "ann", ":", "[", "u", "segmentation", "u", "area", "u", "iscrowd", "u", "image_id", "u", "bbox", "u", "category_id", "u", "id", "]", "iscrowd", ":", "crowd", "instances", "are", "handled", "by", "marking", "their", "overlaps", "with", "all", "categories", "to", "-", "1", "and", "later", "excluded", "in", "training", "bbox", ":", "[", "x1", "y1", "w", "h", "]", ":", "param", "index", ":", "coco", "image", "id", ":", "return", ":", "roidb", "entry" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/rcnn/symimdb/coco.py#L78-L125
train
apache/incubator-mxnet
example/rcnn/symimdb/coco.py
coco._write_coco_results
def _write_coco_results(self, _coco, detections): """ example results [{"image_id": 42, "category_id": 18, "bbox": [258.15,41.29,348.26,243.78], "score": 0.236}, ...] """ cats = [cat['name'] for cat in _coco.loadCats(_coco.getCatIds())] class_to_coco_ind = dict(zip(cats, _coco.getCatIds())) results = [] for cls_ind, cls in enumerate(self.classes): if cls == '__background__': continue logger.info('collecting %s results (%d/%d)' % (cls, cls_ind, self.num_classes - 1)) coco_cat_id = class_to_coco_ind[cls] results.extend(self._coco_results_one_category(detections[cls_ind], coco_cat_id)) logger.info('writing results json to %s' % self._result_file) with open(self._result_file, 'w') as f: json.dump(results, f, sort_keys=True, indent=4)
python
def _write_coco_results(self, _coco, detections): """ example results [{"image_id": 42, "category_id": 18, "bbox": [258.15,41.29,348.26,243.78], "score": 0.236}, ...] """ cats = [cat['name'] for cat in _coco.loadCats(_coco.getCatIds())] class_to_coco_ind = dict(zip(cats, _coco.getCatIds())) results = [] for cls_ind, cls in enumerate(self.classes): if cls == '__background__': continue logger.info('collecting %s results (%d/%d)' % (cls, cls_ind, self.num_classes - 1)) coco_cat_id = class_to_coco_ind[cls] results.extend(self._coco_results_one_category(detections[cls_ind], coco_cat_id)) logger.info('writing results json to %s' % self._result_file) with open(self._result_file, 'w') as f: json.dump(results, f, sort_keys=True, indent=4)
[ "def", "_write_coco_results", "(", "self", ",", "_coco", ",", "detections", ")", ":", "cats", "=", "[", "cat", "[", "'name'", "]", "for", "cat", "in", "_coco", ".", "loadCats", "(", "_coco", ".", "getCatIds", "(", ")", ")", "]", "class_to_coco_ind", "=", "dict", "(", "zip", "(", "cats", ",", "_coco", ".", "getCatIds", "(", ")", ")", ")", "results", "=", "[", "]", "for", "cls_ind", ",", "cls", "in", "enumerate", "(", "self", ".", "classes", ")", ":", "if", "cls", "==", "'__background__'", ":", "continue", "logger", ".", "info", "(", "'collecting %s results (%d/%d)'", "%", "(", "cls", ",", "cls_ind", ",", "self", ".", "num_classes", "-", "1", ")", ")", "coco_cat_id", "=", "class_to_coco_ind", "[", "cls", "]", "results", ".", "extend", "(", "self", ".", "_coco_results_one_category", "(", "detections", "[", "cls_ind", "]", ",", "coco_cat_id", ")", ")", "logger", ".", "info", "(", "'writing results json to %s'", "%", "self", ".", "_result_file", ")", "with", "open", "(", "self", ".", "_result_file", ",", "'w'", ")", "as", "f", ":", "json", ".", "dump", "(", "results", ",", "f", ",", "sort_keys", "=", "True", ",", "indent", "=", "4", ")" ]
example results [{"image_id": 42, "category_id": 18, "bbox": [258.15,41.29,348.26,243.78], "score": 0.236}, ...]
[ "example", "results", "[", "{", "image_id", ":", "42", "category_id", ":", "18", "bbox", ":", "[", "258", ".", "15", "41", ".", "29", "348", ".", "26", "243", ".", "78", "]", "score", ":", "0", ".", "236", "}", "...", "]" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/rcnn/symimdb/coco.py#L132-L150
train
apache/incubator-mxnet
python/mxnet/ndarray/contrib.py
rand_zipfian
def rand_zipfian(true_classes, num_sampled, range_max, ctx=None): """Draw random samples from an approximately log-uniform or Zipfian distribution. This operation randomly samples *num_sampled* candidates the range of integers [0, range_max). The elements of sampled_candidates are drawn with replacement from the base distribution. The base distribution for this operator is an approximately log-uniform or Zipfian distribution: P(class) = (log(class + 2) - log(class + 1)) / log(range_max + 1) This sampler is useful when the true classes approximately follow such a distribution. For example, if the classes represent words in a lexicon sorted in decreasing order of \ frequency. If your classes are not ordered by decreasing frequency, do not use this op. Additionaly, it also returns the number of times each of the \ true classes and the sampled classes is expected to occur. Parameters ---------- true_classes : NDArray A 1-D NDArray of the target classes. num_sampled: int The number of classes to randomly sample. range_max: int The number of possible classes. ctx : Context Device context of output. Default is current context. Returns ------- samples: NDArray The sampled candidate classes in 1-D `int64` dtype. expected_count_true: NDArray The expected count for true classes in 1-D `float64` dtype. expected_count_sample: NDArray The expected count for sampled candidates in 1-D `float64` dtype. Examples -------- >>> true_cls = mx.nd.array([3]) >>> samples, exp_count_true, exp_count_sample = mx.nd.contrib.rand_zipfian(true_cls, 4, 5) >>> samples [1 3 3 3] <NDArray 4 @cpu(0)> >>> exp_count_true [ 0.12453879] <NDArray 1 @cpu(0)> >>> exp_count_sample [ 0.22629439 0.12453879 0.12453879 0.12453879] <NDArray 4 @cpu(0)> """ if ctx is None: ctx = current_context() log_range = math.log(range_max + 1) rand = uniform(0, log_range, shape=(num_sampled,), dtype='float64', ctx=ctx) # make sure sampled_classes are in the range of [0, range_max) sampled_classes = (rand.exp() - 1).astype('int64') % range_max true_cls = true_classes.as_in_context(ctx).astype('float64') expected_count_true = ((true_cls + 2.0) / (true_cls + 1.0)).log() / log_range * num_sampled # cast sampled classes to fp64 to avoid interget division sampled_cls_fp64 = sampled_classes.astype('float64') expected_prob_sampled = ((sampled_cls_fp64 + 2.0) / (sampled_cls_fp64 + 1.0)).log() / log_range expected_count_sampled = expected_prob_sampled * num_sampled return sampled_classes, expected_count_true, expected_count_sampled
python
def rand_zipfian(true_classes, num_sampled, range_max, ctx=None): """Draw random samples from an approximately log-uniform or Zipfian distribution. This operation randomly samples *num_sampled* candidates the range of integers [0, range_max). The elements of sampled_candidates are drawn with replacement from the base distribution. The base distribution for this operator is an approximately log-uniform or Zipfian distribution: P(class) = (log(class + 2) - log(class + 1)) / log(range_max + 1) This sampler is useful when the true classes approximately follow such a distribution. For example, if the classes represent words in a lexicon sorted in decreasing order of \ frequency. If your classes are not ordered by decreasing frequency, do not use this op. Additionaly, it also returns the number of times each of the \ true classes and the sampled classes is expected to occur. Parameters ---------- true_classes : NDArray A 1-D NDArray of the target classes. num_sampled: int The number of classes to randomly sample. range_max: int The number of possible classes. ctx : Context Device context of output. Default is current context. Returns ------- samples: NDArray The sampled candidate classes in 1-D `int64` dtype. expected_count_true: NDArray The expected count for true classes in 1-D `float64` dtype. expected_count_sample: NDArray The expected count for sampled candidates in 1-D `float64` dtype. Examples -------- >>> true_cls = mx.nd.array([3]) >>> samples, exp_count_true, exp_count_sample = mx.nd.contrib.rand_zipfian(true_cls, 4, 5) >>> samples [1 3 3 3] <NDArray 4 @cpu(0)> >>> exp_count_true [ 0.12453879] <NDArray 1 @cpu(0)> >>> exp_count_sample [ 0.22629439 0.12453879 0.12453879 0.12453879] <NDArray 4 @cpu(0)> """ if ctx is None: ctx = current_context() log_range = math.log(range_max + 1) rand = uniform(0, log_range, shape=(num_sampled,), dtype='float64', ctx=ctx) # make sure sampled_classes are in the range of [0, range_max) sampled_classes = (rand.exp() - 1).astype('int64') % range_max true_cls = true_classes.as_in_context(ctx).astype('float64') expected_count_true = ((true_cls + 2.0) / (true_cls + 1.0)).log() / log_range * num_sampled # cast sampled classes to fp64 to avoid interget division sampled_cls_fp64 = sampled_classes.astype('float64') expected_prob_sampled = ((sampled_cls_fp64 + 2.0) / (sampled_cls_fp64 + 1.0)).log() / log_range expected_count_sampled = expected_prob_sampled * num_sampled return sampled_classes, expected_count_true, expected_count_sampled
[ "def", "rand_zipfian", "(", "true_classes", ",", "num_sampled", ",", "range_max", ",", "ctx", "=", "None", ")", ":", "if", "ctx", "is", "None", ":", "ctx", "=", "current_context", "(", ")", "log_range", "=", "math", ".", "log", "(", "range_max", "+", "1", ")", "rand", "=", "uniform", "(", "0", ",", "log_range", ",", "shape", "=", "(", "num_sampled", ",", ")", ",", "dtype", "=", "'float64'", ",", "ctx", "=", "ctx", ")", "# make sure sampled_classes are in the range of [0, range_max)", "sampled_classes", "=", "(", "rand", ".", "exp", "(", ")", "-", "1", ")", ".", "astype", "(", "'int64'", ")", "%", "range_max", "true_cls", "=", "true_classes", ".", "as_in_context", "(", "ctx", ")", ".", "astype", "(", "'float64'", ")", "expected_count_true", "=", "(", "(", "true_cls", "+", "2.0", ")", "/", "(", "true_cls", "+", "1.0", ")", ")", ".", "log", "(", ")", "/", "log_range", "*", "num_sampled", "# cast sampled classes to fp64 to avoid interget division", "sampled_cls_fp64", "=", "sampled_classes", ".", "astype", "(", "'float64'", ")", "expected_prob_sampled", "=", "(", "(", "sampled_cls_fp64", "+", "2.0", ")", "/", "(", "sampled_cls_fp64", "+", "1.0", ")", ")", ".", "log", "(", ")", "/", "log_range", "expected_count_sampled", "=", "expected_prob_sampled", "*", "num_sampled", "return", "sampled_classes", ",", "expected_count_true", ",", "expected_count_sampled" ]
Draw random samples from an approximately log-uniform or Zipfian distribution. This operation randomly samples *num_sampled* candidates the range of integers [0, range_max). The elements of sampled_candidates are drawn with replacement from the base distribution. The base distribution for this operator is an approximately log-uniform or Zipfian distribution: P(class) = (log(class + 2) - log(class + 1)) / log(range_max + 1) This sampler is useful when the true classes approximately follow such a distribution. For example, if the classes represent words in a lexicon sorted in decreasing order of \ frequency. If your classes are not ordered by decreasing frequency, do not use this op. Additionaly, it also returns the number of times each of the \ true classes and the sampled classes is expected to occur. Parameters ---------- true_classes : NDArray A 1-D NDArray of the target classes. num_sampled: int The number of classes to randomly sample. range_max: int The number of possible classes. ctx : Context Device context of output. Default is current context. Returns ------- samples: NDArray The sampled candidate classes in 1-D `int64` dtype. expected_count_true: NDArray The expected count for true classes in 1-D `float64` dtype. expected_count_sample: NDArray The expected count for sampled candidates in 1-D `float64` dtype. Examples -------- >>> true_cls = mx.nd.array([3]) >>> samples, exp_count_true, exp_count_sample = mx.nd.contrib.rand_zipfian(true_cls, 4, 5) >>> samples [1 3 3 3] <NDArray 4 @cpu(0)> >>> exp_count_true [ 0.12453879] <NDArray 1 @cpu(0)> >>> exp_count_sample [ 0.22629439 0.12453879 0.12453879 0.12453879] <NDArray 4 @cpu(0)>
[ "Draw", "random", "samples", "from", "an", "approximately", "log", "-", "uniform", "or", "Zipfian", "distribution", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/ndarray/contrib.py#L36-L100
train
apache/incubator-mxnet
python/mxnet/ndarray/contrib.py
foreach
def foreach(body, data, init_states): """Run a for loop with user-defined computation over NDArrays on dimension 0. This operator simulates a for loop and body has the computation for an iteration of the for loop. It runs the computation in body on each slice from the input NDArrays. body takes two arguments as input and outputs a tuple of two elements, as illustrated below:: out, states = body(data1, states) data1 can be either an NDArray or a list of NDArrays. If data is an NDArray, data1 is an NDArray. Otherwise, data1 is a list of NDArrays and has the same size as data. states is a list of NDArrays and have the same size as init_states. Similarly, out can be either an NDArray or a list of NDArrays, which are concatenated as the first output of foreach; states from the last execution of body are the second output of foreach. The computation done by this operator is equivalent to the pseudo code below when the input data is NDArray:: states = init_states outs = [] for i in data.shape[0]: s = data[i] out, states = body(s, states) outs.append(out) outs = stack(*outs) Parameters ---------- body : a Python function. Define computation in an iteration. data: an NDArray or a list of NDArrays. The input data. init_states: an NDArray or nested lists of NDArrays. The initial values of the loop states. name: string. The name of the operator. Returns ------- outputs: an NDArray or nested lists of NDArrays. The output data concatenated from the output of all iterations. states: an NDArray or nested lists of NDArrays. The loop states in the last iteration. Examples -------- >>> step = lambda data, states: (data + states[0], [states[0] * 2]) >>> data = mx.nd.random.uniform(shape=(2, 10)) >>> states = [mx.nd.random.uniform(shape=(10))] >>> outs, states = mx.nd.contrib.foreach(step, data, states) """ def check_input(inputs, in_type, msg): is_NDArray_or_list = True if isinstance(inputs, list): for i in inputs: if not isinstance(i, in_type): is_NDArray_or_list = False break else: is_NDArray_or_list = isinstance(inputs, in_type) assert is_NDArray_or_list, msg flatten, _ = _flatten(data, "foreach input") check_input(flatten, ndarray.NDArray, "data should be an NDArray or a nested list of NDArrays") flatten, _ = _flatten(init_states, "foreach states") check_input(flatten, ndarray.NDArray, "init_states should be an NDArray or a nested list of NDArrays") not_data_list = isinstance(data, ndarray.NDArray) num_iters = data.shape[0] if not_data_list else data[0].shape[0] states = init_states outputs = [] for i in range(num_iters): if not_data_list: eles = data[i] else: eles = [d[i] for d in data] outs, states = body(eles, states) outs, out_fmt = _flatten(outs, "foreach output") outputs.append(outs) outputs = zip(*outputs) tmp_outputs = [] for out in outputs: tmp_outputs.append(ndarray.op.stack(*out)) outputs = tmp_outputs outputs, _ = _regroup(outputs, out_fmt) return (outputs, states)
python
def foreach(body, data, init_states): """Run a for loop with user-defined computation over NDArrays on dimension 0. This operator simulates a for loop and body has the computation for an iteration of the for loop. It runs the computation in body on each slice from the input NDArrays. body takes two arguments as input and outputs a tuple of two elements, as illustrated below:: out, states = body(data1, states) data1 can be either an NDArray or a list of NDArrays. If data is an NDArray, data1 is an NDArray. Otherwise, data1 is a list of NDArrays and has the same size as data. states is a list of NDArrays and have the same size as init_states. Similarly, out can be either an NDArray or a list of NDArrays, which are concatenated as the first output of foreach; states from the last execution of body are the second output of foreach. The computation done by this operator is equivalent to the pseudo code below when the input data is NDArray:: states = init_states outs = [] for i in data.shape[0]: s = data[i] out, states = body(s, states) outs.append(out) outs = stack(*outs) Parameters ---------- body : a Python function. Define computation in an iteration. data: an NDArray or a list of NDArrays. The input data. init_states: an NDArray or nested lists of NDArrays. The initial values of the loop states. name: string. The name of the operator. Returns ------- outputs: an NDArray or nested lists of NDArrays. The output data concatenated from the output of all iterations. states: an NDArray or nested lists of NDArrays. The loop states in the last iteration. Examples -------- >>> step = lambda data, states: (data + states[0], [states[0] * 2]) >>> data = mx.nd.random.uniform(shape=(2, 10)) >>> states = [mx.nd.random.uniform(shape=(10))] >>> outs, states = mx.nd.contrib.foreach(step, data, states) """ def check_input(inputs, in_type, msg): is_NDArray_or_list = True if isinstance(inputs, list): for i in inputs: if not isinstance(i, in_type): is_NDArray_or_list = False break else: is_NDArray_or_list = isinstance(inputs, in_type) assert is_NDArray_or_list, msg flatten, _ = _flatten(data, "foreach input") check_input(flatten, ndarray.NDArray, "data should be an NDArray or a nested list of NDArrays") flatten, _ = _flatten(init_states, "foreach states") check_input(flatten, ndarray.NDArray, "init_states should be an NDArray or a nested list of NDArrays") not_data_list = isinstance(data, ndarray.NDArray) num_iters = data.shape[0] if not_data_list else data[0].shape[0] states = init_states outputs = [] for i in range(num_iters): if not_data_list: eles = data[i] else: eles = [d[i] for d in data] outs, states = body(eles, states) outs, out_fmt = _flatten(outs, "foreach output") outputs.append(outs) outputs = zip(*outputs) tmp_outputs = [] for out in outputs: tmp_outputs.append(ndarray.op.stack(*out)) outputs = tmp_outputs outputs, _ = _regroup(outputs, out_fmt) return (outputs, states)
[ "def", "foreach", "(", "body", ",", "data", ",", "init_states", ")", ":", "def", "check_input", "(", "inputs", ",", "in_type", ",", "msg", ")", ":", "is_NDArray_or_list", "=", "True", "if", "isinstance", "(", "inputs", ",", "list", ")", ":", "for", "i", "in", "inputs", ":", "if", "not", "isinstance", "(", "i", ",", "in_type", ")", ":", "is_NDArray_or_list", "=", "False", "break", "else", ":", "is_NDArray_or_list", "=", "isinstance", "(", "inputs", ",", "in_type", ")", "assert", "is_NDArray_or_list", ",", "msg", "flatten", ",", "_", "=", "_flatten", "(", "data", ",", "\"foreach input\"", ")", "check_input", "(", "flatten", ",", "ndarray", ".", "NDArray", ",", "\"data should be an NDArray or a nested list of NDArrays\"", ")", "flatten", ",", "_", "=", "_flatten", "(", "init_states", ",", "\"foreach states\"", ")", "check_input", "(", "flatten", ",", "ndarray", ".", "NDArray", ",", "\"init_states should be an NDArray or a nested list of NDArrays\"", ")", "not_data_list", "=", "isinstance", "(", "data", ",", "ndarray", ".", "NDArray", ")", "num_iters", "=", "data", ".", "shape", "[", "0", "]", "if", "not_data_list", "else", "data", "[", "0", "]", ".", "shape", "[", "0", "]", "states", "=", "init_states", "outputs", "=", "[", "]", "for", "i", "in", "range", "(", "num_iters", ")", ":", "if", "not_data_list", ":", "eles", "=", "data", "[", "i", "]", "else", ":", "eles", "=", "[", "d", "[", "i", "]", "for", "d", "in", "data", "]", "outs", ",", "states", "=", "body", "(", "eles", ",", "states", ")", "outs", ",", "out_fmt", "=", "_flatten", "(", "outs", ",", "\"foreach output\"", ")", "outputs", ".", "append", "(", "outs", ")", "outputs", "=", "zip", "(", "*", "outputs", ")", "tmp_outputs", "=", "[", "]", "for", "out", "in", "outputs", ":", "tmp_outputs", ".", "append", "(", "ndarray", ".", "op", ".", "stack", "(", "*", "out", ")", ")", "outputs", "=", "tmp_outputs", "outputs", ",", "_", "=", "_regroup", "(", "outputs", ",", "out_fmt", ")", "return", "(", "outputs", ",", "states", ")" ]
Run a for loop with user-defined computation over NDArrays on dimension 0. This operator simulates a for loop and body has the computation for an iteration of the for loop. It runs the computation in body on each slice from the input NDArrays. body takes two arguments as input and outputs a tuple of two elements, as illustrated below:: out, states = body(data1, states) data1 can be either an NDArray or a list of NDArrays. If data is an NDArray, data1 is an NDArray. Otherwise, data1 is a list of NDArrays and has the same size as data. states is a list of NDArrays and have the same size as init_states. Similarly, out can be either an NDArray or a list of NDArrays, which are concatenated as the first output of foreach; states from the last execution of body are the second output of foreach. The computation done by this operator is equivalent to the pseudo code below when the input data is NDArray:: states = init_states outs = [] for i in data.shape[0]: s = data[i] out, states = body(s, states) outs.append(out) outs = stack(*outs) Parameters ---------- body : a Python function. Define computation in an iteration. data: an NDArray or a list of NDArrays. The input data. init_states: an NDArray or nested lists of NDArrays. The initial values of the loop states. name: string. The name of the operator. Returns ------- outputs: an NDArray or nested lists of NDArrays. The output data concatenated from the output of all iterations. states: an NDArray or nested lists of NDArrays. The loop states in the last iteration. Examples -------- >>> step = lambda data, states: (data + states[0], [states[0] * 2]) >>> data = mx.nd.random.uniform(shape=(2, 10)) >>> states = [mx.nd.random.uniform(shape=(10))] >>> outs, states = mx.nd.contrib.foreach(step, data, states)
[ "Run", "a", "for", "loop", "with", "user", "-", "defined", "computation", "over", "NDArrays", "on", "dimension", "0", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/ndarray/contrib.py#L136-L230
train
apache/incubator-mxnet
python/mxnet/ndarray/contrib.py
while_loop
def while_loop(cond, func, loop_vars, max_iterations=None): """Run a while loop with user-defined computation and loop condition. This operator simulates a while loop which iterately does customized computation as long as the condition is satisfied. `loop_vars` is a list of NDArrays on which the computation uses. `cond` is a user-defined function, used as the loop condition. It consumes `loop_vars`, and produces a scalar MXNet NDArray, indicating the termination of the loop. The loop ends when `cond` returns false (zero). The `cond` is variadic, and its signature should be `cond(*loop_vars) => NDArray`. `func` is a user-defined function, used as the loop body. It also consumes `loop_vars`, and produces `step_output` and `new_loop_vars` at each step. In each step, `step_output` should contain the same number elements. Through all steps, the i-th element of `step_output` should have the same shape and dtype. Also, `new_loop_vars` should contain the same number of elements as `loop_vars`, and the corresponding element should have the same shape and dtype. The `func` is variadic, and its signature should be `func(*loop_vars) => (NDArray or nested List[NDArray] step_output, NDArray or nested List[NDArray] new_loop_vars)`. `max_iterations` is a scalar that defines the maximum number of iterations allowed. This function returns two lists. The first list has the length of `|step_output|`, in which the i-th element are all i-th elements of `step_output` from all steps, stacked along axis 0. The second list has the length of `|loop_vars|`, which represents final states of loop variables. .. warning:: For now, the axis 0 of all NDArrays in the first list are `max_iterations`, due to lack of dynamic shape inference. .. warning:: When `cond` is never satisfied, we assume `step_output` is empty, because it cannot be inferred. This is different from the symbolic version. Parameters ---------- cond: a Python function. The loop condition. func: a Python function. The loop body. loop_vars: an NDArray or nested lists of NDArrays. The initial values of the loop variables. max_iterations: a python int. Maximum number of iterations. Returns ------ outputs: an NDArray or nested lists of NDArrays stacked output from each step states: an NDArray or nested lists of NDArrays final state Examples -------- >>> cond = lambda i, s: i <= 5 >>> func = lambda i, s: ([i + s], [i + 1, s + i]) >>> loop_vars = (mx.nd.array([0], dtype="int64"), mx.nd.array([1], dtype="int64")) >>> outputs, states = mx.nd.contrib.while_loop(cond, func, loop_vars, max_iterations=10) >>> outputs [ [[ 1] [ 2] [ 4] [ 7] [11] [16] [...] # undefined value [...] [...] [...]] <NDArray 6x1 @cpu(0)>] >>> states [ [6] <NDArray 1 @cpu(0)>, [16] <NDArray 1 @cpu(0)>] """ def _to_python_scalar(inputs, type_, name): """Converts "inputs", possibly typed mxnet NDArray, a numpy ndarray, other python types, to the given type """ if isinstance(inputs, ndarray.NDArray): inputs = inputs.asscalar() try: inputs = type_(inputs) except: raise ValueError("Cannot convert %s to python %s" % (name, type_.__name__)) return inputs def _func_wrapper(loop_vars): """This wrapper unifies "func: loop_vars -> new_loop_vars" and "func: loop_vars -> (step_output, new_loop_vars)" into "func: loop_vars -> (None or tuple of step_outputs, tuple of new_loop_vars) """ step_output, new_loop_vars = func(*loop_vars) if step_output is None: step_output = [] if new_loop_vars is None: new_loop_vars = [] if isinstance(step_output, tuple): step_output = list(step_output) if isinstance(new_loop_vars, tuple): new_loop_vars = list(new_loop_vars) new_loop_vars = _as_list(new_loop_vars) if len(loop_vars) != len(new_loop_vars): raise ValueError("The length of loop_vars should be consistent during the loop") return step_output, new_loop_vars if max_iterations is None: raise ValueError("max_iterations should be specified") max_iterations = _to_python_scalar(max_iterations, int, "max_iteration") # It should be work as fine if loop_vars are empty I guess, # but it is semantically unnecessary to include this case. if len(loop_vars) == 0: raise ValueError("loop_vars should contain at least one element") steps = 0 outputs = [] # there might not be an iteration. out_fmt = None not_loop_var_list = isinstance(loop_vars, ndarray.NDArray) loop_vars = _as_list(loop_vars) while steps < max_iterations and \ _to_python_scalar(cond(*loop_vars), bool, "Return value of cond"): # loop condition step_output, loop_vars = _func_wrapper(loop_vars) step_output, out_fmt = _flatten(step_output, "while output") outputs.append(step_output) steps += 1 if len(outputs) != steps or len(step_output) != len(outputs[0]): raise ValueError("Number of elements in step_output should be the same in each step") stacked_outputs = [] for i_th, items in enumerate(zip(*outputs), 1): # `mx.ndarray.pad` only support 4-D or 5-D inputs for now # so we could not use it. items = [x.expand_dims(0) for x in items] if steps != max_iterations and items: pad_shape = [max_iterations - steps] + list(items[0].shape[1: ]) pad = ndarray.empty( shape=pad_shape, ctx=items[0].context, dtype=items[0].dtype, ) items = list(items) + [pad] try: stacked_outputs.append(ndarray.op.concat(*items, dim=0)) except ValueError: raise ValueError("\n".join( ["Shapes of %d-th elements in step_outputs are inconsistent, which are:" % i_th] + [" Step %d, shape is %s" % (i, str(x.shape)) for i, x in enumerate(items)] )) if out_fmt is not None: stacked_outputs, _ = _regroup(stacked_outputs, out_fmt) if not_loop_var_list: loop_vars = loop_vars[0] return stacked_outputs, loop_vars
python
def while_loop(cond, func, loop_vars, max_iterations=None): """Run a while loop with user-defined computation and loop condition. This operator simulates a while loop which iterately does customized computation as long as the condition is satisfied. `loop_vars` is a list of NDArrays on which the computation uses. `cond` is a user-defined function, used as the loop condition. It consumes `loop_vars`, and produces a scalar MXNet NDArray, indicating the termination of the loop. The loop ends when `cond` returns false (zero). The `cond` is variadic, and its signature should be `cond(*loop_vars) => NDArray`. `func` is a user-defined function, used as the loop body. It also consumes `loop_vars`, and produces `step_output` and `new_loop_vars` at each step. In each step, `step_output` should contain the same number elements. Through all steps, the i-th element of `step_output` should have the same shape and dtype. Also, `new_loop_vars` should contain the same number of elements as `loop_vars`, and the corresponding element should have the same shape and dtype. The `func` is variadic, and its signature should be `func(*loop_vars) => (NDArray or nested List[NDArray] step_output, NDArray or nested List[NDArray] new_loop_vars)`. `max_iterations` is a scalar that defines the maximum number of iterations allowed. This function returns two lists. The first list has the length of `|step_output|`, in which the i-th element are all i-th elements of `step_output` from all steps, stacked along axis 0. The second list has the length of `|loop_vars|`, which represents final states of loop variables. .. warning:: For now, the axis 0 of all NDArrays in the first list are `max_iterations`, due to lack of dynamic shape inference. .. warning:: When `cond` is never satisfied, we assume `step_output` is empty, because it cannot be inferred. This is different from the symbolic version. Parameters ---------- cond: a Python function. The loop condition. func: a Python function. The loop body. loop_vars: an NDArray or nested lists of NDArrays. The initial values of the loop variables. max_iterations: a python int. Maximum number of iterations. Returns ------ outputs: an NDArray or nested lists of NDArrays stacked output from each step states: an NDArray or nested lists of NDArrays final state Examples -------- >>> cond = lambda i, s: i <= 5 >>> func = lambda i, s: ([i + s], [i + 1, s + i]) >>> loop_vars = (mx.nd.array([0], dtype="int64"), mx.nd.array([1], dtype="int64")) >>> outputs, states = mx.nd.contrib.while_loop(cond, func, loop_vars, max_iterations=10) >>> outputs [ [[ 1] [ 2] [ 4] [ 7] [11] [16] [...] # undefined value [...] [...] [...]] <NDArray 6x1 @cpu(0)>] >>> states [ [6] <NDArray 1 @cpu(0)>, [16] <NDArray 1 @cpu(0)>] """ def _to_python_scalar(inputs, type_, name): """Converts "inputs", possibly typed mxnet NDArray, a numpy ndarray, other python types, to the given type """ if isinstance(inputs, ndarray.NDArray): inputs = inputs.asscalar() try: inputs = type_(inputs) except: raise ValueError("Cannot convert %s to python %s" % (name, type_.__name__)) return inputs def _func_wrapper(loop_vars): """This wrapper unifies "func: loop_vars -> new_loop_vars" and "func: loop_vars -> (step_output, new_loop_vars)" into "func: loop_vars -> (None or tuple of step_outputs, tuple of new_loop_vars) """ step_output, new_loop_vars = func(*loop_vars) if step_output is None: step_output = [] if new_loop_vars is None: new_loop_vars = [] if isinstance(step_output, tuple): step_output = list(step_output) if isinstance(new_loop_vars, tuple): new_loop_vars = list(new_loop_vars) new_loop_vars = _as_list(new_loop_vars) if len(loop_vars) != len(new_loop_vars): raise ValueError("The length of loop_vars should be consistent during the loop") return step_output, new_loop_vars if max_iterations is None: raise ValueError("max_iterations should be specified") max_iterations = _to_python_scalar(max_iterations, int, "max_iteration") # It should be work as fine if loop_vars are empty I guess, # but it is semantically unnecessary to include this case. if len(loop_vars) == 0: raise ValueError("loop_vars should contain at least one element") steps = 0 outputs = [] # there might not be an iteration. out_fmt = None not_loop_var_list = isinstance(loop_vars, ndarray.NDArray) loop_vars = _as_list(loop_vars) while steps < max_iterations and \ _to_python_scalar(cond(*loop_vars), bool, "Return value of cond"): # loop condition step_output, loop_vars = _func_wrapper(loop_vars) step_output, out_fmt = _flatten(step_output, "while output") outputs.append(step_output) steps += 1 if len(outputs) != steps or len(step_output) != len(outputs[0]): raise ValueError("Number of elements in step_output should be the same in each step") stacked_outputs = [] for i_th, items in enumerate(zip(*outputs), 1): # `mx.ndarray.pad` only support 4-D or 5-D inputs for now # so we could not use it. items = [x.expand_dims(0) for x in items] if steps != max_iterations and items: pad_shape = [max_iterations - steps] + list(items[0].shape[1: ]) pad = ndarray.empty( shape=pad_shape, ctx=items[0].context, dtype=items[0].dtype, ) items = list(items) + [pad] try: stacked_outputs.append(ndarray.op.concat(*items, dim=0)) except ValueError: raise ValueError("\n".join( ["Shapes of %d-th elements in step_outputs are inconsistent, which are:" % i_th] + [" Step %d, shape is %s" % (i, str(x.shape)) for i, x in enumerate(items)] )) if out_fmt is not None: stacked_outputs, _ = _regroup(stacked_outputs, out_fmt) if not_loop_var_list: loop_vars = loop_vars[0] return stacked_outputs, loop_vars
[ "def", "while_loop", "(", "cond", ",", "func", ",", "loop_vars", ",", "max_iterations", "=", "None", ")", ":", "def", "_to_python_scalar", "(", "inputs", ",", "type_", ",", "name", ")", ":", "\"\"\"Converts \"inputs\", possibly typed mxnet NDArray, a numpy ndarray, other python types,\n to the given type\n \"\"\"", "if", "isinstance", "(", "inputs", ",", "ndarray", ".", "NDArray", ")", ":", "inputs", "=", "inputs", ".", "asscalar", "(", ")", "try", ":", "inputs", "=", "type_", "(", "inputs", ")", "except", ":", "raise", "ValueError", "(", "\"Cannot convert %s to python %s\"", "%", "(", "name", ",", "type_", ".", "__name__", ")", ")", "return", "inputs", "def", "_func_wrapper", "(", "loop_vars", ")", ":", "\"\"\"This wrapper unifies\n \"func: loop_vars -> new_loop_vars\"\n and \"func: loop_vars -> (step_output, new_loop_vars)\"\n into \"func: loop_vars -> (None or tuple of step_outputs, tuple of new_loop_vars)\n \"\"\"", "step_output", ",", "new_loop_vars", "=", "func", "(", "*", "loop_vars", ")", "if", "step_output", "is", "None", ":", "step_output", "=", "[", "]", "if", "new_loop_vars", "is", "None", ":", "new_loop_vars", "=", "[", "]", "if", "isinstance", "(", "step_output", ",", "tuple", ")", ":", "step_output", "=", "list", "(", "step_output", ")", "if", "isinstance", "(", "new_loop_vars", ",", "tuple", ")", ":", "new_loop_vars", "=", "list", "(", "new_loop_vars", ")", "new_loop_vars", "=", "_as_list", "(", "new_loop_vars", ")", "if", "len", "(", "loop_vars", ")", "!=", "len", "(", "new_loop_vars", ")", ":", "raise", "ValueError", "(", "\"The length of loop_vars should be consistent during the loop\"", ")", "return", "step_output", ",", "new_loop_vars", "if", "max_iterations", "is", "None", ":", "raise", "ValueError", "(", "\"max_iterations should be specified\"", ")", "max_iterations", "=", "_to_python_scalar", "(", "max_iterations", ",", "int", ",", "\"max_iteration\"", ")", "# It should be work as fine if loop_vars are empty I guess,", "# but it is semantically unnecessary to include this case.", "if", "len", "(", "loop_vars", ")", "==", "0", ":", "raise", "ValueError", "(", "\"loop_vars should contain at least one element\"", ")", "steps", "=", "0", "outputs", "=", "[", "]", "# there might not be an iteration.", "out_fmt", "=", "None", "not_loop_var_list", "=", "isinstance", "(", "loop_vars", ",", "ndarray", ".", "NDArray", ")", "loop_vars", "=", "_as_list", "(", "loop_vars", ")", "while", "steps", "<", "max_iterations", "and", "_to_python_scalar", "(", "cond", "(", "*", "loop_vars", ")", ",", "bool", ",", "\"Return value of cond\"", ")", ":", "# loop condition", "step_output", ",", "loop_vars", "=", "_func_wrapper", "(", "loop_vars", ")", "step_output", ",", "out_fmt", "=", "_flatten", "(", "step_output", ",", "\"while output\"", ")", "outputs", ".", "append", "(", "step_output", ")", "steps", "+=", "1", "if", "len", "(", "outputs", ")", "!=", "steps", "or", "len", "(", "step_output", ")", "!=", "len", "(", "outputs", "[", "0", "]", ")", ":", "raise", "ValueError", "(", "\"Number of elements in step_output should be the same in each step\"", ")", "stacked_outputs", "=", "[", "]", "for", "i_th", ",", "items", "in", "enumerate", "(", "zip", "(", "*", "outputs", ")", ",", "1", ")", ":", "# `mx.ndarray.pad` only support 4-D or 5-D inputs for now", "# so we could not use it.", "items", "=", "[", "x", ".", "expand_dims", "(", "0", ")", "for", "x", "in", "items", "]", "if", "steps", "!=", "max_iterations", "and", "items", ":", "pad_shape", "=", "[", "max_iterations", "-", "steps", "]", "+", "list", "(", "items", "[", "0", "]", ".", "shape", "[", "1", ":", "]", ")", "pad", "=", "ndarray", ".", "empty", "(", "shape", "=", "pad_shape", ",", "ctx", "=", "items", "[", "0", "]", ".", "context", ",", "dtype", "=", "items", "[", "0", "]", ".", "dtype", ",", ")", "items", "=", "list", "(", "items", ")", "+", "[", "pad", "]", "try", ":", "stacked_outputs", ".", "append", "(", "ndarray", ".", "op", ".", "concat", "(", "*", "items", ",", "dim", "=", "0", ")", ")", "except", "ValueError", ":", "raise", "ValueError", "(", "\"\\n\"", ".", "join", "(", "[", "\"Shapes of %d-th elements in step_outputs are inconsistent, which are:\"", "%", "i_th", "]", "+", "[", "\" Step %d, shape is %s\"", "%", "(", "i", ",", "str", "(", "x", ".", "shape", ")", ")", "for", "i", ",", "x", "in", "enumerate", "(", "items", ")", "]", ")", ")", "if", "out_fmt", "is", "not", "None", ":", "stacked_outputs", ",", "_", "=", "_regroup", "(", "stacked_outputs", ",", "out_fmt", ")", "if", "not_loop_var_list", ":", "loop_vars", "=", "loop_vars", "[", "0", "]", "return", "stacked_outputs", ",", "loop_vars" ]
Run a while loop with user-defined computation and loop condition. This operator simulates a while loop which iterately does customized computation as long as the condition is satisfied. `loop_vars` is a list of NDArrays on which the computation uses. `cond` is a user-defined function, used as the loop condition. It consumes `loop_vars`, and produces a scalar MXNet NDArray, indicating the termination of the loop. The loop ends when `cond` returns false (zero). The `cond` is variadic, and its signature should be `cond(*loop_vars) => NDArray`. `func` is a user-defined function, used as the loop body. It also consumes `loop_vars`, and produces `step_output` and `new_loop_vars` at each step. In each step, `step_output` should contain the same number elements. Through all steps, the i-th element of `step_output` should have the same shape and dtype. Also, `new_loop_vars` should contain the same number of elements as `loop_vars`, and the corresponding element should have the same shape and dtype. The `func` is variadic, and its signature should be `func(*loop_vars) => (NDArray or nested List[NDArray] step_output, NDArray or nested List[NDArray] new_loop_vars)`. `max_iterations` is a scalar that defines the maximum number of iterations allowed. This function returns two lists. The first list has the length of `|step_output|`, in which the i-th element are all i-th elements of `step_output` from all steps, stacked along axis 0. The second list has the length of `|loop_vars|`, which represents final states of loop variables. .. warning:: For now, the axis 0 of all NDArrays in the first list are `max_iterations`, due to lack of dynamic shape inference. .. warning:: When `cond` is never satisfied, we assume `step_output` is empty, because it cannot be inferred. This is different from the symbolic version. Parameters ---------- cond: a Python function. The loop condition. func: a Python function. The loop body. loop_vars: an NDArray or nested lists of NDArrays. The initial values of the loop variables. max_iterations: a python int. Maximum number of iterations. Returns ------ outputs: an NDArray or nested lists of NDArrays stacked output from each step states: an NDArray or nested lists of NDArrays final state Examples -------- >>> cond = lambda i, s: i <= 5 >>> func = lambda i, s: ([i + s], [i + 1, s + i]) >>> loop_vars = (mx.nd.array([0], dtype="int64"), mx.nd.array([1], dtype="int64")) >>> outputs, states = mx.nd.contrib.while_loop(cond, func, loop_vars, max_iterations=10) >>> outputs [ [[ 1] [ 2] [ 4] [ 7] [11] [16] [...] # undefined value [...] [...] [...]] <NDArray 6x1 @cpu(0)>] >>> states [ [6] <NDArray 1 @cpu(0)>, [16] <NDArray 1 @cpu(0)>]
[ "Run", "a", "while", "loop", "with", "user", "-", "defined", "computation", "and", "loop", "condition", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/ndarray/contrib.py#L232-L398
train
apache/incubator-mxnet
python/mxnet/ndarray/contrib.py
cond
def cond(pred, then_func, else_func): """Run an if-then-else using user-defined condition and computation This operator simulates a if-like branch which chooses to do one of the two customized computations according to the specified condition. `pred` is a scalar MXNet NDArray, indicating which branch of computation should be used. `then_func` is a user-defined function, used as computation of the then branch. It produces `outputs`, which is a list of NDArrays. The signature of `then_func` should be `then_func() => NDArray or nested List[NDArray]`. `else_func` is a user-defined function, used as computation of the else branch. It produces `outputs`, which is a list of NDArrays. The signature of `else_func` should be `else_func() => NDArray or nested List[NDArray]`. The `outputs` produces by `then_func` and `else_func` should have the same number of elements, all of which should be in the same shape, of the same dtype and stype. This function returns a list of symbols, representing the computation result. Parameters ---------- pred: a MXNet NDArray representing a scalar. The branch condition. then_func: a Python function. The computation to be executed if `pred` is true. else_func: a Python function. The computation to be executed if `pred` is false. Returns ------- outputs: an NDArray or nested lists of NDArrays, representing the result of computation. Examples -------- >>> a, b = mx.nd.array([1]), mx.nd.array([2]) >>> pred = a * b < 5 >>> then_func = lambda: (a + 5) * (b + 5) >>> else_func = lambda: (a - 5) * (b - 5) >>> outputs = mx.nd.contrib.cond(pred, then_func, else_func) >>> outputs[0] [42.] <NDArray 1 @cpu(0)> """ def _to_python_scalar(inputs, type_, name): """Converts "inputs", possibly typed mxnet NDArray, a numpy ndarray, other python types, to the given type """ if hasattr(inputs, "asscalar"): inputs = inputs.asscalar() try: inputs = type_(inputs) except: raise ValueError("Cannot convert %s to python %s" % (name, type_.__name__)) return inputs branch = _to_python_scalar(pred, bool, "pred") if branch: return then_func() else: return else_func()
python
def cond(pred, then_func, else_func): """Run an if-then-else using user-defined condition and computation This operator simulates a if-like branch which chooses to do one of the two customized computations according to the specified condition. `pred` is a scalar MXNet NDArray, indicating which branch of computation should be used. `then_func` is a user-defined function, used as computation of the then branch. It produces `outputs`, which is a list of NDArrays. The signature of `then_func` should be `then_func() => NDArray or nested List[NDArray]`. `else_func` is a user-defined function, used as computation of the else branch. It produces `outputs`, which is a list of NDArrays. The signature of `else_func` should be `else_func() => NDArray or nested List[NDArray]`. The `outputs` produces by `then_func` and `else_func` should have the same number of elements, all of which should be in the same shape, of the same dtype and stype. This function returns a list of symbols, representing the computation result. Parameters ---------- pred: a MXNet NDArray representing a scalar. The branch condition. then_func: a Python function. The computation to be executed if `pred` is true. else_func: a Python function. The computation to be executed if `pred` is false. Returns ------- outputs: an NDArray or nested lists of NDArrays, representing the result of computation. Examples -------- >>> a, b = mx.nd.array([1]), mx.nd.array([2]) >>> pred = a * b < 5 >>> then_func = lambda: (a + 5) * (b + 5) >>> else_func = lambda: (a - 5) * (b - 5) >>> outputs = mx.nd.contrib.cond(pred, then_func, else_func) >>> outputs[0] [42.] <NDArray 1 @cpu(0)> """ def _to_python_scalar(inputs, type_, name): """Converts "inputs", possibly typed mxnet NDArray, a numpy ndarray, other python types, to the given type """ if hasattr(inputs, "asscalar"): inputs = inputs.asscalar() try: inputs = type_(inputs) except: raise ValueError("Cannot convert %s to python %s" % (name, type_.__name__)) return inputs branch = _to_python_scalar(pred, bool, "pred") if branch: return then_func() else: return else_func()
[ "def", "cond", "(", "pred", ",", "then_func", ",", "else_func", ")", ":", "def", "_to_python_scalar", "(", "inputs", ",", "type_", ",", "name", ")", ":", "\"\"\"Converts \"inputs\", possibly typed mxnet NDArray, a numpy ndarray, other python types,\n to the given type\n \"\"\"", "if", "hasattr", "(", "inputs", ",", "\"asscalar\"", ")", ":", "inputs", "=", "inputs", ".", "asscalar", "(", ")", "try", ":", "inputs", "=", "type_", "(", "inputs", ")", "except", ":", "raise", "ValueError", "(", "\"Cannot convert %s to python %s\"", "%", "(", "name", ",", "type_", ".", "__name__", ")", ")", "return", "inputs", "branch", "=", "_to_python_scalar", "(", "pred", ",", "bool", ",", "\"pred\"", ")", "if", "branch", ":", "return", "then_func", "(", ")", "else", ":", "return", "else_func", "(", ")" ]
Run an if-then-else using user-defined condition and computation This operator simulates a if-like branch which chooses to do one of the two customized computations according to the specified condition. `pred` is a scalar MXNet NDArray, indicating which branch of computation should be used. `then_func` is a user-defined function, used as computation of the then branch. It produces `outputs`, which is a list of NDArrays. The signature of `then_func` should be `then_func() => NDArray or nested List[NDArray]`. `else_func` is a user-defined function, used as computation of the else branch. It produces `outputs`, which is a list of NDArrays. The signature of `else_func` should be `else_func() => NDArray or nested List[NDArray]`. The `outputs` produces by `then_func` and `else_func` should have the same number of elements, all of which should be in the same shape, of the same dtype and stype. This function returns a list of symbols, representing the computation result. Parameters ---------- pred: a MXNet NDArray representing a scalar. The branch condition. then_func: a Python function. The computation to be executed if `pred` is true. else_func: a Python function. The computation to be executed if `pred` is false. Returns ------- outputs: an NDArray or nested lists of NDArrays, representing the result of computation. Examples -------- >>> a, b = mx.nd.array([1]), mx.nd.array([2]) >>> pred = a * b < 5 >>> then_func = lambda: (a + 5) * (b + 5) >>> else_func = lambda: (a - 5) * (b - 5) >>> outputs = mx.nd.contrib.cond(pred, then_func, else_func) >>> outputs[0] [42.] <NDArray 1 @cpu(0)>
[ "Run", "an", "if", "-", "then", "-", "else", "using", "user", "-", "defined", "condition", "and", "computation" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/ndarray/contrib.py#L400-L464
train
apache/incubator-mxnet
python/mxnet/ndarray/contrib.py
isfinite
def isfinite(data): """Performs an element-wise check to determine if the NDArray contains an infinite element or not. Parameters ---------- input : NDArray An N-D NDArray. Returns ------- output: NDArray The output NDarray, with same shape as input, where 1 indicates the array element is finite i.e. not equal to positive or negative infinity and 0 in places where it is positive or negative infinity. Examples -------- >>> data = mx.nd.array([np.inf, -np.inf, np.NINF, -1]) >>> output = mx.nd.contrib.isfinite(data) >>> output [0. 0. 0. 1.] <NDArray 4 @cpu(0)> """ is_data_not_nan = data == data is_data_not_infinite = data.abs() != np.inf return ndarray.logical_and(is_data_not_infinite, is_data_not_nan)
python
def isfinite(data): """Performs an element-wise check to determine if the NDArray contains an infinite element or not. Parameters ---------- input : NDArray An N-D NDArray. Returns ------- output: NDArray The output NDarray, with same shape as input, where 1 indicates the array element is finite i.e. not equal to positive or negative infinity and 0 in places where it is positive or negative infinity. Examples -------- >>> data = mx.nd.array([np.inf, -np.inf, np.NINF, -1]) >>> output = mx.nd.contrib.isfinite(data) >>> output [0. 0. 0. 1.] <NDArray 4 @cpu(0)> """ is_data_not_nan = data == data is_data_not_infinite = data.abs() != np.inf return ndarray.logical_and(is_data_not_infinite, is_data_not_nan)
[ "def", "isfinite", "(", "data", ")", ":", "is_data_not_nan", "=", "data", "==", "data", "is_data_not_infinite", "=", "data", ".", "abs", "(", ")", "!=", "np", ".", "inf", "return", "ndarray", ".", "logical_and", "(", "is_data_not_infinite", ",", "is_data_not_nan", ")" ]
Performs an element-wise check to determine if the NDArray contains an infinite element or not. Parameters ---------- input : NDArray An N-D NDArray. Returns ------- output: NDArray The output NDarray, with same shape as input, where 1 indicates the array element is finite i.e. not equal to positive or negative infinity and 0 in places where it is positive or negative infinity. Examples -------- >>> data = mx.nd.array([np.inf, -np.inf, np.NINF, -1]) >>> output = mx.nd.contrib.isfinite(data) >>> output [0. 0. 0. 1.] <NDArray 4 @cpu(0)>
[ "Performs", "an", "element", "-", "wise", "check", "to", "determine", "if", "the", "NDArray", "contains", "an", "infinite", "element", "or", "not", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/ndarray/contrib.py#L492-L519
train
apache/incubator-mxnet
example/speech_recognition/stt_layer_lstm.py
vanilla_lstm
def vanilla_lstm(num_hidden, indata, prev_state, param, seqidx, layeridx, is_batchnorm=False, gamma=None, beta=None, name=None): """LSTM Cell symbol""" i2h = mx.sym.FullyConnected(data=indata, weight=param.i2h_weight, bias=param.i2h_bias, num_hidden=num_hidden * 4, name="t%d_l%d_i2h" % (seqidx, layeridx)) if is_batchnorm: if name is not None: i2h = batchnorm(net=i2h, gamma=gamma, beta=beta, name="%s_batchnorm" % name) else: i2h = batchnorm(net=i2h, gamma=gamma, beta=beta) h2h = mx.sym.FullyConnected(data=prev_state.h, weight=param.h2h_weight, bias=param.h2h_bias, num_hidden=num_hidden * 4, name="t%d_l%d_h2h" % (seqidx, layeridx)) gates = i2h + h2h slice_gates = mx.sym.SliceChannel(gates, num_outputs=4, name="t%d_l%d_slice" % (seqidx, layeridx)) in_gate = mx.sym.Activation(slice_gates[0], act_type="sigmoid") in_transform = mx.sym.Activation(slice_gates[1], act_type="tanh") forget_gate = mx.sym.Activation(slice_gates[2], act_type="sigmoid") out_gate = mx.sym.Activation(slice_gates[3], act_type="sigmoid") next_c = (forget_gate * prev_state.c) + (in_gate * in_transform) next_h = out_gate * mx.sym.Activation(next_c, act_type="tanh") return LSTMState(c=next_c, h=next_h)
python
def vanilla_lstm(num_hidden, indata, prev_state, param, seqidx, layeridx, is_batchnorm=False, gamma=None, beta=None, name=None): """LSTM Cell symbol""" i2h = mx.sym.FullyConnected(data=indata, weight=param.i2h_weight, bias=param.i2h_bias, num_hidden=num_hidden * 4, name="t%d_l%d_i2h" % (seqidx, layeridx)) if is_batchnorm: if name is not None: i2h = batchnorm(net=i2h, gamma=gamma, beta=beta, name="%s_batchnorm" % name) else: i2h = batchnorm(net=i2h, gamma=gamma, beta=beta) h2h = mx.sym.FullyConnected(data=prev_state.h, weight=param.h2h_weight, bias=param.h2h_bias, num_hidden=num_hidden * 4, name="t%d_l%d_h2h" % (seqidx, layeridx)) gates = i2h + h2h slice_gates = mx.sym.SliceChannel(gates, num_outputs=4, name="t%d_l%d_slice" % (seqidx, layeridx)) in_gate = mx.sym.Activation(slice_gates[0], act_type="sigmoid") in_transform = mx.sym.Activation(slice_gates[1], act_type="tanh") forget_gate = mx.sym.Activation(slice_gates[2], act_type="sigmoid") out_gate = mx.sym.Activation(slice_gates[3], act_type="sigmoid") next_c = (forget_gate * prev_state.c) + (in_gate * in_transform) next_h = out_gate * mx.sym.Activation(next_c, act_type="tanh") return LSTMState(c=next_c, h=next_h)
[ "def", "vanilla_lstm", "(", "num_hidden", ",", "indata", ",", "prev_state", ",", "param", ",", "seqidx", ",", "layeridx", ",", "is_batchnorm", "=", "False", ",", "gamma", "=", "None", ",", "beta", "=", "None", ",", "name", "=", "None", ")", ":", "i2h", "=", "mx", ".", "sym", ".", "FullyConnected", "(", "data", "=", "indata", ",", "weight", "=", "param", ".", "i2h_weight", ",", "bias", "=", "param", ".", "i2h_bias", ",", "num_hidden", "=", "num_hidden", "*", "4", ",", "name", "=", "\"t%d_l%d_i2h\"", "%", "(", "seqidx", ",", "layeridx", ")", ")", "if", "is_batchnorm", ":", "if", "name", "is", "not", "None", ":", "i2h", "=", "batchnorm", "(", "net", "=", "i2h", ",", "gamma", "=", "gamma", ",", "beta", "=", "beta", ",", "name", "=", "\"%s_batchnorm\"", "%", "name", ")", "else", ":", "i2h", "=", "batchnorm", "(", "net", "=", "i2h", ",", "gamma", "=", "gamma", ",", "beta", "=", "beta", ")", "h2h", "=", "mx", ".", "sym", ".", "FullyConnected", "(", "data", "=", "prev_state", ".", "h", ",", "weight", "=", "param", ".", "h2h_weight", ",", "bias", "=", "param", ".", "h2h_bias", ",", "num_hidden", "=", "num_hidden", "*", "4", ",", "name", "=", "\"t%d_l%d_h2h\"", "%", "(", "seqidx", ",", "layeridx", ")", ")", "gates", "=", "i2h", "+", "h2h", "slice_gates", "=", "mx", ".", "sym", ".", "SliceChannel", "(", "gates", ",", "num_outputs", "=", "4", ",", "name", "=", "\"t%d_l%d_slice\"", "%", "(", "seqidx", ",", "layeridx", ")", ")", "in_gate", "=", "mx", ".", "sym", ".", "Activation", "(", "slice_gates", "[", "0", "]", ",", "act_type", "=", "\"sigmoid\"", ")", "in_transform", "=", "mx", ".", "sym", ".", "Activation", "(", "slice_gates", "[", "1", "]", ",", "act_type", "=", "\"tanh\"", ")", "forget_gate", "=", "mx", ".", "sym", ".", "Activation", "(", "slice_gates", "[", "2", "]", ",", "act_type", "=", "\"sigmoid\"", ")", "out_gate", "=", "mx", ".", "sym", ".", "Activation", "(", "slice_gates", "[", "3", "]", ",", "act_type", "=", "\"sigmoid\"", ")", "next_c", "=", "(", "forget_gate", "*", "prev_state", ".", "c", ")", "+", "(", "in_gate", "*", "in_transform", ")", "next_h", "=", "out_gate", "*", "mx", ".", "sym", ".", "Activation", "(", "next_c", ",", "act_type", "=", "\"tanh\"", ")", "return", "LSTMState", "(", "c", "=", "next_c", ",", "h", "=", "next_h", ")" ]
LSTM Cell symbol
[ "LSTM", "Cell", "symbol" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/speech_recognition/stt_layer_lstm.py#L36-L62
train
apache/incubator-mxnet
example/speech_recognition/stt_layer_lstm.py
lstm
def lstm(num_hidden, indata, prev_state, param, seqidx, layeridx, dropout=0., num_hidden_proj=0, is_batchnorm=False, gamma=None, beta=None, name=None): """LSTM Cell symbol""" # dropout input if dropout > 0.: indata = mx.sym.Dropout(data=indata, p=dropout) i2h = mx.sym.FullyConnected(data=indata, weight=param.i2h_weight, bias=param.i2h_bias, num_hidden=num_hidden * 4, name="t%d_l%d_i2h" % (seqidx, layeridx)) if is_batchnorm: if name is not None: i2h = batchnorm(net=i2h, gamma=gamma, beta=beta, name="%s_batchnorm" % name) else: i2h = batchnorm(net=i2h, gamma=gamma, beta=beta) h2h = mx.sym.FullyConnected(data=prev_state.h, weight=param.h2h_weight, # bias=param.h2h_bias, no_bias=True, num_hidden=num_hidden * 4, name="t%d_l%d_h2h" % (seqidx, layeridx)) gates = i2h + h2h slice_gates = mx.sym.SliceChannel(gates, num_outputs=4, name="t%d_l%d_slice" % (seqidx, layeridx)) Wcidc = mx.sym.broadcast_mul(param.c2i_bias, prev_state.c) + slice_gates[0] in_gate = mx.sym.Activation(Wcidc, act_type="sigmoid") in_transform = mx.sym.Activation(slice_gates[1], act_type="tanh") Wcfdc = mx.sym.broadcast_mul(param.c2f_bias, prev_state.c) + slice_gates[2] forget_gate = mx.sym.Activation(Wcfdc, act_type="sigmoid") next_c = (forget_gate * prev_state.c) + (in_gate * in_transform) Wcoct = mx.sym.broadcast_mul(param.c2o_bias, next_c) + slice_gates[3] out_gate = mx.sym.Activation(Wcoct, act_type="sigmoid") next_h = out_gate * mx.sym.Activation(next_c, act_type="tanh") if num_hidden_proj > 0: proj_next_h = mx.sym.FullyConnected(data=next_h, weight=param.ph2h_weight, no_bias=True, num_hidden=num_hidden_proj, name="t%d_l%d_ph2h" % (seqidx, layeridx)) return LSTMState(c=next_c, h=proj_next_h) else: return LSTMState(c=next_c, h=next_h)
python
def lstm(num_hidden, indata, prev_state, param, seqidx, layeridx, dropout=0., num_hidden_proj=0, is_batchnorm=False, gamma=None, beta=None, name=None): """LSTM Cell symbol""" # dropout input if dropout > 0.: indata = mx.sym.Dropout(data=indata, p=dropout) i2h = mx.sym.FullyConnected(data=indata, weight=param.i2h_weight, bias=param.i2h_bias, num_hidden=num_hidden * 4, name="t%d_l%d_i2h" % (seqidx, layeridx)) if is_batchnorm: if name is not None: i2h = batchnorm(net=i2h, gamma=gamma, beta=beta, name="%s_batchnorm" % name) else: i2h = batchnorm(net=i2h, gamma=gamma, beta=beta) h2h = mx.sym.FullyConnected(data=prev_state.h, weight=param.h2h_weight, # bias=param.h2h_bias, no_bias=True, num_hidden=num_hidden * 4, name="t%d_l%d_h2h" % (seqidx, layeridx)) gates = i2h + h2h slice_gates = mx.sym.SliceChannel(gates, num_outputs=4, name="t%d_l%d_slice" % (seqidx, layeridx)) Wcidc = mx.sym.broadcast_mul(param.c2i_bias, prev_state.c) + slice_gates[0] in_gate = mx.sym.Activation(Wcidc, act_type="sigmoid") in_transform = mx.sym.Activation(slice_gates[1], act_type="tanh") Wcfdc = mx.sym.broadcast_mul(param.c2f_bias, prev_state.c) + slice_gates[2] forget_gate = mx.sym.Activation(Wcfdc, act_type="sigmoid") next_c = (forget_gate * prev_state.c) + (in_gate * in_transform) Wcoct = mx.sym.broadcast_mul(param.c2o_bias, next_c) + slice_gates[3] out_gate = mx.sym.Activation(Wcoct, act_type="sigmoid") next_h = out_gate * mx.sym.Activation(next_c, act_type="tanh") if num_hidden_proj > 0: proj_next_h = mx.sym.FullyConnected(data=next_h, weight=param.ph2h_weight, no_bias=True, num_hidden=num_hidden_proj, name="t%d_l%d_ph2h" % (seqidx, layeridx)) return LSTMState(c=next_c, h=proj_next_h) else: return LSTMState(c=next_c, h=next_h)
[ "def", "lstm", "(", "num_hidden", ",", "indata", ",", "prev_state", ",", "param", ",", "seqidx", ",", "layeridx", ",", "dropout", "=", "0.", ",", "num_hidden_proj", "=", "0", ",", "is_batchnorm", "=", "False", ",", "gamma", "=", "None", ",", "beta", "=", "None", ",", "name", "=", "None", ")", ":", "# dropout input", "if", "dropout", ">", "0.", ":", "indata", "=", "mx", ".", "sym", ".", "Dropout", "(", "data", "=", "indata", ",", "p", "=", "dropout", ")", "i2h", "=", "mx", ".", "sym", ".", "FullyConnected", "(", "data", "=", "indata", ",", "weight", "=", "param", ".", "i2h_weight", ",", "bias", "=", "param", ".", "i2h_bias", ",", "num_hidden", "=", "num_hidden", "*", "4", ",", "name", "=", "\"t%d_l%d_i2h\"", "%", "(", "seqidx", ",", "layeridx", ")", ")", "if", "is_batchnorm", ":", "if", "name", "is", "not", "None", ":", "i2h", "=", "batchnorm", "(", "net", "=", "i2h", ",", "gamma", "=", "gamma", ",", "beta", "=", "beta", ",", "name", "=", "\"%s_batchnorm\"", "%", "name", ")", "else", ":", "i2h", "=", "batchnorm", "(", "net", "=", "i2h", ",", "gamma", "=", "gamma", ",", "beta", "=", "beta", ")", "h2h", "=", "mx", ".", "sym", ".", "FullyConnected", "(", "data", "=", "prev_state", ".", "h", ",", "weight", "=", "param", ".", "h2h_weight", ",", "# bias=param.h2h_bias,", "no_bias", "=", "True", ",", "num_hidden", "=", "num_hidden", "*", "4", ",", "name", "=", "\"t%d_l%d_h2h\"", "%", "(", "seqidx", ",", "layeridx", ")", ")", "gates", "=", "i2h", "+", "h2h", "slice_gates", "=", "mx", ".", "sym", ".", "SliceChannel", "(", "gates", ",", "num_outputs", "=", "4", ",", "name", "=", "\"t%d_l%d_slice\"", "%", "(", "seqidx", ",", "layeridx", ")", ")", "Wcidc", "=", "mx", ".", "sym", ".", "broadcast_mul", "(", "param", ".", "c2i_bias", ",", "prev_state", ".", "c", ")", "+", "slice_gates", "[", "0", "]", "in_gate", "=", "mx", ".", "sym", ".", "Activation", "(", "Wcidc", ",", "act_type", "=", "\"sigmoid\"", ")", "in_transform", "=", "mx", ".", "sym", ".", "Activation", "(", "slice_gates", "[", "1", "]", ",", "act_type", "=", "\"tanh\"", ")", "Wcfdc", "=", "mx", ".", "sym", ".", "broadcast_mul", "(", "param", ".", "c2f_bias", ",", "prev_state", ".", "c", ")", "+", "slice_gates", "[", "2", "]", "forget_gate", "=", "mx", ".", "sym", ".", "Activation", "(", "Wcfdc", ",", "act_type", "=", "\"sigmoid\"", ")", "next_c", "=", "(", "forget_gate", "*", "prev_state", ".", "c", ")", "+", "(", "in_gate", "*", "in_transform", ")", "Wcoct", "=", "mx", ".", "sym", ".", "broadcast_mul", "(", "param", ".", "c2o_bias", ",", "next_c", ")", "+", "slice_gates", "[", "3", "]", "out_gate", "=", "mx", ".", "sym", ".", "Activation", "(", "Wcoct", ",", "act_type", "=", "\"sigmoid\"", ")", "next_h", "=", "out_gate", "*", "mx", ".", "sym", ".", "Activation", "(", "next_c", ",", "act_type", "=", "\"tanh\"", ")", "if", "num_hidden_proj", ">", "0", ":", "proj_next_h", "=", "mx", ".", "sym", ".", "FullyConnected", "(", "data", "=", "next_h", ",", "weight", "=", "param", ".", "ph2h_weight", ",", "no_bias", "=", "True", ",", "num_hidden", "=", "num_hidden_proj", ",", "name", "=", "\"t%d_l%d_ph2h\"", "%", "(", "seqidx", ",", "layeridx", ")", ")", "return", "LSTMState", "(", "c", "=", "next_c", ",", "h", "=", "proj_next_h", ")", "else", ":", "return", "LSTMState", "(", "c", "=", "next_c", ",", "h", "=", "next_h", ")" ]
LSTM Cell symbol
[ "LSTM", "Cell", "symbol" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/speech_recognition/stt_layer_lstm.py#L65-L118
train
apache/incubator-mxnet
example/rcnn/symdata/image.py
get_image
def get_image(roi_rec, short, max_size, mean, std): """ read, resize, transform image, return im_tensor, im_info, gt_boxes roi_rec should have keys: ["image", "boxes", "gt_classes", "flipped"] 0 --- x (width, second dim of im) | y (height, first dim of im) """ im = imdecode(roi_rec['image']) if roi_rec["flipped"]: im = im[:, ::-1, :] im, im_scale = resize(im, short, max_size) height, width = im.shape[:2] im_info = np.array([height, width, im_scale], dtype=np.float32) im_tensor = transform(im, mean, std) # gt boxes: (x1, y1, x2, y2, cls) if roi_rec['gt_classes'].size > 0: gt_inds = np.where(roi_rec['gt_classes'] != 0)[0] gt_boxes = np.empty((len(gt_inds), 5), dtype=np.float32) gt_boxes[:, 0:4] = roi_rec['boxes'][gt_inds, :] gt_boxes[:, 4] = roi_rec['gt_classes'][gt_inds] # scale gt_boxes gt_boxes[:, 0:4] *= im_scale else: gt_boxes = np.empty((0, 5), dtype=np.float32) return im_tensor, im_info, gt_boxes
python
def get_image(roi_rec, short, max_size, mean, std): """ read, resize, transform image, return im_tensor, im_info, gt_boxes roi_rec should have keys: ["image", "boxes", "gt_classes", "flipped"] 0 --- x (width, second dim of im) | y (height, first dim of im) """ im = imdecode(roi_rec['image']) if roi_rec["flipped"]: im = im[:, ::-1, :] im, im_scale = resize(im, short, max_size) height, width = im.shape[:2] im_info = np.array([height, width, im_scale], dtype=np.float32) im_tensor = transform(im, mean, std) # gt boxes: (x1, y1, x2, y2, cls) if roi_rec['gt_classes'].size > 0: gt_inds = np.where(roi_rec['gt_classes'] != 0)[0] gt_boxes = np.empty((len(gt_inds), 5), dtype=np.float32) gt_boxes[:, 0:4] = roi_rec['boxes'][gt_inds, :] gt_boxes[:, 4] = roi_rec['gt_classes'][gt_inds] # scale gt_boxes gt_boxes[:, 0:4] *= im_scale else: gt_boxes = np.empty((0, 5), dtype=np.float32) return im_tensor, im_info, gt_boxes
[ "def", "get_image", "(", "roi_rec", ",", "short", ",", "max_size", ",", "mean", ",", "std", ")", ":", "im", "=", "imdecode", "(", "roi_rec", "[", "'image'", "]", ")", "if", "roi_rec", "[", "\"flipped\"", "]", ":", "im", "=", "im", "[", ":", ",", ":", ":", "-", "1", ",", ":", "]", "im", ",", "im_scale", "=", "resize", "(", "im", ",", "short", ",", "max_size", ")", "height", ",", "width", "=", "im", ".", "shape", "[", ":", "2", "]", "im_info", "=", "np", ".", "array", "(", "[", "height", ",", "width", ",", "im_scale", "]", ",", "dtype", "=", "np", ".", "float32", ")", "im_tensor", "=", "transform", "(", "im", ",", "mean", ",", "std", ")", "# gt boxes: (x1, y1, x2, y2, cls)", "if", "roi_rec", "[", "'gt_classes'", "]", ".", "size", ">", "0", ":", "gt_inds", "=", "np", ".", "where", "(", "roi_rec", "[", "'gt_classes'", "]", "!=", "0", ")", "[", "0", "]", "gt_boxes", "=", "np", ".", "empty", "(", "(", "len", "(", "gt_inds", ")", ",", "5", ")", ",", "dtype", "=", "np", ".", "float32", ")", "gt_boxes", "[", ":", ",", "0", ":", "4", "]", "=", "roi_rec", "[", "'boxes'", "]", "[", "gt_inds", ",", ":", "]", "gt_boxes", "[", ":", ",", "4", "]", "=", "roi_rec", "[", "'gt_classes'", "]", "[", "gt_inds", "]", "# scale gt_boxes", "gt_boxes", "[", ":", ",", "0", ":", "4", "]", "*=", "im_scale", "else", ":", "gt_boxes", "=", "np", ".", "empty", "(", "(", "0", ",", "5", ")", ",", "dtype", "=", "np", ".", "float32", ")", "return", "im_tensor", ",", "im_info", ",", "gt_boxes" ]
read, resize, transform image, return im_tensor, im_info, gt_boxes roi_rec should have keys: ["image", "boxes", "gt_classes", "flipped"] 0 --- x (width, second dim of im) | y (height, first dim of im)
[ "read", "resize", "transform", "image", "return", "im_tensor", "im_info", "gt_boxes", "roi_rec", "should", "have", "keys", ":", "[", "image", "boxes", "gt_classes", "flipped", "]", "0", "---", "x", "(", "width", "second", "dim", "of", "im", ")", "|", "y", "(", "height", "first", "dim", "of", "im", ")" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/rcnn/symdata/image.py#L22-L49
train
apache/incubator-mxnet
example/rcnn/symdata/image.py
imdecode
def imdecode(image_path): """Return BGR image read by opencv""" import os assert os.path.exists(image_path), image_path + ' not found' im = cv2.imread(image_path) return im
python
def imdecode(image_path): """Return BGR image read by opencv""" import os assert os.path.exists(image_path), image_path + ' not found' im = cv2.imread(image_path) return im
[ "def", "imdecode", "(", "image_path", ")", ":", "import", "os", "assert", "os", ".", "path", ".", "exists", "(", "image_path", ")", ",", "image_path", "+", "' not found'", "im", "=", "cv2", ".", "imread", "(", "image_path", ")", "return", "im" ]
Return BGR image read by opencv
[ "Return", "BGR", "image", "read", "by", "opencv" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/rcnn/symdata/image.py#L52-L57
train
apache/incubator-mxnet
example/rcnn/symdata/image.py
resize
def resize(im, short, max_size): """ only resize input image to target size and return scale :param im: BGR image input by opencv :param short: one dimensional size (the short side) :param max_size: one dimensional max size (the long side) :return: resized image (NDArray) and scale (float) """ im_shape = im.shape im_size_min = np.min(im_shape[0:2]) im_size_max = np.max(im_shape[0:2]) im_scale = float(short) / float(im_size_min) # prevent bigger axis from being more than max_size: if np.round(im_scale * im_size_max) > max_size: im_scale = float(max_size) / float(im_size_max) im = cv2.resize(im, None, None, fx=im_scale, fy=im_scale, interpolation=cv2.INTER_LINEAR) return im, im_scale
python
def resize(im, short, max_size): """ only resize input image to target size and return scale :param im: BGR image input by opencv :param short: one dimensional size (the short side) :param max_size: one dimensional max size (the long side) :return: resized image (NDArray) and scale (float) """ im_shape = im.shape im_size_min = np.min(im_shape[0:2]) im_size_max = np.max(im_shape[0:2]) im_scale = float(short) / float(im_size_min) # prevent bigger axis from being more than max_size: if np.round(im_scale * im_size_max) > max_size: im_scale = float(max_size) / float(im_size_max) im = cv2.resize(im, None, None, fx=im_scale, fy=im_scale, interpolation=cv2.INTER_LINEAR) return im, im_scale
[ "def", "resize", "(", "im", ",", "short", ",", "max_size", ")", ":", "im_shape", "=", "im", ".", "shape", "im_size_min", "=", "np", ".", "min", "(", "im_shape", "[", "0", ":", "2", "]", ")", "im_size_max", "=", "np", ".", "max", "(", "im_shape", "[", "0", ":", "2", "]", ")", "im_scale", "=", "float", "(", "short", ")", "/", "float", "(", "im_size_min", ")", "# prevent bigger axis from being more than max_size:", "if", "np", ".", "round", "(", "im_scale", "*", "im_size_max", ")", ">", "max_size", ":", "im_scale", "=", "float", "(", "max_size", ")", "/", "float", "(", "im_size_max", ")", "im", "=", "cv2", ".", "resize", "(", "im", ",", "None", ",", "None", ",", "fx", "=", "im_scale", ",", "fy", "=", "im_scale", ",", "interpolation", "=", "cv2", ".", "INTER_LINEAR", ")", "return", "im", ",", "im_scale" ]
only resize input image to target size and return scale :param im: BGR image input by opencv :param short: one dimensional size (the short side) :param max_size: one dimensional max size (the long side) :return: resized image (NDArray) and scale (float)
[ "only", "resize", "input", "image", "to", "target", "size", "and", "return", "scale", ":", "param", "im", ":", "BGR", "image", "input", "by", "opencv", ":", "param", "short", ":", "one", "dimensional", "size", "(", "the", "short", "side", ")", ":", "param", "max_size", ":", "one", "dimensional", "max", "size", "(", "the", "long", "side", ")", ":", "return", ":", "resized", "image", "(", "NDArray", ")", "and", "scale", "(", "float", ")" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/rcnn/symdata/image.py#L60-L76
train
apache/incubator-mxnet
example/rcnn/symdata/image.py
transform
def transform(im, mean, std): """ transform into mxnet tensor, subtract pixel size and transform to correct format :param im: [height, width, channel] in BGR :param mean: [RGB pixel mean] :param std: [RGB pixel std var] :return: [batch, channel, height, width] """ im_tensor = np.zeros((3, im.shape[0], im.shape[1])) for i in range(3): im_tensor[i, :, :] = (im[:, :, 2 - i] - mean[i]) / std[i] return im_tensor
python
def transform(im, mean, std): """ transform into mxnet tensor, subtract pixel size and transform to correct format :param im: [height, width, channel] in BGR :param mean: [RGB pixel mean] :param std: [RGB pixel std var] :return: [batch, channel, height, width] """ im_tensor = np.zeros((3, im.shape[0], im.shape[1])) for i in range(3): im_tensor[i, :, :] = (im[:, :, 2 - i] - mean[i]) / std[i] return im_tensor
[ "def", "transform", "(", "im", ",", "mean", ",", "std", ")", ":", "im_tensor", "=", "np", ".", "zeros", "(", "(", "3", ",", "im", ".", "shape", "[", "0", "]", ",", "im", ".", "shape", "[", "1", "]", ")", ")", "for", "i", "in", "range", "(", "3", ")", ":", "im_tensor", "[", "i", ",", ":", ",", ":", "]", "=", "(", "im", "[", ":", ",", ":", ",", "2", "-", "i", "]", "-", "mean", "[", "i", "]", ")", "/", "std", "[", "i", "]", "return", "im_tensor" ]
transform into mxnet tensor, subtract pixel size and transform to correct format :param im: [height, width, channel] in BGR :param mean: [RGB pixel mean] :param std: [RGB pixel std var] :return: [batch, channel, height, width]
[ "transform", "into", "mxnet", "tensor", "subtract", "pixel", "size", "and", "transform", "to", "correct", "format", ":", "param", "im", ":", "[", "height", "width", "channel", "]", "in", "BGR", ":", "param", "mean", ":", "[", "RGB", "pixel", "mean", "]", ":", "param", "std", ":", "[", "RGB", "pixel", "std", "var", "]", ":", "return", ":", "[", "batch", "channel", "height", "width", "]" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/rcnn/symdata/image.py#L79-L91
train
apache/incubator-mxnet
example/rcnn/symdata/image.py
transform_inverse
def transform_inverse(im_tensor, mean, std): """ transform from mxnet im_tensor to ordinary RGB image im_tensor is limited to one image :param im_tensor: [batch, channel, height, width] :param mean: [RGB pixel mean] :param std: [RGB pixel std var] :return: im [height, width, channel(RGB)] """ assert im_tensor.shape[0] == 3 im = im_tensor.transpose((1, 2, 0)) im = im * std + mean im = im.astype(np.uint8) return im
python
def transform_inverse(im_tensor, mean, std): """ transform from mxnet im_tensor to ordinary RGB image im_tensor is limited to one image :param im_tensor: [batch, channel, height, width] :param mean: [RGB pixel mean] :param std: [RGB pixel std var] :return: im [height, width, channel(RGB)] """ assert im_tensor.shape[0] == 3 im = im_tensor.transpose((1, 2, 0)) im = im * std + mean im = im.astype(np.uint8) return im
[ "def", "transform_inverse", "(", "im_tensor", ",", "mean", ",", "std", ")", ":", "assert", "im_tensor", ".", "shape", "[", "0", "]", "==", "3", "im", "=", "im_tensor", ".", "transpose", "(", "(", "1", ",", "2", ",", "0", ")", ")", "im", "=", "im", "*", "std", "+", "mean", "im", "=", "im", ".", "astype", "(", "np", ".", "uint8", ")", "return", "im" ]
transform from mxnet im_tensor to ordinary RGB image im_tensor is limited to one image :param im_tensor: [batch, channel, height, width] :param mean: [RGB pixel mean] :param std: [RGB pixel std var] :return: im [height, width, channel(RGB)]
[ "transform", "from", "mxnet", "im_tensor", "to", "ordinary", "RGB", "image", "im_tensor", "is", "limited", "to", "one", "image", ":", "param", "im_tensor", ":", "[", "batch", "channel", "height", "width", "]", ":", "param", "mean", ":", "[", "RGB", "pixel", "mean", "]", ":", "param", "std", ":", "[", "RGB", "pixel", "std", "var", "]", ":", "return", ":", "im", "[", "height", "width", "channel", "(", "RGB", ")", "]" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/rcnn/symdata/image.py#L94-L107
train
apache/incubator-mxnet
example/rcnn/symdata/image.py
tensor_vstack
def tensor_vstack(tensor_list, pad=0): """ vertically stack tensors by adding a new axis expand dims if only 1 tensor :param tensor_list: list of tensor to be stacked vertically :param pad: label to pad with :return: tensor with max shape """ if len(tensor_list) == 1: return tensor_list[0][np.newaxis, :] ndim = len(tensor_list[0].shape) dimensions = [len(tensor_list)] # first dim is batch size for dim in range(ndim): dimensions.append(max([tensor.shape[dim] for tensor in tensor_list])) dtype = tensor_list[0].dtype if pad == 0: all_tensor = np.zeros(tuple(dimensions), dtype=dtype) elif pad == 1: all_tensor = np.ones(tuple(dimensions), dtype=dtype) else: all_tensor = np.full(tuple(dimensions), pad, dtype=dtype) if ndim == 1: for ind, tensor in enumerate(tensor_list): all_tensor[ind, :tensor.shape[0]] = tensor elif ndim == 2: for ind, tensor in enumerate(tensor_list): all_tensor[ind, :tensor.shape[0], :tensor.shape[1]] = tensor elif ndim == 3: for ind, tensor in enumerate(tensor_list): all_tensor[ind, :tensor.shape[0], :tensor.shape[1], :tensor.shape[2]] = tensor else: raise Exception('Sorry, unimplemented.') return all_tensor
python
def tensor_vstack(tensor_list, pad=0): """ vertically stack tensors by adding a new axis expand dims if only 1 tensor :param tensor_list: list of tensor to be stacked vertically :param pad: label to pad with :return: tensor with max shape """ if len(tensor_list) == 1: return tensor_list[0][np.newaxis, :] ndim = len(tensor_list[0].shape) dimensions = [len(tensor_list)] # first dim is batch size for dim in range(ndim): dimensions.append(max([tensor.shape[dim] for tensor in tensor_list])) dtype = tensor_list[0].dtype if pad == 0: all_tensor = np.zeros(tuple(dimensions), dtype=dtype) elif pad == 1: all_tensor = np.ones(tuple(dimensions), dtype=dtype) else: all_tensor = np.full(tuple(dimensions), pad, dtype=dtype) if ndim == 1: for ind, tensor in enumerate(tensor_list): all_tensor[ind, :tensor.shape[0]] = tensor elif ndim == 2: for ind, tensor in enumerate(tensor_list): all_tensor[ind, :tensor.shape[0], :tensor.shape[1]] = tensor elif ndim == 3: for ind, tensor in enumerate(tensor_list): all_tensor[ind, :tensor.shape[0], :tensor.shape[1], :tensor.shape[2]] = tensor else: raise Exception('Sorry, unimplemented.') return all_tensor
[ "def", "tensor_vstack", "(", "tensor_list", ",", "pad", "=", "0", ")", ":", "if", "len", "(", "tensor_list", ")", "==", "1", ":", "return", "tensor_list", "[", "0", "]", "[", "np", ".", "newaxis", ",", ":", "]", "ndim", "=", "len", "(", "tensor_list", "[", "0", "]", ".", "shape", ")", "dimensions", "=", "[", "len", "(", "tensor_list", ")", "]", "# first dim is batch size", "for", "dim", "in", "range", "(", "ndim", ")", ":", "dimensions", ".", "append", "(", "max", "(", "[", "tensor", ".", "shape", "[", "dim", "]", "for", "tensor", "in", "tensor_list", "]", ")", ")", "dtype", "=", "tensor_list", "[", "0", "]", ".", "dtype", "if", "pad", "==", "0", ":", "all_tensor", "=", "np", ".", "zeros", "(", "tuple", "(", "dimensions", ")", ",", "dtype", "=", "dtype", ")", "elif", "pad", "==", "1", ":", "all_tensor", "=", "np", ".", "ones", "(", "tuple", "(", "dimensions", ")", ",", "dtype", "=", "dtype", ")", "else", ":", "all_tensor", "=", "np", ".", "full", "(", "tuple", "(", "dimensions", ")", ",", "pad", ",", "dtype", "=", "dtype", ")", "if", "ndim", "==", "1", ":", "for", "ind", ",", "tensor", "in", "enumerate", "(", "tensor_list", ")", ":", "all_tensor", "[", "ind", ",", ":", "tensor", ".", "shape", "[", "0", "]", "]", "=", "tensor", "elif", "ndim", "==", "2", ":", "for", "ind", ",", "tensor", "in", "enumerate", "(", "tensor_list", ")", ":", "all_tensor", "[", "ind", ",", ":", "tensor", ".", "shape", "[", "0", "]", ",", ":", "tensor", ".", "shape", "[", "1", "]", "]", "=", "tensor", "elif", "ndim", "==", "3", ":", "for", "ind", ",", "tensor", "in", "enumerate", "(", "tensor_list", ")", ":", "all_tensor", "[", "ind", ",", ":", "tensor", ".", "shape", "[", "0", "]", ",", ":", "tensor", ".", "shape", "[", "1", "]", ",", ":", "tensor", ".", "shape", "[", "2", "]", "]", "=", "tensor", "else", ":", "raise", "Exception", "(", "'Sorry, unimplemented.'", ")", "return", "all_tensor" ]
vertically stack tensors by adding a new axis expand dims if only 1 tensor :param tensor_list: list of tensor to be stacked vertically :param pad: label to pad with :return: tensor with max shape
[ "vertically", "stack", "tensors", "by", "adding", "a", "new", "axis", "expand", "dims", "if", "only", "1", "tensor", ":", "param", "tensor_list", ":", "list", "of", "tensor", "to", "be", "stacked", "vertically", ":", "param", "pad", ":", "label", "to", "pad", "with", ":", "return", ":", "tensor", "with", "max", "shape" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/rcnn/symdata/image.py#L110-L144
train
apache/incubator-mxnet
example/gluon/embedding_learning/train.py
get_distance_matrix
def get_distance_matrix(x): """Get distance matrix given a matrix. Used in testing.""" square = nd.sum(x ** 2.0, axis=1, keepdims=True) distance_square = square + square.transpose() - (2.0 * nd.dot(x, x.transpose())) return nd.sqrt(distance_square)
python
def get_distance_matrix(x): """Get distance matrix given a matrix. Used in testing.""" square = nd.sum(x ** 2.0, axis=1, keepdims=True) distance_square = square + square.transpose() - (2.0 * nd.dot(x, x.transpose())) return nd.sqrt(distance_square)
[ "def", "get_distance_matrix", "(", "x", ")", ":", "square", "=", "nd", ".", "sum", "(", "x", "**", "2.0", ",", "axis", "=", "1", ",", "keepdims", "=", "True", ")", "distance_square", "=", "square", "+", "square", ".", "transpose", "(", ")", "-", "(", "2.0", "*", "nd", ".", "dot", "(", "x", ",", "x", ".", "transpose", "(", ")", ")", ")", "return", "nd", ".", "sqrt", "(", "distance_square", ")" ]
Get distance matrix given a matrix. Used in testing.
[ "Get", "distance", "matrix", "given", "a", "matrix", ".", "Used", "in", "testing", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/embedding_learning/train.py#L116-L120
train
apache/incubator-mxnet
example/gluon/embedding_learning/train.py
evaluate_emb
def evaluate_emb(emb, labels): """Evaluate embeddings based on Recall@k.""" d_mat = get_distance_matrix(emb) d_mat = d_mat.asnumpy() labels = labels.asnumpy() names = [] accs = [] for k in [1, 2, 4, 8, 16]: names.append('Recall@%d' % k) correct, cnt = 0.0, 0.0 for i in range(emb.shape[0]): d_mat[i, i] = 1e10 nns = argpartition(d_mat[i], k)[:k] if any(labels[i] == labels[nn] for nn in nns): correct += 1 cnt += 1 accs.append(correct/cnt) return names, accs
python
def evaluate_emb(emb, labels): """Evaluate embeddings based on Recall@k.""" d_mat = get_distance_matrix(emb) d_mat = d_mat.asnumpy() labels = labels.asnumpy() names = [] accs = [] for k in [1, 2, 4, 8, 16]: names.append('Recall@%d' % k) correct, cnt = 0.0, 0.0 for i in range(emb.shape[0]): d_mat[i, i] = 1e10 nns = argpartition(d_mat[i], k)[:k] if any(labels[i] == labels[nn] for nn in nns): correct += 1 cnt += 1 accs.append(correct/cnt) return names, accs
[ "def", "evaluate_emb", "(", "emb", ",", "labels", ")", ":", "d_mat", "=", "get_distance_matrix", "(", "emb", ")", "d_mat", "=", "d_mat", ".", "asnumpy", "(", ")", "labels", "=", "labels", ".", "asnumpy", "(", ")", "names", "=", "[", "]", "accs", "=", "[", "]", "for", "k", "in", "[", "1", ",", "2", ",", "4", ",", "8", ",", "16", "]", ":", "names", ".", "append", "(", "'Recall@%d'", "%", "k", ")", "correct", ",", "cnt", "=", "0.0", ",", "0.0", "for", "i", "in", "range", "(", "emb", ".", "shape", "[", "0", "]", ")", ":", "d_mat", "[", "i", ",", "i", "]", "=", "1e10", "nns", "=", "argpartition", "(", "d_mat", "[", "i", "]", ",", "k", ")", "[", ":", "k", "]", "if", "any", "(", "labels", "[", "i", "]", "==", "labels", "[", "nn", "]", "for", "nn", "in", "nns", ")", ":", "correct", "+=", "1", "cnt", "+=", "1", "accs", ".", "append", "(", "correct", "/", "cnt", ")", "return", "names", ",", "accs" ]
Evaluate embeddings based on Recall@k.
[ "Evaluate", "embeddings", "based", "on", "Recall" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/embedding_learning/train.py#L123-L141
train
apache/incubator-mxnet
example/gluon/embedding_learning/train.py
get_lr
def get_lr(lr, epoch, steps, factor): """Get learning rate based on schedule.""" for s in steps: if epoch >= s: lr *= factor return lr
python
def get_lr(lr, epoch, steps, factor): """Get learning rate based on schedule.""" for s in steps: if epoch >= s: lr *= factor return lr
[ "def", "get_lr", "(", "lr", ",", "epoch", ",", "steps", ",", "factor", ")", ":", "for", "s", "in", "steps", ":", "if", "epoch", ">=", "s", ":", "lr", "*=", "factor", "return", "lr" ]
Get learning rate based on schedule.
[ "Get", "learning", "rate", "based", "on", "schedule", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/embedding_learning/train.py#L161-L166
train
apache/incubator-mxnet
example/gluon/embedding_learning/train.py
train
def train(epochs, ctx): """Training function.""" if isinstance(ctx, mx.Context): ctx = [ctx] net.initialize(mx.init.Xavier(magnitude=2), ctx=ctx) opt_options = {'learning_rate': opt.lr, 'wd': opt.wd} if opt.optimizer == 'sgd': opt_options['momentum'] = 0.9 if opt.optimizer == 'adam': opt_options['epsilon'] = 1e-7 trainer = gluon.Trainer(net.collect_params(), opt.optimizer, opt_options, kvstore=opt.kvstore) if opt.lr_beta > 0.0: # Jointly train class-specific beta. # See "sampling matters in deep embedding learning" paper for details. beta.initialize(mx.init.Constant(opt.beta), ctx=ctx) trainer_beta = gluon.Trainer([beta], 'sgd', {'learning_rate': opt.lr_beta, 'momentum': 0.9}, kvstore=opt.kvstore) loss = MarginLoss(margin=opt.margin, nu=opt.nu) best_val = 0.0 for epoch in range(epochs): tic = time.time() prev_loss, cumulative_loss = 0.0, 0.0 # Learning rate schedule. trainer.set_learning_rate(get_lr(opt.lr, epoch, steps, opt.factor)) logging.info('Epoch %d learning rate=%f', epoch, trainer.learning_rate) if opt.lr_beta > 0.0: trainer_beta.set_learning_rate(get_lr(opt.lr_beta, epoch, steps, opt.factor)) logging.info('Epoch %d beta learning rate=%f', epoch, trainer_beta.learning_rate) # Inner training loop. for i in range(200): batch = train_data.next() data = gluon.utils.split_and_load(batch.data[0], ctx_list=ctx, batch_axis=0) label = gluon.utils.split_and_load(batch.label[0], ctx_list=ctx, batch_axis=0) Ls = [] with ag.record(): for x, y in zip(data, label): a_indices, anchors, positives, negatives, _ = net(x) if opt.lr_beta > 0.0: L = loss(anchors, positives, negatives, beta, y[a_indices]) else: L = loss(anchors, positives, negatives, opt.beta, None) # Store the loss and do backward after we have done forward # on all GPUs for better speed on multiple GPUs. Ls.append(L) cumulative_loss += nd.mean(L).asscalar() for L in Ls: L.backward() # Update. trainer.step(batch.data[0].shape[0]) if opt.lr_beta > 0.0: trainer_beta.step(batch.data[0].shape[0]) if (i+1) % opt.log_interval == 0: logging.info('[Epoch %d, Iter %d] training loss=%f' % ( epoch, i+1, cumulative_loss - prev_loss)) prev_loss = cumulative_loss logging.info('[Epoch %d] training loss=%f'%(epoch, cumulative_loss)) logging.info('[Epoch %d] time cost: %f'%(epoch, time.time()-tic)) names, val_accs = test(ctx) for name, val_acc in zip(names, val_accs): logging.info('[Epoch %d] validation: %s=%f'%(epoch, name, val_acc)) if val_accs[0] > best_val: best_val = val_accs[0] logging.info('Saving %s.' % opt.save_model_prefix) net.save_parameters('%s.params' % opt.save_model_prefix) return best_val
python
def train(epochs, ctx): """Training function.""" if isinstance(ctx, mx.Context): ctx = [ctx] net.initialize(mx.init.Xavier(magnitude=2), ctx=ctx) opt_options = {'learning_rate': opt.lr, 'wd': opt.wd} if opt.optimizer == 'sgd': opt_options['momentum'] = 0.9 if opt.optimizer == 'adam': opt_options['epsilon'] = 1e-7 trainer = gluon.Trainer(net.collect_params(), opt.optimizer, opt_options, kvstore=opt.kvstore) if opt.lr_beta > 0.0: # Jointly train class-specific beta. # See "sampling matters in deep embedding learning" paper for details. beta.initialize(mx.init.Constant(opt.beta), ctx=ctx) trainer_beta = gluon.Trainer([beta], 'sgd', {'learning_rate': opt.lr_beta, 'momentum': 0.9}, kvstore=opt.kvstore) loss = MarginLoss(margin=opt.margin, nu=opt.nu) best_val = 0.0 for epoch in range(epochs): tic = time.time() prev_loss, cumulative_loss = 0.0, 0.0 # Learning rate schedule. trainer.set_learning_rate(get_lr(opt.lr, epoch, steps, opt.factor)) logging.info('Epoch %d learning rate=%f', epoch, trainer.learning_rate) if opt.lr_beta > 0.0: trainer_beta.set_learning_rate(get_lr(opt.lr_beta, epoch, steps, opt.factor)) logging.info('Epoch %d beta learning rate=%f', epoch, trainer_beta.learning_rate) # Inner training loop. for i in range(200): batch = train_data.next() data = gluon.utils.split_and_load(batch.data[0], ctx_list=ctx, batch_axis=0) label = gluon.utils.split_and_load(batch.label[0], ctx_list=ctx, batch_axis=0) Ls = [] with ag.record(): for x, y in zip(data, label): a_indices, anchors, positives, negatives, _ = net(x) if opt.lr_beta > 0.0: L = loss(anchors, positives, negatives, beta, y[a_indices]) else: L = loss(anchors, positives, negatives, opt.beta, None) # Store the loss and do backward after we have done forward # on all GPUs for better speed on multiple GPUs. Ls.append(L) cumulative_loss += nd.mean(L).asscalar() for L in Ls: L.backward() # Update. trainer.step(batch.data[0].shape[0]) if opt.lr_beta > 0.0: trainer_beta.step(batch.data[0].shape[0]) if (i+1) % opt.log_interval == 0: logging.info('[Epoch %d, Iter %d] training loss=%f' % ( epoch, i+1, cumulative_loss - prev_loss)) prev_loss = cumulative_loss logging.info('[Epoch %d] training loss=%f'%(epoch, cumulative_loss)) logging.info('[Epoch %d] time cost: %f'%(epoch, time.time()-tic)) names, val_accs = test(ctx) for name, val_acc in zip(names, val_accs): logging.info('[Epoch %d] validation: %s=%f'%(epoch, name, val_acc)) if val_accs[0] > best_val: best_val = val_accs[0] logging.info('Saving %s.' % opt.save_model_prefix) net.save_parameters('%s.params' % opt.save_model_prefix) return best_val
[ "def", "train", "(", "epochs", ",", "ctx", ")", ":", "if", "isinstance", "(", "ctx", ",", "mx", ".", "Context", ")", ":", "ctx", "=", "[", "ctx", "]", "net", ".", "initialize", "(", "mx", ".", "init", ".", "Xavier", "(", "magnitude", "=", "2", ")", ",", "ctx", "=", "ctx", ")", "opt_options", "=", "{", "'learning_rate'", ":", "opt", ".", "lr", ",", "'wd'", ":", "opt", ".", "wd", "}", "if", "opt", ".", "optimizer", "==", "'sgd'", ":", "opt_options", "[", "'momentum'", "]", "=", "0.9", "if", "opt", ".", "optimizer", "==", "'adam'", ":", "opt_options", "[", "'epsilon'", "]", "=", "1e-7", "trainer", "=", "gluon", ".", "Trainer", "(", "net", ".", "collect_params", "(", ")", ",", "opt", ".", "optimizer", ",", "opt_options", ",", "kvstore", "=", "opt", ".", "kvstore", ")", "if", "opt", ".", "lr_beta", ">", "0.0", ":", "# Jointly train class-specific beta.", "# See \"sampling matters in deep embedding learning\" paper for details.", "beta", ".", "initialize", "(", "mx", ".", "init", ".", "Constant", "(", "opt", ".", "beta", ")", ",", "ctx", "=", "ctx", ")", "trainer_beta", "=", "gluon", ".", "Trainer", "(", "[", "beta", "]", ",", "'sgd'", ",", "{", "'learning_rate'", ":", "opt", ".", "lr_beta", ",", "'momentum'", ":", "0.9", "}", ",", "kvstore", "=", "opt", ".", "kvstore", ")", "loss", "=", "MarginLoss", "(", "margin", "=", "opt", ".", "margin", ",", "nu", "=", "opt", ".", "nu", ")", "best_val", "=", "0.0", "for", "epoch", "in", "range", "(", "epochs", ")", ":", "tic", "=", "time", ".", "time", "(", ")", "prev_loss", ",", "cumulative_loss", "=", "0.0", ",", "0.0", "# Learning rate schedule.", "trainer", ".", "set_learning_rate", "(", "get_lr", "(", "opt", ".", "lr", ",", "epoch", ",", "steps", ",", "opt", ".", "factor", ")", ")", "logging", ".", "info", "(", "'Epoch %d learning rate=%f'", ",", "epoch", ",", "trainer", ".", "learning_rate", ")", "if", "opt", ".", "lr_beta", ">", "0.0", ":", "trainer_beta", ".", "set_learning_rate", "(", "get_lr", "(", "opt", ".", "lr_beta", ",", "epoch", ",", "steps", ",", "opt", ".", "factor", ")", ")", "logging", ".", "info", "(", "'Epoch %d beta learning rate=%f'", ",", "epoch", ",", "trainer_beta", ".", "learning_rate", ")", "# Inner training loop.", "for", "i", "in", "range", "(", "200", ")", ":", "batch", "=", "train_data", ".", "next", "(", ")", "data", "=", "gluon", ".", "utils", ".", "split_and_load", "(", "batch", ".", "data", "[", "0", "]", ",", "ctx_list", "=", "ctx", ",", "batch_axis", "=", "0", ")", "label", "=", "gluon", ".", "utils", ".", "split_and_load", "(", "batch", ".", "label", "[", "0", "]", ",", "ctx_list", "=", "ctx", ",", "batch_axis", "=", "0", ")", "Ls", "=", "[", "]", "with", "ag", ".", "record", "(", ")", ":", "for", "x", ",", "y", "in", "zip", "(", "data", ",", "label", ")", ":", "a_indices", ",", "anchors", ",", "positives", ",", "negatives", ",", "_", "=", "net", "(", "x", ")", "if", "opt", ".", "lr_beta", ">", "0.0", ":", "L", "=", "loss", "(", "anchors", ",", "positives", ",", "negatives", ",", "beta", ",", "y", "[", "a_indices", "]", ")", "else", ":", "L", "=", "loss", "(", "anchors", ",", "positives", ",", "negatives", ",", "opt", ".", "beta", ",", "None", ")", "# Store the loss and do backward after we have done forward", "# on all GPUs for better speed on multiple GPUs.", "Ls", ".", "append", "(", "L", ")", "cumulative_loss", "+=", "nd", ".", "mean", "(", "L", ")", ".", "asscalar", "(", ")", "for", "L", "in", "Ls", ":", "L", ".", "backward", "(", ")", "# Update.", "trainer", ".", "step", "(", "batch", ".", "data", "[", "0", "]", ".", "shape", "[", "0", "]", ")", "if", "opt", ".", "lr_beta", ">", "0.0", ":", "trainer_beta", ".", "step", "(", "batch", ".", "data", "[", "0", "]", ".", "shape", "[", "0", "]", ")", "if", "(", "i", "+", "1", ")", "%", "opt", ".", "log_interval", "==", "0", ":", "logging", ".", "info", "(", "'[Epoch %d, Iter %d] training loss=%f'", "%", "(", "epoch", ",", "i", "+", "1", ",", "cumulative_loss", "-", "prev_loss", ")", ")", "prev_loss", "=", "cumulative_loss", "logging", ".", "info", "(", "'[Epoch %d] training loss=%f'", "%", "(", "epoch", ",", "cumulative_loss", ")", ")", "logging", ".", "info", "(", "'[Epoch %d] time cost: %f'", "%", "(", "epoch", ",", "time", ".", "time", "(", ")", "-", "tic", ")", ")", "names", ",", "val_accs", "=", "test", "(", "ctx", ")", "for", "name", ",", "val_acc", "in", "zip", "(", "names", ",", "val_accs", ")", ":", "logging", ".", "info", "(", "'[Epoch %d] validation: %s=%f'", "%", "(", "epoch", ",", "name", ",", "val_acc", ")", ")", "if", "val_accs", "[", "0", "]", ">", "best_val", ":", "best_val", "=", "val_accs", "[", "0", "]", "logging", ".", "info", "(", "'Saving %s.'", "%", "opt", ".", "save_model_prefix", ")", "net", ".", "save_parameters", "(", "'%s.params'", "%", "opt", ".", "save_model_prefix", ")", "return", "best_val" ]
Training function.
[ "Training", "function", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/gluon/embedding_learning/train.py#L169-L250
train
apache/incubator-mxnet
example/ctc/lstm.py
_lstm_unroll_base
def _lstm_unroll_base(num_lstm_layer, seq_len, num_hidden): """ Returns symbol for LSTM model up to loss/softmax""" param_cells = [] last_states = [] for i in range(num_lstm_layer): param_cells.append(LSTMParam(i2h_weight=mx.sym.Variable("l%d_i2h_weight" % i), i2h_bias=mx.sym.Variable("l%d_i2h_bias" % i), h2h_weight=mx.sym.Variable("l%d_h2h_weight" % i), h2h_bias=mx.sym.Variable("l%d_h2h_bias" % i))) state = LSTMState(c=mx.sym.Variable("l%d_init_c" % i), h=mx.sym.Variable("l%d_init_h" % i)) last_states.append(state) assert len(last_states) == num_lstm_layer # embedding layer data = mx.sym.Variable('data') wordvec = mx.sym.SliceChannel(data=data, num_outputs=seq_len, squeeze_axis=1) hidden_all = [] for seqidx in range(seq_len): hidden = wordvec[seqidx] for i in range(num_lstm_layer): next_state = _lstm( num_hidden=num_hidden, indata=hidden, prev_state=last_states[i], param=param_cells[i], seqidx=seqidx, layeridx=i) hidden = next_state.h last_states[i] = next_state hidden_all.append(hidden) hidden_concat = mx.sym.Concat(*hidden_all, dim=0) pred_fc = mx.sym.FullyConnected(data=hidden_concat, num_hidden=11, name="pred_fc") return pred_fc
python
def _lstm_unroll_base(num_lstm_layer, seq_len, num_hidden): """ Returns symbol for LSTM model up to loss/softmax""" param_cells = [] last_states = [] for i in range(num_lstm_layer): param_cells.append(LSTMParam(i2h_weight=mx.sym.Variable("l%d_i2h_weight" % i), i2h_bias=mx.sym.Variable("l%d_i2h_bias" % i), h2h_weight=mx.sym.Variable("l%d_h2h_weight" % i), h2h_bias=mx.sym.Variable("l%d_h2h_bias" % i))) state = LSTMState(c=mx.sym.Variable("l%d_init_c" % i), h=mx.sym.Variable("l%d_init_h" % i)) last_states.append(state) assert len(last_states) == num_lstm_layer # embedding layer data = mx.sym.Variable('data') wordvec = mx.sym.SliceChannel(data=data, num_outputs=seq_len, squeeze_axis=1) hidden_all = [] for seqidx in range(seq_len): hidden = wordvec[seqidx] for i in range(num_lstm_layer): next_state = _lstm( num_hidden=num_hidden, indata=hidden, prev_state=last_states[i], param=param_cells[i], seqidx=seqidx, layeridx=i) hidden = next_state.h last_states[i] = next_state hidden_all.append(hidden) hidden_concat = mx.sym.Concat(*hidden_all, dim=0) pred_fc = mx.sym.FullyConnected(data=hidden_concat, num_hidden=11, name="pred_fc") return pred_fc
[ "def", "_lstm_unroll_base", "(", "num_lstm_layer", ",", "seq_len", ",", "num_hidden", ")", ":", "param_cells", "=", "[", "]", "last_states", "=", "[", "]", "for", "i", "in", "range", "(", "num_lstm_layer", ")", ":", "param_cells", ".", "append", "(", "LSTMParam", "(", "i2h_weight", "=", "mx", ".", "sym", ".", "Variable", "(", "\"l%d_i2h_weight\"", "%", "i", ")", ",", "i2h_bias", "=", "mx", ".", "sym", ".", "Variable", "(", "\"l%d_i2h_bias\"", "%", "i", ")", ",", "h2h_weight", "=", "mx", ".", "sym", ".", "Variable", "(", "\"l%d_h2h_weight\"", "%", "i", ")", ",", "h2h_bias", "=", "mx", ".", "sym", ".", "Variable", "(", "\"l%d_h2h_bias\"", "%", "i", ")", ")", ")", "state", "=", "LSTMState", "(", "c", "=", "mx", ".", "sym", ".", "Variable", "(", "\"l%d_init_c\"", "%", "i", ")", ",", "h", "=", "mx", ".", "sym", ".", "Variable", "(", "\"l%d_init_h\"", "%", "i", ")", ")", "last_states", ".", "append", "(", "state", ")", "assert", "len", "(", "last_states", ")", "==", "num_lstm_layer", "# embedding layer", "data", "=", "mx", ".", "sym", ".", "Variable", "(", "'data'", ")", "wordvec", "=", "mx", ".", "sym", ".", "SliceChannel", "(", "data", "=", "data", ",", "num_outputs", "=", "seq_len", ",", "squeeze_axis", "=", "1", ")", "hidden_all", "=", "[", "]", "for", "seqidx", "in", "range", "(", "seq_len", ")", ":", "hidden", "=", "wordvec", "[", "seqidx", "]", "for", "i", "in", "range", "(", "num_lstm_layer", ")", ":", "next_state", "=", "_lstm", "(", "num_hidden", "=", "num_hidden", ",", "indata", "=", "hidden", ",", "prev_state", "=", "last_states", "[", "i", "]", ",", "param", "=", "param_cells", "[", "i", "]", ",", "seqidx", "=", "seqidx", ",", "layeridx", "=", "i", ")", "hidden", "=", "next_state", ".", "h", "last_states", "[", "i", "]", "=", "next_state", "hidden_all", ".", "append", "(", "hidden", ")", "hidden_concat", "=", "mx", ".", "sym", ".", "Concat", "(", "*", "hidden_all", ",", "dim", "=", "0", ")", "pred_fc", "=", "mx", ".", "sym", ".", "FullyConnected", "(", "data", "=", "hidden_concat", ",", "num_hidden", "=", "11", ",", "name", "=", "\"pred_fc\"", ")", "return", "pred_fc" ]
Returns symbol for LSTM model up to loss/softmax
[ "Returns", "symbol", "for", "LSTM", "model", "up", "to", "loss", "/", "softmax" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ctc/lstm.py#L58-L93
train
apache/incubator-mxnet
example/ctc/lstm.py
_add_warp_ctc_loss
def _add_warp_ctc_loss(pred, seq_len, num_label, label): """ Adds Symbol.contrib.ctc_loss on top of pred symbol and returns the resulting symbol """ label = mx.sym.Reshape(data=label, shape=(-1,)) label = mx.sym.Cast(data=label, dtype='int32') return mx.sym.WarpCTC(data=pred, label=label, label_length=num_label, input_length=seq_len)
python
def _add_warp_ctc_loss(pred, seq_len, num_label, label): """ Adds Symbol.contrib.ctc_loss on top of pred symbol and returns the resulting symbol """ label = mx.sym.Reshape(data=label, shape=(-1,)) label = mx.sym.Cast(data=label, dtype='int32') return mx.sym.WarpCTC(data=pred, label=label, label_length=num_label, input_length=seq_len)
[ "def", "_add_warp_ctc_loss", "(", "pred", ",", "seq_len", ",", "num_label", ",", "label", ")", ":", "label", "=", "mx", ".", "sym", ".", "Reshape", "(", "data", "=", "label", ",", "shape", "=", "(", "-", "1", ",", ")", ")", "label", "=", "mx", ".", "sym", ".", "Cast", "(", "data", "=", "label", ",", "dtype", "=", "'int32'", ")", "return", "mx", ".", "sym", ".", "WarpCTC", "(", "data", "=", "pred", ",", "label", "=", "label", ",", "label_length", "=", "num_label", ",", "input_length", "=", "seq_len", ")" ]
Adds Symbol.contrib.ctc_loss on top of pred symbol and returns the resulting symbol
[ "Adds", "Symbol", ".", "contrib", ".", "ctc_loss", "on", "top", "of", "pred", "symbol", "and", "returns", "the", "resulting", "symbol" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ctc/lstm.py#L96-L100
train
apache/incubator-mxnet
example/ctc/lstm.py
_add_mxnet_ctc_loss
def _add_mxnet_ctc_loss(pred, seq_len, label): """ Adds Symbol.WapCTC on top of pred symbol and returns the resulting symbol """ pred_ctc = mx.sym.Reshape(data=pred, shape=(-4, seq_len, -1, 0)) loss = mx.sym.contrib.ctc_loss(data=pred_ctc, label=label) ctc_loss = mx.sym.MakeLoss(loss) softmax_class = mx.symbol.SoftmaxActivation(data=pred) softmax_loss = mx.sym.MakeLoss(softmax_class) softmax_loss = mx.sym.BlockGrad(softmax_loss) return mx.sym.Group([softmax_loss, ctc_loss])
python
def _add_mxnet_ctc_loss(pred, seq_len, label): """ Adds Symbol.WapCTC on top of pred symbol and returns the resulting symbol """ pred_ctc = mx.sym.Reshape(data=pred, shape=(-4, seq_len, -1, 0)) loss = mx.sym.contrib.ctc_loss(data=pred_ctc, label=label) ctc_loss = mx.sym.MakeLoss(loss) softmax_class = mx.symbol.SoftmaxActivation(data=pred) softmax_loss = mx.sym.MakeLoss(softmax_class) softmax_loss = mx.sym.BlockGrad(softmax_loss) return mx.sym.Group([softmax_loss, ctc_loss])
[ "def", "_add_mxnet_ctc_loss", "(", "pred", ",", "seq_len", ",", "label", ")", ":", "pred_ctc", "=", "mx", ".", "sym", ".", "Reshape", "(", "data", "=", "pred", ",", "shape", "=", "(", "-", "4", ",", "seq_len", ",", "-", "1", ",", "0", ")", ")", "loss", "=", "mx", ".", "sym", ".", "contrib", ".", "ctc_loss", "(", "data", "=", "pred_ctc", ",", "label", "=", "label", ")", "ctc_loss", "=", "mx", ".", "sym", ".", "MakeLoss", "(", "loss", ")", "softmax_class", "=", "mx", ".", "symbol", ".", "SoftmaxActivation", "(", "data", "=", "pred", ")", "softmax_loss", "=", "mx", ".", "sym", ".", "MakeLoss", "(", "softmax_class", ")", "softmax_loss", "=", "mx", ".", "sym", ".", "BlockGrad", "(", "softmax_loss", ")", "return", "mx", ".", "sym", ".", "Group", "(", "[", "softmax_loss", ",", "ctc_loss", "]", ")" ]
Adds Symbol.WapCTC on top of pred symbol and returns the resulting symbol
[ "Adds", "Symbol", ".", "WapCTC", "on", "top", "of", "pred", "symbol", "and", "returns", "the", "resulting", "symbol" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ctc/lstm.py#L103-L113
train
apache/incubator-mxnet
example/ctc/lstm.py
_add_ctc_loss
def _add_ctc_loss(pred, seq_len, num_label, loss_type): """ Adds CTC loss on top of pred symbol and returns the resulting symbol """ label = mx.sym.Variable('label') if loss_type == 'warpctc': print("Using WarpCTC Loss") sm = _add_warp_ctc_loss(pred, seq_len, num_label, label) else: print("Using MXNet CTC Loss") assert loss_type == 'ctc' sm = _add_mxnet_ctc_loss(pred, seq_len, label) return sm
python
def _add_ctc_loss(pred, seq_len, num_label, loss_type): """ Adds CTC loss on top of pred symbol and returns the resulting symbol """ label = mx.sym.Variable('label') if loss_type == 'warpctc': print("Using WarpCTC Loss") sm = _add_warp_ctc_loss(pred, seq_len, num_label, label) else: print("Using MXNet CTC Loss") assert loss_type == 'ctc' sm = _add_mxnet_ctc_loss(pred, seq_len, label) return sm
[ "def", "_add_ctc_loss", "(", "pred", ",", "seq_len", ",", "num_label", ",", "loss_type", ")", ":", "label", "=", "mx", ".", "sym", ".", "Variable", "(", "'label'", ")", "if", "loss_type", "==", "'warpctc'", ":", "print", "(", "\"Using WarpCTC Loss\"", ")", "sm", "=", "_add_warp_ctc_loss", "(", "pred", ",", "seq_len", ",", "num_label", ",", "label", ")", "else", ":", "print", "(", "\"Using MXNet CTC Loss\"", ")", "assert", "loss_type", "==", "'ctc'", "sm", "=", "_add_mxnet_ctc_loss", "(", "pred", ",", "seq_len", ",", "label", ")", "return", "sm" ]
Adds CTC loss on top of pred symbol and returns the resulting symbol
[ "Adds", "CTC", "loss", "on", "top", "of", "pred", "symbol", "and", "returns", "the", "resulting", "symbol" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ctc/lstm.py#L116-L126
train
apache/incubator-mxnet
example/ctc/lstm.py
lstm_unroll
def lstm_unroll(num_lstm_layer, seq_len, num_hidden, num_label, loss_type=None): """ Creates an unrolled LSTM symbol for inference if loss_type is not specified, and for training if loss_type is specified. loss_type must be one of 'ctc' or 'warpctc' Parameters ---------- num_lstm_layer: int seq_len: int num_hidden: int num_label: int loss_type: str 'ctc' or 'warpctc' Returns ------- mxnet.symbol.symbol.Symbol """ # Create the base (shared between training and inference) and add loss to the end pred = _lstm_unroll_base(num_lstm_layer, seq_len, num_hidden) if loss_type: # Training mode, add loss return _add_ctc_loss(pred, seq_len, num_label, loss_type) else: # Inference mode, add softmax return mx.sym.softmax(data=pred, name='softmax')
python
def lstm_unroll(num_lstm_layer, seq_len, num_hidden, num_label, loss_type=None): """ Creates an unrolled LSTM symbol for inference if loss_type is not specified, and for training if loss_type is specified. loss_type must be one of 'ctc' or 'warpctc' Parameters ---------- num_lstm_layer: int seq_len: int num_hidden: int num_label: int loss_type: str 'ctc' or 'warpctc' Returns ------- mxnet.symbol.symbol.Symbol """ # Create the base (shared between training and inference) and add loss to the end pred = _lstm_unroll_base(num_lstm_layer, seq_len, num_hidden) if loss_type: # Training mode, add loss return _add_ctc_loss(pred, seq_len, num_label, loss_type) else: # Inference mode, add softmax return mx.sym.softmax(data=pred, name='softmax')
[ "def", "lstm_unroll", "(", "num_lstm_layer", ",", "seq_len", ",", "num_hidden", ",", "num_label", ",", "loss_type", "=", "None", ")", ":", "# Create the base (shared between training and inference) and add loss to the end", "pred", "=", "_lstm_unroll_base", "(", "num_lstm_layer", ",", "seq_len", ",", "num_hidden", ")", "if", "loss_type", ":", "# Training mode, add loss", "return", "_add_ctc_loss", "(", "pred", ",", "seq_len", ",", "num_label", ",", "loss_type", ")", "else", ":", "# Inference mode, add softmax", "return", "mx", ".", "sym", ".", "softmax", "(", "data", "=", "pred", ",", "name", "=", "'softmax'", ")" ]
Creates an unrolled LSTM symbol for inference if loss_type is not specified, and for training if loss_type is specified. loss_type must be one of 'ctc' or 'warpctc' Parameters ---------- num_lstm_layer: int seq_len: int num_hidden: int num_label: int loss_type: str 'ctc' or 'warpctc' Returns ------- mxnet.symbol.symbol.Symbol
[ "Creates", "an", "unrolled", "LSTM", "symbol", "for", "inference", "if", "loss_type", "is", "not", "specified", "and", "for", "training", "if", "loss_type", "is", "specified", ".", "loss_type", "must", "be", "one", "of", "ctc", "or", "warpctc" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ctc/lstm.py#L129-L155
train
apache/incubator-mxnet
example/ctc/lstm.py
init_states
def init_states(batch_size, num_lstm_layer, num_hidden): """ Returns name and shape of init states of LSTM network Parameters ---------- batch_size: list of tuple of str and tuple of int and int num_lstm_layer: int num_hidden: int Returns ------- list of tuple of str and tuple of int and int """ init_c = [('l%d_init_c' % l, (batch_size, num_hidden)) for l in range(num_lstm_layer)] init_h = [('l%d_init_h' % l, (batch_size, num_hidden)) for l in range(num_lstm_layer)] return init_c + init_h
python
def init_states(batch_size, num_lstm_layer, num_hidden): """ Returns name and shape of init states of LSTM network Parameters ---------- batch_size: list of tuple of str and tuple of int and int num_lstm_layer: int num_hidden: int Returns ------- list of tuple of str and tuple of int and int """ init_c = [('l%d_init_c' % l, (batch_size, num_hidden)) for l in range(num_lstm_layer)] init_h = [('l%d_init_h' % l, (batch_size, num_hidden)) for l in range(num_lstm_layer)] return init_c + init_h
[ "def", "init_states", "(", "batch_size", ",", "num_lstm_layer", ",", "num_hidden", ")", ":", "init_c", "=", "[", "(", "'l%d_init_c'", "%", "l", ",", "(", "batch_size", ",", "num_hidden", ")", ")", "for", "l", "in", "range", "(", "num_lstm_layer", ")", "]", "init_h", "=", "[", "(", "'l%d_init_h'", "%", "l", ",", "(", "batch_size", ",", "num_hidden", ")", ")", "for", "l", "in", "range", "(", "num_lstm_layer", ")", "]", "return", "init_c", "+", "init_h" ]
Returns name and shape of init states of LSTM network Parameters ---------- batch_size: list of tuple of str and tuple of int and int num_lstm_layer: int num_hidden: int Returns ------- list of tuple of str and tuple of int and int
[ "Returns", "name", "and", "shape", "of", "init", "states", "of", "LSTM", "network" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ctc/lstm.py#L158-L174
train
apache/incubator-mxnet
python/mxnet/_ctypes/ndarray.py
_imperative_invoke
def _imperative_invoke(handle, ndargs, keys, vals, out): """ctypes implementation of imperative invoke wrapper""" if out is not None: original_output = out if isinstance(out, NDArrayBase): out = (out,) num_output = ctypes.c_int(len(out)) output_vars = c_handle_array(out) output_vars = ctypes.cast(output_vars, ctypes.POINTER(NDArrayHandle)) else: original_output = None output_vars = ctypes.POINTER(NDArrayHandle)() num_output = ctypes.c_int(0) # return output stypes to avoid the c_api call for checking # a handle's stype in _ndarray_cls out_stypes = ctypes.POINTER(ctypes.c_int)() check_call(_LIB.MXImperativeInvokeEx( ctypes.c_void_p(handle), ctypes.c_int(len(ndargs)), c_handle_array(ndargs), ctypes.byref(num_output), ctypes.byref(output_vars), ctypes.c_int(len(keys)), c_str_array(keys), c_str_array([str(s) for s in vals]), ctypes.byref(out_stypes))) if original_output is not None: return original_output if num_output.value == 1: return _ndarray_cls(ctypes.cast(output_vars[0], NDArrayHandle), stype=out_stypes[0]) else: return [_ndarray_cls(ctypes.cast(output_vars[i], NDArrayHandle), stype=out_stypes[i]) for i in range(num_output.value)]
python
def _imperative_invoke(handle, ndargs, keys, vals, out): """ctypes implementation of imperative invoke wrapper""" if out is not None: original_output = out if isinstance(out, NDArrayBase): out = (out,) num_output = ctypes.c_int(len(out)) output_vars = c_handle_array(out) output_vars = ctypes.cast(output_vars, ctypes.POINTER(NDArrayHandle)) else: original_output = None output_vars = ctypes.POINTER(NDArrayHandle)() num_output = ctypes.c_int(0) # return output stypes to avoid the c_api call for checking # a handle's stype in _ndarray_cls out_stypes = ctypes.POINTER(ctypes.c_int)() check_call(_LIB.MXImperativeInvokeEx( ctypes.c_void_p(handle), ctypes.c_int(len(ndargs)), c_handle_array(ndargs), ctypes.byref(num_output), ctypes.byref(output_vars), ctypes.c_int(len(keys)), c_str_array(keys), c_str_array([str(s) for s in vals]), ctypes.byref(out_stypes))) if original_output is not None: return original_output if num_output.value == 1: return _ndarray_cls(ctypes.cast(output_vars[0], NDArrayHandle), stype=out_stypes[0]) else: return [_ndarray_cls(ctypes.cast(output_vars[i], NDArrayHandle), stype=out_stypes[i]) for i in range(num_output.value)]
[ "def", "_imperative_invoke", "(", "handle", ",", "ndargs", ",", "keys", ",", "vals", ",", "out", ")", ":", "if", "out", "is", "not", "None", ":", "original_output", "=", "out", "if", "isinstance", "(", "out", ",", "NDArrayBase", ")", ":", "out", "=", "(", "out", ",", ")", "num_output", "=", "ctypes", ".", "c_int", "(", "len", "(", "out", ")", ")", "output_vars", "=", "c_handle_array", "(", "out", ")", "output_vars", "=", "ctypes", ".", "cast", "(", "output_vars", ",", "ctypes", ".", "POINTER", "(", "NDArrayHandle", ")", ")", "else", ":", "original_output", "=", "None", "output_vars", "=", "ctypes", ".", "POINTER", "(", "NDArrayHandle", ")", "(", ")", "num_output", "=", "ctypes", ".", "c_int", "(", "0", ")", "# return output stypes to avoid the c_api call for checking", "# a handle's stype in _ndarray_cls", "out_stypes", "=", "ctypes", ".", "POINTER", "(", "ctypes", ".", "c_int", ")", "(", ")", "check_call", "(", "_LIB", ".", "MXImperativeInvokeEx", "(", "ctypes", ".", "c_void_p", "(", "handle", ")", ",", "ctypes", ".", "c_int", "(", "len", "(", "ndargs", ")", ")", ",", "c_handle_array", "(", "ndargs", ")", ",", "ctypes", ".", "byref", "(", "num_output", ")", ",", "ctypes", ".", "byref", "(", "output_vars", ")", ",", "ctypes", ".", "c_int", "(", "len", "(", "keys", ")", ")", ",", "c_str_array", "(", "keys", ")", ",", "c_str_array", "(", "[", "str", "(", "s", ")", "for", "s", "in", "vals", "]", ")", ",", "ctypes", ".", "byref", "(", "out_stypes", ")", ")", ")", "if", "original_output", "is", "not", "None", ":", "return", "original_output", "if", "num_output", ".", "value", "==", "1", ":", "return", "_ndarray_cls", "(", "ctypes", ".", "cast", "(", "output_vars", "[", "0", "]", ",", "NDArrayHandle", ")", ",", "stype", "=", "out_stypes", "[", "0", "]", ")", "else", ":", "return", "[", "_ndarray_cls", "(", "ctypes", ".", "cast", "(", "output_vars", "[", "i", "]", ",", "NDArrayHandle", ")", ",", "stype", "=", "out_stypes", "[", "i", "]", ")", "for", "i", "in", "range", "(", "num_output", ".", "value", ")", "]" ]
ctypes implementation of imperative invoke wrapper
[ "ctypes", "implementation", "of", "imperative", "invoke", "wrapper" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/_ctypes/ndarray.py#L65-L102
train
apache/incubator-mxnet
python/mxnet/contrib/autograd.py
set_is_training
def set_is_training(is_train): """Set status to training/not training. When training, graph will be constructed for gradient computation. Operators will also run with ctx.is_train=True. For example, Dropout will drop inputs randomly when is_train=True while simply passing through if is_train=False. Parameters ---------- is_train: bool Returns ------- previous state before this set. """ prev = ctypes.c_int() check_call(_LIB.MXAutogradSetIsTraining( ctypes.c_int(is_train), ctypes.byref(prev))) check_call(_LIB.MXAutogradSetIsRecording( ctypes.c_int(is_train), ctypes.byref(prev))) return bool(prev.value)
python
def set_is_training(is_train): """Set status to training/not training. When training, graph will be constructed for gradient computation. Operators will also run with ctx.is_train=True. For example, Dropout will drop inputs randomly when is_train=True while simply passing through if is_train=False. Parameters ---------- is_train: bool Returns ------- previous state before this set. """ prev = ctypes.c_int() check_call(_LIB.MXAutogradSetIsTraining( ctypes.c_int(is_train), ctypes.byref(prev))) check_call(_LIB.MXAutogradSetIsRecording( ctypes.c_int(is_train), ctypes.byref(prev))) return bool(prev.value)
[ "def", "set_is_training", "(", "is_train", ")", ":", "prev", "=", "ctypes", ".", "c_int", "(", ")", "check_call", "(", "_LIB", ".", "MXAutogradSetIsTraining", "(", "ctypes", ".", "c_int", "(", "is_train", ")", ",", "ctypes", ".", "byref", "(", "prev", ")", ")", ")", "check_call", "(", "_LIB", ".", "MXAutogradSetIsRecording", "(", "ctypes", ".", "c_int", "(", "is_train", ")", ",", "ctypes", ".", "byref", "(", "prev", ")", ")", ")", "return", "bool", "(", "prev", ".", "value", ")" ]
Set status to training/not training. When training, graph will be constructed for gradient computation. Operators will also run with ctx.is_train=True. For example, Dropout will drop inputs randomly when is_train=True while simply passing through if is_train=False. Parameters ---------- is_train: bool Returns ------- previous state before this set.
[ "Set", "status", "to", "training", "/", "not", "training", ".", "When", "training", "graph", "will", "be", "constructed", "for", "gradient", "computation", ".", "Operators", "will", "also", "run", "with", "ctx", ".", "is_train", "=", "True", ".", "For", "example", "Dropout", "will", "drop", "inputs", "randomly", "when", "is_train", "=", "True", "while", "simply", "passing", "through", "if", "is_train", "=", "False", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/contrib/autograd.py#L32-L51
train
apache/incubator-mxnet
python/mxnet/contrib/autograd.py
backward
def backward(outputs, out_grads=None, retain_graph=False): """Compute the gradients of outputs w.r.t variables. Parameters ---------- outputs: list of NDArray out_grads: list of NDArray or None """ assert isinstance(outputs, (list, tuple)), \ "outputs must be a list or tuple of NDArrays" if out_grads is None: check_call(_LIB.MXAutogradBackward( len(outputs), c_handle_array(outputs), ctypes.c_void_p(0), ctypes.c_int(retain_graph))) return ograd_handles = [] for arr in out_grads: if arr is not None: ograd_handles.append(arr.handle) else: ograd_handles.append(NDArrayHandle(0)) assert len(ograd_handles) == len(outputs), \ "outputs and out_grads must have the same length" check_call(_LIB.MXAutogradBackward( len(outputs), c_handle_array(outputs), c_array(NDArrayHandle, ograd_handles), ctypes.c_int(retain_graph)))
python
def backward(outputs, out_grads=None, retain_graph=False): """Compute the gradients of outputs w.r.t variables. Parameters ---------- outputs: list of NDArray out_grads: list of NDArray or None """ assert isinstance(outputs, (list, tuple)), \ "outputs must be a list or tuple of NDArrays" if out_grads is None: check_call(_LIB.MXAutogradBackward( len(outputs), c_handle_array(outputs), ctypes.c_void_p(0), ctypes.c_int(retain_graph))) return ograd_handles = [] for arr in out_grads: if arr is not None: ograd_handles.append(arr.handle) else: ograd_handles.append(NDArrayHandle(0)) assert len(ograd_handles) == len(outputs), \ "outputs and out_grads must have the same length" check_call(_LIB.MXAutogradBackward( len(outputs), c_handle_array(outputs), c_array(NDArrayHandle, ograd_handles), ctypes.c_int(retain_graph)))
[ "def", "backward", "(", "outputs", ",", "out_grads", "=", "None", ",", "retain_graph", "=", "False", ")", ":", "assert", "isinstance", "(", "outputs", ",", "(", "list", ",", "tuple", ")", ")", ",", "\"outputs must be a list or tuple of NDArrays\"", "if", "out_grads", "is", "None", ":", "check_call", "(", "_LIB", ".", "MXAutogradBackward", "(", "len", "(", "outputs", ")", ",", "c_handle_array", "(", "outputs", ")", ",", "ctypes", ".", "c_void_p", "(", "0", ")", ",", "ctypes", ".", "c_int", "(", "retain_graph", ")", ")", ")", "return", "ograd_handles", "=", "[", "]", "for", "arr", "in", "out_grads", ":", "if", "arr", "is", "not", "None", ":", "ograd_handles", ".", "append", "(", "arr", ".", "handle", ")", "else", ":", "ograd_handles", ".", "append", "(", "NDArrayHandle", "(", "0", ")", ")", "assert", "len", "(", "ograd_handles", ")", "==", "len", "(", "outputs", ")", ",", "\"outputs and out_grads must have the same length\"", "check_call", "(", "_LIB", ".", "MXAutogradBackward", "(", "len", "(", "outputs", ")", ",", "c_handle_array", "(", "outputs", ")", ",", "c_array", "(", "NDArrayHandle", ",", "ograd_handles", ")", ",", "ctypes", ".", "c_int", "(", "retain_graph", ")", ")", ")" ]
Compute the gradients of outputs w.r.t variables. Parameters ---------- outputs: list of NDArray out_grads: list of NDArray or None
[ "Compute", "the", "gradients", "of", "outputs", "w", ".", "r", ".", "t", "variables", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/contrib/autograd.py#L123-L155
train
apache/incubator-mxnet
python/mxnet/contrib/autograd.py
grad_and_loss
def grad_and_loss(func, argnum=None): """Return function that computes both gradient of arguments and loss value. Parameters ---------- func: a python function The forward (loss) function. argnum: an int or a list of int The index of argument to calculate gradient for. Returns ------- grad_and_loss_func: a python function A function that would compute both the gradient of arguments and loss value. """ @functools.wraps(func) def wrapped(*args): """Wrapped function.""" variables = args if argnum is not None: argnum_ = argnum if isinstance(argnum, list) else [argnum] variables = [args[i] for i in argnum_] for x in variables: assert isinstance(x, NDArray), "type of autograd input should NDArray." grads = [zeros_like(x) for x in variables] mark_variables(variables, grads) with train_section(): outputs = func(*args) compute_gradient([outputs] if isinstance(outputs, NDArray) else outputs) return grads, outputs return wrapped
python
def grad_and_loss(func, argnum=None): """Return function that computes both gradient of arguments and loss value. Parameters ---------- func: a python function The forward (loss) function. argnum: an int or a list of int The index of argument to calculate gradient for. Returns ------- grad_and_loss_func: a python function A function that would compute both the gradient of arguments and loss value. """ @functools.wraps(func) def wrapped(*args): """Wrapped function.""" variables = args if argnum is not None: argnum_ = argnum if isinstance(argnum, list) else [argnum] variables = [args[i] for i in argnum_] for x in variables: assert isinstance(x, NDArray), "type of autograd input should NDArray." grads = [zeros_like(x) for x in variables] mark_variables(variables, grads) with train_section(): outputs = func(*args) compute_gradient([outputs] if isinstance(outputs, NDArray) else outputs) return grads, outputs return wrapped
[ "def", "grad_and_loss", "(", "func", ",", "argnum", "=", "None", ")", ":", "@", "functools", ".", "wraps", "(", "func", ")", "def", "wrapped", "(", "*", "args", ")", ":", "\"\"\"Wrapped function.\"\"\"", "variables", "=", "args", "if", "argnum", "is", "not", "None", ":", "argnum_", "=", "argnum", "if", "isinstance", "(", "argnum", ",", "list", ")", "else", "[", "argnum", "]", "variables", "=", "[", "args", "[", "i", "]", "for", "i", "in", "argnum_", "]", "for", "x", "in", "variables", ":", "assert", "isinstance", "(", "x", ",", "NDArray", ")", ",", "\"type of autograd input should NDArray.\"", "grads", "=", "[", "zeros_like", "(", "x", ")", "for", "x", "in", "variables", "]", "mark_variables", "(", "variables", ",", "grads", ")", "with", "train_section", "(", ")", ":", "outputs", "=", "func", "(", "*", "args", ")", "compute_gradient", "(", "[", "outputs", "]", "if", "isinstance", "(", "outputs", ",", "NDArray", ")", "else", "outputs", ")", "return", "grads", ",", "outputs", "return", "wrapped" ]
Return function that computes both gradient of arguments and loss value. Parameters ---------- func: a python function The forward (loss) function. argnum: an int or a list of int The index of argument to calculate gradient for. Returns ------- grad_and_loss_func: a python function A function that would compute both the gradient of arguments and loss value.
[ "Return", "function", "that", "computes", "both", "gradient", "of", "arguments", "and", "loss", "value", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/contrib/autograd.py#L163-L193
train
apache/incubator-mxnet
python/mxnet/contrib/autograd.py
grad
def grad(func, argnum=None): """Return function that computes gradient of arguments. Parameters ---------- func: a python function The forward (loss) function. argnum: an int or a list of int The index of argument to calculate gradient for. Returns ------- grad_func: a python function A function that would compute the gradient of arguments. Examples -------- >>> # autograd supports dynamic graph which is changed >>> # every instance >>> def func(x): >>> r = random.randint(0, 1) >>> if r % 2: >>> return x**2 >>> else: >>> return x/3 >>> # use `grad(func)` to get the gradient function >>> for x in range(10): >>> grad_func = grad(func) >>> inputs = nd.array([[1, 2, 3], [4, 5, 6]]) >>> grad_vals = grad_func(inputs) """ grad_with_loss_func = grad_and_loss(func, argnum) @functools.wraps(grad_with_loss_func) def wrapped(*args): return grad_with_loss_func(*args)[0] return wrapped
python
def grad(func, argnum=None): """Return function that computes gradient of arguments. Parameters ---------- func: a python function The forward (loss) function. argnum: an int or a list of int The index of argument to calculate gradient for. Returns ------- grad_func: a python function A function that would compute the gradient of arguments. Examples -------- >>> # autograd supports dynamic graph which is changed >>> # every instance >>> def func(x): >>> r = random.randint(0, 1) >>> if r % 2: >>> return x**2 >>> else: >>> return x/3 >>> # use `grad(func)` to get the gradient function >>> for x in range(10): >>> grad_func = grad(func) >>> inputs = nd.array([[1, 2, 3], [4, 5, 6]]) >>> grad_vals = grad_func(inputs) """ grad_with_loss_func = grad_and_loss(func, argnum) @functools.wraps(grad_with_loss_func) def wrapped(*args): return grad_with_loss_func(*args)[0] return wrapped
[ "def", "grad", "(", "func", ",", "argnum", "=", "None", ")", ":", "grad_with_loss_func", "=", "grad_and_loss", "(", "func", ",", "argnum", ")", "@", "functools", ".", "wraps", "(", "grad_with_loss_func", ")", "def", "wrapped", "(", "*", "args", ")", ":", "return", "grad_with_loss_func", "(", "*", "args", ")", "[", "0", "]", "return", "wrapped" ]
Return function that computes gradient of arguments. Parameters ---------- func: a python function The forward (loss) function. argnum: an int or a list of int The index of argument to calculate gradient for. Returns ------- grad_func: a python function A function that would compute the gradient of arguments. Examples -------- >>> # autograd supports dynamic graph which is changed >>> # every instance >>> def func(x): >>> r = random.randint(0, 1) >>> if r % 2: >>> return x**2 >>> else: >>> return x/3 >>> # use `grad(func)` to get the gradient function >>> for x in range(10): >>> grad_func = grad(func) >>> inputs = nd.array([[1, 2, 3], [4, 5, 6]]) >>> grad_vals = grad_func(inputs)
[ "Return", "function", "that", "computes", "gradient", "of", "arguments", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/contrib/autograd.py#L195-L230
train
apache/incubator-mxnet
python/mxnet/gluon/utils.py
split_data
def split_data(data, num_slice, batch_axis=0, even_split=True): """Splits an NDArray into `num_slice` slices along `batch_axis`. Usually used for data parallelism where each slices is sent to one device (i.e. GPU). Parameters ---------- data : NDArray A batch of data. num_slice : int Number of desired slices. batch_axis : int, default 0 The axis along which to slice. even_split : bool, default True Whether to force all slices to have the same number of elements. If `True`, an error will be raised when `num_slice` does not evenly divide `data.shape[batch_axis]`. Returns ------- list of NDArray Return value is a list even if `num_slice` is 1. """ size = data.shape[batch_axis] if even_split and size % num_slice != 0: raise ValueError( "data with shape %s cannot be evenly split into %d slices along axis %d. " \ "Use a batch size that's multiple of %d or set even_split=False to allow " \ "uneven partitioning of data."%( str(data.shape), num_slice, batch_axis, num_slice)) step = size // num_slice # If size < num_slice, make fewer slices if not even_split and size < num_slice: step = 1 num_slice = size if batch_axis == 0: slices = [data[i*step:(i+1)*step] if i < num_slice - 1 else data[i*step:size] for i in range(num_slice)] elif even_split: slices = ndarray.split(data, num_outputs=num_slice, axis=batch_axis) else: slices = [ndarray.slice_axis(data, batch_axis, i*step, (i+1)*step) if i < num_slice - 1 else ndarray.slice_axis(data, batch_axis, i*step, size) for i in range(num_slice)] return slices
python
def split_data(data, num_slice, batch_axis=0, even_split=True): """Splits an NDArray into `num_slice` slices along `batch_axis`. Usually used for data parallelism where each slices is sent to one device (i.e. GPU). Parameters ---------- data : NDArray A batch of data. num_slice : int Number of desired slices. batch_axis : int, default 0 The axis along which to slice. even_split : bool, default True Whether to force all slices to have the same number of elements. If `True`, an error will be raised when `num_slice` does not evenly divide `data.shape[batch_axis]`. Returns ------- list of NDArray Return value is a list even if `num_slice` is 1. """ size = data.shape[batch_axis] if even_split and size % num_slice != 0: raise ValueError( "data with shape %s cannot be evenly split into %d slices along axis %d. " \ "Use a batch size that's multiple of %d or set even_split=False to allow " \ "uneven partitioning of data."%( str(data.shape), num_slice, batch_axis, num_slice)) step = size // num_slice # If size < num_slice, make fewer slices if not even_split and size < num_slice: step = 1 num_slice = size if batch_axis == 0: slices = [data[i*step:(i+1)*step] if i < num_slice - 1 else data[i*step:size] for i in range(num_slice)] elif even_split: slices = ndarray.split(data, num_outputs=num_slice, axis=batch_axis) else: slices = [ndarray.slice_axis(data, batch_axis, i*step, (i+1)*step) if i < num_slice - 1 else ndarray.slice_axis(data, batch_axis, i*step, size) for i in range(num_slice)] return slices
[ "def", "split_data", "(", "data", ",", "num_slice", ",", "batch_axis", "=", "0", ",", "even_split", "=", "True", ")", ":", "size", "=", "data", ".", "shape", "[", "batch_axis", "]", "if", "even_split", "and", "size", "%", "num_slice", "!=", "0", ":", "raise", "ValueError", "(", "\"data with shape %s cannot be evenly split into %d slices along axis %d. \"", "\"Use a batch size that's multiple of %d or set even_split=False to allow \"", "\"uneven partitioning of data.\"", "%", "(", "str", "(", "data", ".", "shape", ")", ",", "num_slice", ",", "batch_axis", ",", "num_slice", ")", ")", "step", "=", "size", "//", "num_slice", "# If size < num_slice, make fewer slices", "if", "not", "even_split", "and", "size", "<", "num_slice", ":", "step", "=", "1", "num_slice", "=", "size", "if", "batch_axis", "==", "0", ":", "slices", "=", "[", "data", "[", "i", "*", "step", ":", "(", "i", "+", "1", ")", "*", "step", "]", "if", "i", "<", "num_slice", "-", "1", "else", "data", "[", "i", "*", "step", ":", "size", "]", "for", "i", "in", "range", "(", "num_slice", ")", "]", "elif", "even_split", ":", "slices", "=", "ndarray", ".", "split", "(", "data", ",", "num_outputs", "=", "num_slice", ",", "axis", "=", "batch_axis", ")", "else", ":", "slices", "=", "[", "ndarray", ".", "slice_axis", "(", "data", ",", "batch_axis", ",", "i", "*", "step", ",", "(", "i", "+", "1", ")", "*", "step", ")", "if", "i", "<", "num_slice", "-", "1", "else", "ndarray", ".", "slice_axis", "(", "data", ",", "batch_axis", ",", "i", "*", "step", ",", "size", ")", "for", "i", "in", "range", "(", "num_slice", ")", "]", "return", "slices" ]
Splits an NDArray into `num_slice` slices along `batch_axis`. Usually used for data parallelism where each slices is sent to one device (i.e. GPU). Parameters ---------- data : NDArray A batch of data. num_slice : int Number of desired slices. batch_axis : int, default 0 The axis along which to slice. even_split : bool, default True Whether to force all slices to have the same number of elements. If `True`, an error will be raised when `num_slice` does not evenly divide `data.shape[batch_axis]`. Returns ------- list of NDArray Return value is a list even if `num_slice` is 1.
[ "Splits", "an", "NDArray", "into", "num_slice", "slices", "along", "batch_axis", ".", "Usually", "used", "for", "data", "parallelism", "where", "each", "slices", "is", "sent", "to", "one", "device", "(", "i", ".", "e", ".", "GPU", ")", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/utils.py#L42-L90
train
apache/incubator-mxnet
python/mxnet/gluon/utils.py
split_and_load
def split_and_load(data, ctx_list, batch_axis=0, even_split=True): """Splits an NDArray into `len(ctx_list)` slices along `batch_axis` and loads each slice to one context in `ctx_list`. Parameters ---------- data : NDArray A batch of data. ctx_list : list of Context A list of Contexts. batch_axis : int, default 0 The axis along which to slice. even_split : bool, default True Whether to force all slices to have the same number of elements. Returns ------- list of NDArray Each corresponds to a context in `ctx_list`. """ if not isinstance(data, ndarray.NDArray): data = ndarray.array(data, ctx=ctx_list[0]) if len(ctx_list) == 1: return [data.as_in_context(ctx_list[0])] slices = split_data(data, len(ctx_list), batch_axis, even_split) return [i.as_in_context(ctx) for i, ctx in zip(slices, ctx_list)]
python
def split_and_load(data, ctx_list, batch_axis=0, even_split=True): """Splits an NDArray into `len(ctx_list)` slices along `batch_axis` and loads each slice to one context in `ctx_list`. Parameters ---------- data : NDArray A batch of data. ctx_list : list of Context A list of Contexts. batch_axis : int, default 0 The axis along which to slice. even_split : bool, default True Whether to force all slices to have the same number of elements. Returns ------- list of NDArray Each corresponds to a context in `ctx_list`. """ if not isinstance(data, ndarray.NDArray): data = ndarray.array(data, ctx=ctx_list[0]) if len(ctx_list) == 1: return [data.as_in_context(ctx_list[0])] slices = split_data(data, len(ctx_list), batch_axis, even_split) return [i.as_in_context(ctx) for i, ctx in zip(slices, ctx_list)]
[ "def", "split_and_load", "(", "data", ",", "ctx_list", ",", "batch_axis", "=", "0", ",", "even_split", "=", "True", ")", ":", "if", "not", "isinstance", "(", "data", ",", "ndarray", ".", "NDArray", ")", ":", "data", "=", "ndarray", ".", "array", "(", "data", ",", "ctx", "=", "ctx_list", "[", "0", "]", ")", "if", "len", "(", "ctx_list", ")", "==", "1", ":", "return", "[", "data", ".", "as_in_context", "(", "ctx_list", "[", "0", "]", ")", "]", "slices", "=", "split_data", "(", "data", ",", "len", "(", "ctx_list", ")", ",", "batch_axis", ",", "even_split", ")", "return", "[", "i", ".", "as_in_context", "(", "ctx", ")", "for", "i", ",", "ctx", "in", "zip", "(", "slices", ",", "ctx_list", ")", "]" ]
Splits an NDArray into `len(ctx_list)` slices along `batch_axis` and loads each slice to one context in `ctx_list`. Parameters ---------- data : NDArray A batch of data. ctx_list : list of Context A list of Contexts. batch_axis : int, default 0 The axis along which to slice. even_split : bool, default True Whether to force all slices to have the same number of elements. Returns ------- list of NDArray Each corresponds to a context in `ctx_list`.
[ "Splits", "an", "NDArray", "into", "len", "(", "ctx_list", ")", "slices", "along", "batch_axis", "and", "loads", "each", "slice", "to", "one", "context", "in", "ctx_list", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/utils.py#L93-L119
train
apache/incubator-mxnet
python/mxnet/gluon/utils.py
clip_global_norm
def clip_global_norm(arrays, max_norm, check_isfinite=True): """Rescales NDArrays so that the sum of their 2-norm is smaller than `max_norm`. Parameters ---------- arrays : list of NDArray max_norm : float check_isfinite : bool, default True If True, check that the total_norm is finite (not nan or inf). This requires a blocking .asscalar() call. Returns ------- NDArray or float Total norm. Return type is NDArray of shape (1,) if check_isfinite is False. Otherwise a float is returned. """ def _norm(array): if array.stype == 'default': x = array.reshape((-1,)) return ndarray.dot(x, x) return array.norm().square() assert len(arrays) > 0 ctx = arrays[0].context total_norm = ndarray.add_n(*[_norm(arr).as_in_context(ctx) for arr in arrays]) total_norm = ndarray.sqrt(total_norm) if check_isfinite: if not np.isfinite(total_norm.asscalar()): warnings.warn( UserWarning('nan or inf is detected. ' 'Clipping results will be undefined.'), stacklevel=2) scale = max_norm / (total_norm + 1e-8) scale = ndarray.min(ndarray.concat(scale, ndarray.ones(1, ctx=ctx), dim=0)) for arr in arrays: arr *= scale.as_in_context(arr.context) if check_isfinite: return total_norm.asscalar() else: return total_norm
python
def clip_global_norm(arrays, max_norm, check_isfinite=True): """Rescales NDArrays so that the sum of their 2-norm is smaller than `max_norm`. Parameters ---------- arrays : list of NDArray max_norm : float check_isfinite : bool, default True If True, check that the total_norm is finite (not nan or inf). This requires a blocking .asscalar() call. Returns ------- NDArray or float Total norm. Return type is NDArray of shape (1,) if check_isfinite is False. Otherwise a float is returned. """ def _norm(array): if array.stype == 'default': x = array.reshape((-1,)) return ndarray.dot(x, x) return array.norm().square() assert len(arrays) > 0 ctx = arrays[0].context total_norm = ndarray.add_n(*[_norm(arr).as_in_context(ctx) for arr in arrays]) total_norm = ndarray.sqrt(total_norm) if check_isfinite: if not np.isfinite(total_norm.asscalar()): warnings.warn( UserWarning('nan or inf is detected. ' 'Clipping results will be undefined.'), stacklevel=2) scale = max_norm / (total_norm + 1e-8) scale = ndarray.min(ndarray.concat(scale, ndarray.ones(1, ctx=ctx), dim=0)) for arr in arrays: arr *= scale.as_in_context(arr.context) if check_isfinite: return total_norm.asscalar() else: return total_norm
[ "def", "clip_global_norm", "(", "arrays", ",", "max_norm", ",", "check_isfinite", "=", "True", ")", ":", "def", "_norm", "(", "array", ")", ":", "if", "array", ".", "stype", "==", "'default'", ":", "x", "=", "array", ".", "reshape", "(", "(", "-", "1", ",", ")", ")", "return", "ndarray", ".", "dot", "(", "x", ",", "x", ")", "return", "array", ".", "norm", "(", ")", ".", "square", "(", ")", "assert", "len", "(", "arrays", ")", ">", "0", "ctx", "=", "arrays", "[", "0", "]", ".", "context", "total_norm", "=", "ndarray", ".", "add_n", "(", "*", "[", "_norm", "(", "arr", ")", ".", "as_in_context", "(", "ctx", ")", "for", "arr", "in", "arrays", "]", ")", "total_norm", "=", "ndarray", ".", "sqrt", "(", "total_norm", ")", "if", "check_isfinite", ":", "if", "not", "np", ".", "isfinite", "(", "total_norm", ".", "asscalar", "(", ")", ")", ":", "warnings", ".", "warn", "(", "UserWarning", "(", "'nan or inf is detected. '", "'Clipping results will be undefined.'", ")", ",", "stacklevel", "=", "2", ")", "scale", "=", "max_norm", "/", "(", "total_norm", "+", "1e-8", ")", "scale", "=", "ndarray", ".", "min", "(", "ndarray", ".", "concat", "(", "scale", ",", "ndarray", ".", "ones", "(", "1", ",", "ctx", "=", "ctx", ")", ",", "dim", "=", "0", ")", ")", "for", "arr", "in", "arrays", ":", "arr", "*=", "scale", ".", "as_in_context", "(", "arr", ".", "context", ")", "if", "check_isfinite", ":", "return", "total_norm", ".", "asscalar", "(", ")", "else", ":", "return", "total_norm" ]
Rescales NDArrays so that the sum of their 2-norm is smaller than `max_norm`. Parameters ---------- arrays : list of NDArray max_norm : float check_isfinite : bool, default True If True, check that the total_norm is finite (not nan or inf). This requires a blocking .asscalar() call. Returns ------- NDArray or float Total norm. Return type is NDArray of shape (1,) if check_isfinite is False. Otherwise a float is returned.
[ "Rescales", "NDArrays", "so", "that", "the", "sum", "of", "their", "2", "-", "norm", "is", "smaller", "than", "max_norm", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/utils.py#L122-L161
train
apache/incubator-mxnet
python/mxnet/gluon/utils.py
_indent
def _indent(s_, numSpaces): """Indent string """ s = s_.split('\n') if len(s) == 1: return s_ first = s.pop(0) s = [first] + [(numSpaces * ' ') + line for line in s] s = '\n'.join(s) return s
python
def _indent(s_, numSpaces): """Indent string """ s = s_.split('\n') if len(s) == 1: return s_ first = s.pop(0) s = [first] + [(numSpaces * ' ') + line for line in s] s = '\n'.join(s) return s
[ "def", "_indent", "(", "s_", ",", "numSpaces", ")", ":", "s", "=", "s_", ".", "split", "(", "'\\n'", ")", "if", "len", "(", "s", ")", "==", "1", ":", "return", "s_", "first", "=", "s", ".", "pop", "(", "0", ")", "s", "=", "[", "first", "]", "+", "[", "(", "numSpaces", "*", "' '", ")", "+", "line", "for", "line", "in", "s", "]", "s", "=", "'\\n'", ".", "join", "(", "s", ")", "return", "s" ]
Indent string
[ "Indent", "string" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/utils.py#L164-L173
train
apache/incubator-mxnet
python/mxnet/gluon/utils.py
check_sha1
def check_sha1(filename, sha1_hash): """Check whether the sha1 hash of the file content matches the expected hash. Parameters ---------- filename : str Path to the file. sha1_hash : str Expected sha1 hash in hexadecimal digits. Returns ------- bool Whether the file content matches the expected hash. """ sha1 = hashlib.sha1() with open(filename, 'rb') as f: while True: data = f.read(1048576) if not data: break sha1.update(data) return sha1.hexdigest() == sha1_hash
python
def check_sha1(filename, sha1_hash): """Check whether the sha1 hash of the file content matches the expected hash. Parameters ---------- filename : str Path to the file. sha1_hash : str Expected sha1 hash in hexadecimal digits. Returns ------- bool Whether the file content matches the expected hash. """ sha1 = hashlib.sha1() with open(filename, 'rb') as f: while True: data = f.read(1048576) if not data: break sha1.update(data) return sha1.hexdigest() == sha1_hash
[ "def", "check_sha1", "(", "filename", ",", "sha1_hash", ")", ":", "sha1", "=", "hashlib", ".", "sha1", "(", ")", "with", "open", "(", "filename", ",", "'rb'", ")", "as", "f", ":", "while", "True", ":", "data", "=", "f", ".", "read", "(", "1048576", ")", "if", "not", "data", ":", "break", "sha1", ".", "update", "(", "data", ")", "return", "sha1", ".", "hexdigest", "(", ")", "==", "sha1_hash" ]
Check whether the sha1 hash of the file content matches the expected hash. Parameters ---------- filename : str Path to the file. sha1_hash : str Expected sha1 hash in hexadecimal digits. Returns ------- bool Whether the file content matches the expected hash.
[ "Check", "whether", "the", "sha1", "hash", "of", "the", "file", "content", "matches", "the", "expected", "hash", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/utils.py#L176-L199
train
apache/incubator-mxnet
python/mxnet/gluon/utils.py
download
def download(url, path=None, overwrite=False, sha1_hash=None, retries=5, verify_ssl=True): """Download an given URL Parameters ---------- url : str URL to download path : str, optional Destination path to store downloaded file. By default stores to the current directory with same name as in url. overwrite : bool, optional Whether to overwrite destination file if already exists. sha1_hash : str, optional Expected sha1 hash in hexadecimal digits. Will ignore existing file when hash is specified but doesn't match. retries : integer, default 5 The number of times to attempt the download in case of failure or non 200 return codes verify_ssl : bool, default True Verify SSL certificates. Returns ------- str The file path of the downloaded file. """ if path is None: fname = url.split('/')[-1] # Empty filenames are invalid assert fname, 'Can\'t construct file-name from this URL. ' \ 'Please set the `path` option manually.' else: path = os.path.expanduser(path) if os.path.isdir(path): fname = os.path.join(path, url.split('/')[-1]) else: fname = path assert retries >= 0, "Number of retries should be at least 0, currently it's {}".format( retries) if not verify_ssl: warnings.warn( 'Unverified HTTPS request is being made (verify_ssl=False). ' 'Adding certificate verification is strongly advised.') if overwrite or not os.path.exists(fname) or (sha1_hash and not check_sha1(fname, sha1_hash)): dirname = os.path.dirname(os.path.abspath(os.path.expanduser(fname))) if not os.path.exists(dirname): os.makedirs(dirname) while retries + 1 > 0: # Disable pyling too broad Exception # pylint: disable=W0703 try: print('Downloading {} from {}...'.format(fname, url)) r = requests.get(url, stream=True, verify=verify_ssl) if r.status_code != 200: raise RuntimeError('Failed downloading url {}'.format(url)) # create uuid for temporary files random_uuid = str(uuid.uuid4()) with open('{}.{}'.format(fname, random_uuid), 'wb') as f: for chunk in r.iter_content(chunk_size=1024): if chunk: # filter out keep-alive new chunks f.write(chunk) # if the target file exists(created by other processes) # and have the same hash with target file # delete the temporary file if not os.path.exists(fname) or (sha1_hash and not check_sha1(fname, sha1_hash)): # atmoic operation in the same file system _replace_atomic('{}.{}'.format(fname, random_uuid), fname) else: try: os.remove('{}.{}'.format(fname, random_uuid)) except OSError: pass finally: warnings.warn( 'File {} exists in file system so the downloaded file is deleted'.format(fname)) if sha1_hash and not check_sha1(fname, sha1_hash): raise UserWarning( 'File {} is downloaded but the content hash does not match.' ' The repo may be outdated or download may be incomplete. ' 'If the "repo_url" is overridden, consider switching to ' 'the default repo.'.format(fname)) break except Exception as e: retries -= 1 if retries <= 0: raise e else: print('download failed due to {}, retrying, {} attempt{} left' .format(repr(e), retries, 's' if retries > 1 else '')) return fname
python
def download(url, path=None, overwrite=False, sha1_hash=None, retries=5, verify_ssl=True): """Download an given URL Parameters ---------- url : str URL to download path : str, optional Destination path to store downloaded file. By default stores to the current directory with same name as in url. overwrite : bool, optional Whether to overwrite destination file if already exists. sha1_hash : str, optional Expected sha1 hash in hexadecimal digits. Will ignore existing file when hash is specified but doesn't match. retries : integer, default 5 The number of times to attempt the download in case of failure or non 200 return codes verify_ssl : bool, default True Verify SSL certificates. Returns ------- str The file path of the downloaded file. """ if path is None: fname = url.split('/')[-1] # Empty filenames are invalid assert fname, 'Can\'t construct file-name from this URL. ' \ 'Please set the `path` option manually.' else: path = os.path.expanduser(path) if os.path.isdir(path): fname = os.path.join(path, url.split('/')[-1]) else: fname = path assert retries >= 0, "Number of retries should be at least 0, currently it's {}".format( retries) if not verify_ssl: warnings.warn( 'Unverified HTTPS request is being made (verify_ssl=False). ' 'Adding certificate verification is strongly advised.') if overwrite or not os.path.exists(fname) or (sha1_hash and not check_sha1(fname, sha1_hash)): dirname = os.path.dirname(os.path.abspath(os.path.expanduser(fname))) if not os.path.exists(dirname): os.makedirs(dirname) while retries + 1 > 0: # Disable pyling too broad Exception # pylint: disable=W0703 try: print('Downloading {} from {}...'.format(fname, url)) r = requests.get(url, stream=True, verify=verify_ssl) if r.status_code != 200: raise RuntimeError('Failed downloading url {}'.format(url)) # create uuid for temporary files random_uuid = str(uuid.uuid4()) with open('{}.{}'.format(fname, random_uuid), 'wb') as f: for chunk in r.iter_content(chunk_size=1024): if chunk: # filter out keep-alive new chunks f.write(chunk) # if the target file exists(created by other processes) # and have the same hash with target file # delete the temporary file if not os.path.exists(fname) or (sha1_hash and not check_sha1(fname, sha1_hash)): # atmoic operation in the same file system _replace_atomic('{}.{}'.format(fname, random_uuid), fname) else: try: os.remove('{}.{}'.format(fname, random_uuid)) except OSError: pass finally: warnings.warn( 'File {} exists in file system so the downloaded file is deleted'.format(fname)) if sha1_hash and not check_sha1(fname, sha1_hash): raise UserWarning( 'File {} is downloaded but the content hash does not match.' ' The repo may be outdated or download may be incomplete. ' 'If the "repo_url" is overridden, consider switching to ' 'the default repo.'.format(fname)) break except Exception as e: retries -= 1 if retries <= 0: raise e else: print('download failed due to {}, retrying, {} attempt{} left' .format(repr(e), retries, 's' if retries > 1 else '')) return fname
[ "def", "download", "(", "url", ",", "path", "=", "None", ",", "overwrite", "=", "False", ",", "sha1_hash", "=", "None", ",", "retries", "=", "5", ",", "verify_ssl", "=", "True", ")", ":", "if", "path", "is", "None", ":", "fname", "=", "url", ".", "split", "(", "'/'", ")", "[", "-", "1", "]", "# Empty filenames are invalid", "assert", "fname", ",", "'Can\\'t construct file-name from this URL. '", "'Please set the `path` option manually.'", "else", ":", "path", "=", "os", ".", "path", ".", "expanduser", "(", "path", ")", "if", "os", ".", "path", ".", "isdir", "(", "path", ")", ":", "fname", "=", "os", ".", "path", ".", "join", "(", "path", ",", "url", ".", "split", "(", "'/'", ")", "[", "-", "1", "]", ")", "else", ":", "fname", "=", "path", "assert", "retries", ">=", "0", ",", "\"Number of retries should be at least 0, currently it's {}\"", ".", "format", "(", "retries", ")", "if", "not", "verify_ssl", ":", "warnings", ".", "warn", "(", "'Unverified HTTPS request is being made (verify_ssl=False). '", "'Adding certificate verification is strongly advised.'", ")", "if", "overwrite", "or", "not", "os", ".", "path", ".", "exists", "(", "fname", ")", "or", "(", "sha1_hash", "and", "not", "check_sha1", "(", "fname", ",", "sha1_hash", ")", ")", ":", "dirname", "=", "os", ".", "path", ".", "dirname", "(", "os", ".", "path", ".", "abspath", "(", "os", ".", "path", ".", "expanduser", "(", "fname", ")", ")", ")", "if", "not", "os", ".", "path", ".", "exists", "(", "dirname", ")", ":", "os", ".", "makedirs", "(", "dirname", ")", "while", "retries", "+", "1", ">", "0", ":", "# Disable pyling too broad Exception", "# pylint: disable=W0703", "try", ":", "print", "(", "'Downloading {} from {}...'", ".", "format", "(", "fname", ",", "url", ")", ")", "r", "=", "requests", ".", "get", "(", "url", ",", "stream", "=", "True", ",", "verify", "=", "verify_ssl", ")", "if", "r", ".", "status_code", "!=", "200", ":", "raise", "RuntimeError", "(", "'Failed downloading url {}'", ".", "format", "(", "url", ")", ")", "# create uuid for temporary files", "random_uuid", "=", "str", "(", "uuid", ".", "uuid4", "(", ")", ")", "with", "open", "(", "'{}.{}'", ".", "format", "(", "fname", ",", "random_uuid", ")", ",", "'wb'", ")", "as", "f", ":", "for", "chunk", "in", "r", ".", "iter_content", "(", "chunk_size", "=", "1024", ")", ":", "if", "chunk", ":", "# filter out keep-alive new chunks", "f", ".", "write", "(", "chunk", ")", "# if the target file exists(created by other processes)", "# and have the same hash with target file", "# delete the temporary file", "if", "not", "os", ".", "path", ".", "exists", "(", "fname", ")", "or", "(", "sha1_hash", "and", "not", "check_sha1", "(", "fname", ",", "sha1_hash", ")", ")", ":", "# atmoic operation in the same file system", "_replace_atomic", "(", "'{}.{}'", ".", "format", "(", "fname", ",", "random_uuid", ")", ",", "fname", ")", "else", ":", "try", ":", "os", ".", "remove", "(", "'{}.{}'", ".", "format", "(", "fname", ",", "random_uuid", ")", ")", "except", "OSError", ":", "pass", "finally", ":", "warnings", ".", "warn", "(", "'File {} exists in file system so the downloaded file is deleted'", ".", "format", "(", "fname", ")", ")", "if", "sha1_hash", "and", "not", "check_sha1", "(", "fname", ",", "sha1_hash", ")", ":", "raise", "UserWarning", "(", "'File {} is downloaded but the content hash does not match.'", "' The repo may be outdated or download may be incomplete. '", "'If the \"repo_url\" is overridden, consider switching to '", "'the default repo.'", ".", "format", "(", "fname", ")", ")", "break", "except", "Exception", "as", "e", ":", "retries", "-=", "1", "if", "retries", "<=", "0", ":", "raise", "e", "else", ":", "print", "(", "'download failed due to {}, retrying, {} attempt{} left'", ".", "format", "(", "repr", "(", "e", ")", ",", "retries", ",", "'s'", "if", "retries", ">", "1", "else", "''", ")", ")", "return", "fname" ]
Download an given URL Parameters ---------- url : str URL to download path : str, optional Destination path to store downloaded file. By default stores to the current directory with same name as in url. overwrite : bool, optional Whether to overwrite destination file if already exists. sha1_hash : str, optional Expected sha1 hash in hexadecimal digits. Will ignore existing file when hash is specified but doesn't match. retries : integer, default 5 The number of times to attempt the download in case of failure or non 200 return codes verify_ssl : bool, default True Verify SSL certificates. Returns ------- str The file path of the downloaded file.
[ "Download", "an", "given", "URL" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/utils.py#L258-L349
train
apache/incubator-mxnet
python/mxnet/gluon/utils.py
_get_repo_url
def _get_repo_url(): """Return the base URL for Gluon dataset and model repository.""" default_repo = 'https://apache-mxnet.s3-accelerate.dualstack.amazonaws.com/' repo_url = os.environ.get('MXNET_GLUON_REPO', default_repo) if repo_url[-1] != '/': repo_url = repo_url+'/' return repo_url
python
def _get_repo_url(): """Return the base URL for Gluon dataset and model repository.""" default_repo = 'https://apache-mxnet.s3-accelerate.dualstack.amazonaws.com/' repo_url = os.environ.get('MXNET_GLUON_REPO', default_repo) if repo_url[-1] != '/': repo_url = repo_url+'/' return repo_url
[ "def", "_get_repo_url", "(", ")", ":", "default_repo", "=", "'https://apache-mxnet.s3-accelerate.dualstack.amazonaws.com/'", "repo_url", "=", "os", ".", "environ", ".", "get", "(", "'MXNET_GLUON_REPO'", ",", "default_repo", ")", "if", "repo_url", "[", "-", "1", "]", "!=", "'/'", ":", "repo_url", "=", "repo_url", "+", "'/'", "return", "repo_url" ]
Return the base URL for Gluon dataset and model repository.
[ "Return", "the", "base", "URL", "for", "Gluon", "dataset", "and", "model", "repository", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/utils.py#L351-L357
train
apache/incubator-mxnet
python/mxnet/gluon/utils.py
_get_repo_file_url
def _get_repo_file_url(namespace, filename): """Return the URL for hosted file in Gluon repository. Parameters ---------- namespace : str Namespace of the file. filename : str Name of the file """ return '{base_url}{namespace}/{filename}'.format(base_url=_get_repo_url(), namespace=namespace, filename=filename)
python
def _get_repo_file_url(namespace, filename): """Return the URL for hosted file in Gluon repository. Parameters ---------- namespace : str Namespace of the file. filename : str Name of the file """ return '{base_url}{namespace}/{filename}'.format(base_url=_get_repo_url(), namespace=namespace, filename=filename)
[ "def", "_get_repo_file_url", "(", "namespace", ",", "filename", ")", ":", "return", "'{base_url}{namespace}/{filename}'", ".", "format", "(", "base_url", "=", "_get_repo_url", "(", ")", ",", "namespace", "=", "namespace", ",", "filename", "=", "filename", ")" ]
Return the URL for hosted file in Gluon repository. Parameters ---------- namespace : str Namespace of the file. filename : str Name of the file
[ "Return", "the", "URL", "for", "hosted", "file", "in", "Gluon", "repository", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/utils.py#L359-L371
train
apache/incubator-mxnet
python/mxnet/gluon/utils.py
_brief_print_list
def _brief_print_list(lst, limit=7): """Print at most `limit` elements of list.""" lst = list(lst) if len(lst) > limit: return _brief_print_list(lst[:limit//2], limit) + ', ..., ' + \ _brief_print_list(lst[-limit//2:], limit) return ', '.join(["'%s'"%str(i) for i in lst])
python
def _brief_print_list(lst, limit=7): """Print at most `limit` elements of list.""" lst = list(lst) if len(lst) > limit: return _brief_print_list(lst[:limit//2], limit) + ', ..., ' + \ _brief_print_list(lst[-limit//2:], limit) return ', '.join(["'%s'"%str(i) for i in lst])
[ "def", "_brief_print_list", "(", "lst", ",", "limit", "=", "7", ")", ":", "lst", "=", "list", "(", "lst", ")", "if", "len", "(", "lst", ")", ">", "limit", ":", "return", "_brief_print_list", "(", "lst", "[", ":", "limit", "//", "2", "]", ",", "limit", ")", "+", "', ..., '", "+", "_brief_print_list", "(", "lst", "[", "-", "limit", "//", "2", ":", "]", ",", "limit", ")", "return", "', '", ".", "join", "(", "[", "\"'%s'\"", "%", "str", "(", "i", ")", "for", "i", "in", "lst", "]", ")" ]
Print at most `limit` elements of list.
[ "Print", "at", "most", "limit", "elements", "of", "list", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/utils.py#L373-L379
train
apache/incubator-mxnet
python/mxnet/symbol/register.py
_make_symbol_function
def _make_symbol_function(handle, name, func_name): """Create a symbol function by handle and function name.""" code, doc_str = _generate_symbol_function_code(handle, name, func_name) local = {} exec(code, None, local) # pylint: disable=exec-used symbol_function = local[func_name] symbol_function.__name__ = func_name symbol_function.__doc__ = doc_str symbol_function.__module__ = 'mxnet.symbol' return symbol_function
python
def _make_symbol_function(handle, name, func_name): """Create a symbol function by handle and function name.""" code, doc_str = _generate_symbol_function_code(handle, name, func_name) local = {} exec(code, None, local) # pylint: disable=exec-used symbol_function = local[func_name] symbol_function.__name__ = func_name symbol_function.__doc__ = doc_str symbol_function.__module__ = 'mxnet.symbol' return symbol_function
[ "def", "_make_symbol_function", "(", "handle", ",", "name", ",", "func_name", ")", ":", "code", ",", "doc_str", "=", "_generate_symbol_function_code", "(", "handle", ",", "name", ",", "func_name", ")", "local", "=", "{", "}", "exec", "(", "code", ",", "None", ",", "local", ")", "# pylint: disable=exec-used", "symbol_function", "=", "local", "[", "func_name", "]", "symbol_function", ".", "__name__", "=", "func_name", "symbol_function", ".", "__doc__", "=", "doc_str", "symbol_function", ".", "__module__", "=", "'mxnet.symbol'", "return", "symbol_function" ]
Create a symbol function by handle and function name.
[ "Create", "a", "symbol", "function", "by", "handle", "and", "function", "name", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/symbol/register.py#L199-L209
train
apache/incubator-mxnet
example/sparse/matrix_factorization/train.py
batch_row_ids
def batch_row_ids(data_batch): """ Generate row ids based on the current mini-batch """ item = data_batch.data[0] user = data_batch.data[1] return {'user_weight': user.astype(np.int64), 'item_weight': item.astype(np.int64)}
python
def batch_row_ids(data_batch): """ Generate row ids based on the current mini-batch """ item = data_batch.data[0] user = data_batch.data[1] return {'user_weight': user.astype(np.int64), 'item_weight': item.astype(np.int64)}
[ "def", "batch_row_ids", "(", "data_batch", ")", ":", "item", "=", "data_batch", ".", "data", "[", "0", "]", "user", "=", "data_batch", ".", "data", "[", "1", "]", "return", "{", "'user_weight'", ":", "user", ".", "astype", "(", "np", ".", "int64", ")", ",", "'item_weight'", ":", "item", ".", "astype", "(", "np", ".", "int64", ")", "}" ]
Generate row ids based on the current mini-batch
[ "Generate", "row", "ids", "based", "on", "the", "current", "mini", "-", "batch" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/sparse/matrix_factorization/train.py#L52-L57
train
apache/incubator-mxnet
example/sparse/matrix_factorization/train.py
all_row_ids
def all_row_ids(data_batch): """ Generate row ids for all rows """ all_users = mx.nd.arange(0, MOVIELENS['max_user'], dtype='int64') all_movies = mx.nd.arange(0, MOVIELENS['max_movie'], dtype='int64') return {'user_weight': all_users, 'item_weight': all_movies}
python
def all_row_ids(data_batch): """ Generate row ids for all rows """ all_users = mx.nd.arange(0, MOVIELENS['max_user'], dtype='int64') all_movies = mx.nd.arange(0, MOVIELENS['max_movie'], dtype='int64') return {'user_weight': all_users, 'item_weight': all_movies}
[ "def", "all_row_ids", "(", "data_batch", ")", ":", "all_users", "=", "mx", ".", "nd", ".", "arange", "(", "0", ",", "MOVIELENS", "[", "'max_user'", "]", ",", "dtype", "=", "'int64'", ")", "all_movies", "=", "mx", ".", "nd", ".", "arange", "(", "0", ",", "MOVIELENS", "[", "'max_movie'", "]", ",", "dtype", "=", "'int64'", ")", "return", "{", "'user_weight'", ":", "all_users", ",", "'item_weight'", ":", "all_movies", "}" ]
Generate row ids for all rows
[ "Generate", "row", "ids", "for", "all", "rows" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/sparse/matrix_factorization/train.py#L59-L63
train
apache/incubator-mxnet
example/ssd/tools/caffe_converter/convert_model.py
convert_model
def convert_model(prototxt_fname, caffemodel_fname, output_prefix=None): """Convert caffe model Parameters ---------- prototxt_fname : str Filename of the prototxt model definition caffemodel_fname : str Filename of the binary caffe model output_prefix : str, optinoal If given, then save the converted MXNet into output_prefx+'.json' and output_prefx+'.params' Returns ------- sym : Symbol Symbol convereted from prototxt arg_params : list of NDArray Argument parameters aux_params : list of NDArray Aux parameters input_dim : tuple Input dimension """ sym, input_dim = convert_symbol(prototxt_fname) arg_shapes, _, aux_shapes = sym.infer_shape(data=tuple(input_dim)) arg_names = sym.list_arguments() aux_names = sym.list_auxiliary_states() arg_shape_dic = dict(zip(arg_names, arg_shapes)) aux_shape_dic = dict(zip(aux_names, aux_shapes)) arg_params = {} aux_params = {} first_conv = True layers, names = caffe_parser.read_caffemodel(prototxt_fname, caffemodel_fname) layer_iter = caffe_parser.layer_iter(layers, names) layers_proto = caffe_parser.get_layers(caffe_parser.read_prototxt(prototxt_fname)) for layer_name, layer_type, layer_blobs in layer_iter: if layer_type == 'Convolution' or layer_type == 'InnerProduct' \ or layer_type == 4 or layer_type == 14 or layer_type == 'PReLU' \ or layer_type == 'Deconvolution' or layer_type == 39 or layer_type == 'Normalize': if layer_type == 'PReLU': assert (len(layer_blobs) == 1) wmat = layer_blobs[0].data weight_name = layer_name + '_gamma' arg_params[weight_name] = mx.nd.zeros(wmat.shape) arg_params[weight_name][:] = wmat continue if layer_type == 'Normalize': assert (len(layer_blobs) == 1) weight_name = layer_name + '_scale' wmat = layer_blobs[0].data arg_params[weight_name] = mx.nd.zeros((1, len(wmat), 1, 1)) arg_params[weight_name][:] = np.array(list(wmat)).reshape((1, len(wmat), 1, 1)) continue wmat_dim = [] if getattr(layer_blobs[0].shape, 'dim', None) is not None: if len(layer_blobs[0].shape.dim) > 0: wmat_dim = layer_blobs[0].shape.dim else: wmat_dim = [layer_blobs[0].num, layer_blobs[0].channels, layer_blobs[0].height, layer_blobs[0].width] else: wmat_dim = list(layer_blobs[0].shape) wmat = np.array(layer_blobs[0].data).reshape(wmat_dim) channels = wmat_dim[1] if channels == 3 or channels == 4: # RGB or RGBA if first_conv: # Swapping BGR of caffe into RGB in mxnet wmat[:, [0, 2], :, :] = wmat[:, [2, 0], :, :] assert(wmat.flags['C_CONTIGUOUS'] is True) sys.stdout.write('converting layer {0}, wmat shape = {1}'.format( layer_name, wmat.shape)) if len(layer_blobs) == 2: bias = np.array(layer_blobs[1].data) bias = bias.reshape((bias.shape[0], 1)) assert(bias.flags['C_CONTIGUOUS'] is True) bias_name = layer_name + "_bias" if bias_name not in arg_shape_dic: print(bias_name + ' not found in arg_shape_dic.') continue bias = bias.reshape(arg_shape_dic[bias_name]) arg_params[bias_name] = mx.nd.zeros(bias.shape) arg_params[bias_name][:] = bias sys.stdout.write(', bias shape = {}'.format(bias.shape)) sys.stdout.write('\n') sys.stdout.flush() wmat = wmat.reshape((wmat.shape[0], -1)) weight_name = layer_name + "_weight" if weight_name not in arg_shape_dic: print(weight_name + ' not found in arg_shape_dic.') continue wmat = wmat.reshape(arg_shape_dic[weight_name]) arg_params[weight_name] = mx.nd.zeros(wmat.shape) arg_params[weight_name][:] = wmat if first_conv and (layer_type == 'Convolution' or layer_type == 4): first_conv = False elif layer_type == 'Scale': if 'scale' in layer_name: bn_name = layer_name.replace('scale', 'bn') elif 'sc' in layer_name: bn_name = layer_name.replace('sc', 'bn') else: assert False, 'Unknown name convention for bn/scale' gamma = np.array(layer_blobs[0].data) beta = np.array(layer_blobs[1].data) # beta = np.expand_dims(beta, 1) beta_name = '{}_beta'.format(bn_name) gamma_name = '{}_gamma'.format(bn_name) beta = beta.reshape(arg_shape_dic[beta_name]) gamma = gamma.reshape(arg_shape_dic[gamma_name]) arg_params[beta_name] = mx.nd.zeros(beta.shape) arg_params[gamma_name] = mx.nd.zeros(gamma.shape) arg_params[beta_name][:] = beta arg_params[gamma_name][:] = gamma assert gamma.flags['C_CONTIGUOUS'] is True assert beta.flags['C_CONTIGUOUS'] is True print('converting scale layer, beta shape = {}, gamma shape = {}'.format( beta.shape, gamma.shape)) elif layer_type == 'BatchNorm': bn_name = layer_name mean = np.array(layer_blobs[0].data) var = np.array(layer_blobs[1].data) rescale_factor = layer_blobs[2].data[0] if rescale_factor != 0: rescale_factor = 1 / rescale_factor mean_name = '{}_moving_mean'.format(bn_name) var_name = '{}_moving_var'.format(bn_name) mean = mean.reshape(aux_shape_dic[mean_name]) var = var.reshape(aux_shape_dic[var_name]) aux_params[mean_name] = mx.nd.zeros(mean.shape) aux_params[var_name] = mx.nd.zeros(var.shape) # Get the original epsilon for idx, layer in enumerate(layers_proto): if layer.name == bn_name: bn_index = idx eps_caffe = layers_proto[bn_index].batch_norm_param.eps # Compensate for the epsilon shift performed in convert_symbol eps_symbol = float(sym.attr_dict()[bn_name + '_moving_mean']['eps']) eps_correction = eps_caffe - eps_symbol # Fill parameters aux_params[mean_name][:] = mean * rescale_factor aux_params[var_name][:] = var * rescale_factor + eps_correction assert var.flags['C_CONTIGUOUS'] is True assert mean.flags['C_CONTIGUOUS'] is True print('converting batchnorm layer, mean shape = {}, var shape = {}'.format( mean.shape, var.shape)) fix_gamma = layers_proto[bn_index+1].type != 'Scale' if fix_gamma: gamma_name = '{}_gamma'.format(bn_name) gamma = np.array(np.ones(arg_shape_dic[gamma_name])) beta_name = '{}_beta'.format(bn_name) beta = np.array(np.zeros(arg_shape_dic[beta_name])) arg_params[beta_name] = mx.nd.zeros(beta.shape) arg_params[gamma_name] = mx.nd.zeros(gamma.shape) arg_params[beta_name][:] = beta arg_params[gamma_name][:] = gamma assert gamma.flags['C_CONTIGUOUS'] is True assert beta.flags['C_CONTIGUOUS'] is True else: print('\tskipping layer {} of type {}'.format(layer_name, layer_type)) assert len(layer_blobs) == 0 if output_prefix is not None: model = mx.mod.Module(symbol=sym, label_names=None) model.bind(data_shapes=[('data', tuple(input_dim))]) model.init_params(arg_params=arg_params, aux_params=aux_params) model.save_checkpoint(output_prefix, 0) return sym, arg_params, aux_params, input_dim
python
def convert_model(prototxt_fname, caffemodel_fname, output_prefix=None): """Convert caffe model Parameters ---------- prototxt_fname : str Filename of the prototxt model definition caffemodel_fname : str Filename of the binary caffe model output_prefix : str, optinoal If given, then save the converted MXNet into output_prefx+'.json' and output_prefx+'.params' Returns ------- sym : Symbol Symbol convereted from prototxt arg_params : list of NDArray Argument parameters aux_params : list of NDArray Aux parameters input_dim : tuple Input dimension """ sym, input_dim = convert_symbol(prototxt_fname) arg_shapes, _, aux_shapes = sym.infer_shape(data=tuple(input_dim)) arg_names = sym.list_arguments() aux_names = sym.list_auxiliary_states() arg_shape_dic = dict(zip(arg_names, arg_shapes)) aux_shape_dic = dict(zip(aux_names, aux_shapes)) arg_params = {} aux_params = {} first_conv = True layers, names = caffe_parser.read_caffemodel(prototxt_fname, caffemodel_fname) layer_iter = caffe_parser.layer_iter(layers, names) layers_proto = caffe_parser.get_layers(caffe_parser.read_prototxt(prototxt_fname)) for layer_name, layer_type, layer_blobs in layer_iter: if layer_type == 'Convolution' or layer_type == 'InnerProduct' \ or layer_type == 4 or layer_type == 14 or layer_type == 'PReLU' \ or layer_type == 'Deconvolution' or layer_type == 39 or layer_type == 'Normalize': if layer_type == 'PReLU': assert (len(layer_blobs) == 1) wmat = layer_blobs[0].data weight_name = layer_name + '_gamma' arg_params[weight_name] = mx.nd.zeros(wmat.shape) arg_params[weight_name][:] = wmat continue if layer_type == 'Normalize': assert (len(layer_blobs) == 1) weight_name = layer_name + '_scale' wmat = layer_blobs[0].data arg_params[weight_name] = mx.nd.zeros((1, len(wmat), 1, 1)) arg_params[weight_name][:] = np.array(list(wmat)).reshape((1, len(wmat), 1, 1)) continue wmat_dim = [] if getattr(layer_blobs[0].shape, 'dim', None) is not None: if len(layer_blobs[0].shape.dim) > 0: wmat_dim = layer_blobs[0].shape.dim else: wmat_dim = [layer_blobs[0].num, layer_blobs[0].channels, layer_blobs[0].height, layer_blobs[0].width] else: wmat_dim = list(layer_blobs[0].shape) wmat = np.array(layer_blobs[0].data).reshape(wmat_dim) channels = wmat_dim[1] if channels == 3 or channels == 4: # RGB or RGBA if first_conv: # Swapping BGR of caffe into RGB in mxnet wmat[:, [0, 2], :, :] = wmat[:, [2, 0], :, :] assert(wmat.flags['C_CONTIGUOUS'] is True) sys.stdout.write('converting layer {0}, wmat shape = {1}'.format( layer_name, wmat.shape)) if len(layer_blobs) == 2: bias = np.array(layer_blobs[1].data) bias = bias.reshape((bias.shape[0], 1)) assert(bias.flags['C_CONTIGUOUS'] is True) bias_name = layer_name + "_bias" if bias_name not in arg_shape_dic: print(bias_name + ' not found in arg_shape_dic.') continue bias = bias.reshape(arg_shape_dic[bias_name]) arg_params[bias_name] = mx.nd.zeros(bias.shape) arg_params[bias_name][:] = bias sys.stdout.write(', bias shape = {}'.format(bias.shape)) sys.stdout.write('\n') sys.stdout.flush() wmat = wmat.reshape((wmat.shape[0], -1)) weight_name = layer_name + "_weight" if weight_name not in arg_shape_dic: print(weight_name + ' not found in arg_shape_dic.') continue wmat = wmat.reshape(arg_shape_dic[weight_name]) arg_params[weight_name] = mx.nd.zeros(wmat.shape) arg_params[weight_name][:] = wmat if first_conv and (layer_type == 'Convolution' or layer_type == 4): first_conv = False elif layer_type == 'Scale': if 'scale' in layer_name: bn_name = layer_name.replace('scale', 'bn') elif 'sc' in layer_name: bn_name = layer_name.replace('sc', 'bn') else: assert False, 'Unknown name convention for bn/scale' gamma = np.array(layer_blobs[0].data) beta = np.array(layer_blobs[1].data) # beta = np.expand_dims(beta, 1) beta_name = '{}_beta'.format(bn_name) gamma_name = '{}_gamma'.format(bn_name) beta = beta.reshape(arg_shape_dic[beta_name]) gamma = gamma.reshape(arg_shape_dic[gamma_name]) arg_params[beta_name] = mx.nd.zeros(beta.shape) arg_params[gamma_name] = mx.nd.zeros(gamma.shape) arg_params[beta_name][:] = beta arg_params[gamma_name][:] = gamma assert gamma.flags['C_CONTIGUOUS'] is True assert beta.flags['C_CONTIGUOUS'] is True print('converting scale layer, beta shape = {}, gamma shape = {}'.format( beta.shape, gamma.shape)) elif layer_type == 'BatchNorm': bn_name = layer_name mean = np.array(layer_blobs[0].data) var = np.array(layer_blobs[1].data) rescale_factor = layer_blobs[2].data[0] if rescale_factor != 0: rescale_factor = 1 / rescale_factor mean_name = '{}_moving_mean'.format(bn_name) var_name = '{}_moving_var'.format(bn_name) mean = mean.reshape(aux_shape_dic[mean_name]) var = var.reshape(aux_shape_dic[var_name]) aux_params[mean_name] = mx.nd.zeros(mean.shape) aux_params[var_name] = mx.nd.zeros(var.shape) # Get the original epsilon for idx, layer in enumerate(layers_proto): if layer.name == bn_name: bn_index = idx eps_caffe = layers_proto[bn_index].batch_norm_param.eps # Compensate for the epsilon shift performed in convert_symbol eps_symbol = float(sym.attr_dict()[bn_name + '_moving_mean']['eps']) eps_correction = eps_caffe - eps_symbol # Fill parameters aux_params[mean_name][:] = mean * rescale_factor aux_params[var_name][:] = var * rescale_factor + eps_correction assert var.flags['C_CONTIGUOUS'] is True assert mean.flags['C_CONTIGUOUS'] is True print('converting batchnorm layer, mean shape = {}, var shape = {}'.format( mean.shape, var.shape)) fix_gamma = layers_proto[bn_index+1].type != 'Scale' if fix_gamma: gamma_name = '{}_gamma'.format(bn_name) gamma = np.array(np.ones(arg_shape_dic[gamma_name])) beta_name = '{}_beta'.format(bn_name) beta = np.array(np.zeros(arg_shape_dic[beta_name])) arg_params[beta_name] = mx.nd.zeros(beta.shape) arg_params[gamma_name] = mx.nd.zeros(gamma.shape) arg_params[beta_name][:] = beta arg_params[gamma_name][:] = gamma assert gamma.flags['C_CONTIGUOUS'] is True assert beta.flags['C_CONTIGUOUS'] is True else: print('\tskipping layer {} of type {}'.format(layer_name, layer_type)) assert len(layer_blobs) == 0 if output_prefix is not None: model = mx.mod.Module(symbol=sym, label_names=None) model.bind(data_shapes=[('data', tuple(input_dim))]) model.init_params(arg_params=arg_params, aux_params=aux_params) model.save_checkpoint(output_prefix, 0) return sym, arg_params, aux_params, input_dim
[ "def", "convert_model", "(", "prototxt_fname", ",", "caffemodel_fname", ",", "output_prefix", "=", "None", ")", ":", "sym", ",", "input_dim", "=", "convert_symbol", "(", "prototxt_fname", ")", "arg_shapes", ",", "_", ",", "aux_shapes", "=", "sym", ".", "infer_shape", "(", "data", "=", "tuple", "(", "input_dim", ")", ")", "arg_names", "=", "sym", ".", "list_arguments", "(", ")", "aux_names", "=", "sym", ".", "list_auxiliary_states", "(", ")", "arg_shape_dic", "=", "dict", "(", "zip", "(", "arg_names", ",", "arg_shapes", ")", ")", "aux_shape_dic", "=", "dict", "(", "zip", "(", "aux_names", ",", "aux_shapes", ")", ")", "arg_params", "=", "{", "}", "aux_params", "=", "{", "}", "first_conv", "=", "True", "layers", ",", "names", "=", "caffe_parser", ".", "read_caffemodel", "(", "prototxt_fname", ",", "caffemodel_fname", ")", "layer_iter", "=", "caffe_parser", ".", "layer_iter", "(", "layers", ",", "names", ")", "layers_proto", "=", "caffe_parser", ".", "get_layers", "(", "caffe_parser", ".", "read_prototxt", "(", "prototxt_fname", ")", ")", "for", "layer_name", ",", "layer_type", ",", "layer_blobs", "in", "layer_iter", ":", "if", "layer_type", "==", "'Convolution'", "or", "layer_type", "==", "'InnerProduct'", "or", "layer_type", "==", "4", "or", "layer_type", "==", "14", "or", "layer_type", "==", "'PReLU'", "or", "layer_type", "==", "'Deconvolution'", "or", "layer_type", "==", "39", "or", "layer_type", "==", "'Normalize'", ":", "if", "layer_type", "==", "'PReLU'", ":", "assert", "(", "len", "(", "layer_blobs", ")", "==", "1", ")", "wmat", "=", "layer_blobs", "[", "0", "]", ".", "data", "weight_name", "=", "layer_name", "+", "'_gamma'", "arg_params", "[", "weight_name", "]", "=", "mx", ".", "nd", ".", "zeros", "(", "wmat", ".", "shape", ")", "arg_params", "[", "weight_name", "]", "[", ":", "]", "=", "wmat", "continue", "if", "layer_type", "==", "'Normalize'", ":", "assert", "(", "len", "(", "layer_blobs", ")", "==", "1", ")", "weight_name", "=", "layer_name", "+", "'_scale'", "wmat", "=", "layer_blobs", "[", "0", "]", ".", "data", "arg_params", "[", "weight_name", "]", "=", "mx", ".", "nd", ".", "zeros", "(", "(", "1", ",", "len", "(", "wmat", ")", ",", "1", ",", "1", ")", ")", "arg_params", "[", "weight_name", "]", "[", ":", "]", "=", "np", ".", "array", "(", "list", "(", "wmat", ")", ")", ".", "reshape", "(", "(", "1", ",", "len", "(", "wmat", ")", ",", "1", ",", "1", ")", ")", "continue", "wmat_dim", "=", "[", "]", "if", "getattr", "(", "layer_blobs", "[", "0", "]", ".", "shape", ",", "'dim'", ",", "None", ")", "is", "not", "None", ":", "if", "len", "(", "layer_blobs", "[", "0", "]", ".", "shape", ".", "dim", ")", ">", "0", ":", "wmat_dim", "=", "layer_blobs", "[", "0", "]", ".", "shape", ".", "dim", "else", ":", "wmat_dim", "=", "[", "layer_blobs", "[", "0", "]", ".", "num", ",", "layer_blobs", "[", "0", "]", ".", "channels", ",", "layer_blobs", "[", "0", "]", ".", "height", ",", "layer_blobs", "[", "0", "]", ".", "width", "]", "else", ":", "wmat_dim", "=", "list", "(", "layer_blobs", "[", "0", "]", ".", "shape", ")", "wmat", "=", "np", ".", "array", "(", "layer_blobs", "[", "0", "]", ".", "data", ")", ".", "reshape", "(", "wmat_dim", ")", "channels", "=", "wmat_dim", "[", "1", "]", "if", "channels", "==", "3", "or", "channels", "==", "4", ":", "# RGB or RGBA", "if", "first_conv", ":", "# Swapping BGR of caffe into RGB in mxnet", "wmat", "[", ":", ",", "[", "0", ",", "2", "]", ",", ":", ",", ":", "]", "=", "wmat", "[", ":", ",", "[", "2", ",", "0", "]", ",", ":", ",", ":", "]", "assert", "(", "wmat", ".", "flags", "[", "'C_CONTIGUOUS'", "]", "is", "True", ")", "sys", ".", "stdout", ".", "write", "(", "'converting layer {0}, wmat shape = {1}'", ".", "format", "(", "layer_name", ",", "wmat", ".", "shape", ")", ")", "if", "len", "(", "layer_blobs", ")", "==", "2", ":", "bias", "=", "np", ".", "array", "(", "layer_blobs", "[", "1", "]", ".", "data", ")", "bias", "=", "bias", ".", "reshape", "(", "(", "bias", ".", "shape", "[", "0", "]", ",", "1", ")", ")", "assert", "(", "bias", ".", "flags", "[", "'C_CONTIGUOUS'", "]", "is", "True", ")", "bias_name", "=", "layer_name", "+", "\"_bias\"", "if", "bias_name", "not", "in", "arg_shape_dic", ":", "print", "(", "bias_name", "+", "' not found in arg_shape_dic.'", ")", "continue", "bias", "=", "bias", ".", "reshape", "(", "arg_shape_dic", "[", "bias_name", "]", ")", "arg_params", "[", "bias_name", "]", "=", "mx", ".", "nd", ".", "zeros", "(", "bias", ".", "shape", ")", "arg_params", "[", "bias_name", "]", "[", ":", "]", "=", "bias", "sys", ".", "stdout", ".", "write", "(", "', bias shape = {}'", ".", "format", "(", "bias", ".", "shape", ")", ")", "sys", ".", "stdout", ".", "write", "(", "'\\n'", ")", "sys", ".", "stdout", ".", "flush", "(", ")", "wmat", "=", "wmat", ".", "reshape", "(", "(", "wmat", ".", "shape", "[", "0", "]", ",", "-", "1", ")", ")", "weight_name", "=", "layer_name", "+", "\"_weight\"", "if", "weight_name", "not", "in", "arg_shape_dic", ":", "print", "(", "weight_name", "+", "' not found in arg_shape_dic.'", ")", "continue", "wmat", "=", "wmat", ".", "reshape", "(", "arg_shape_dic", "[", "weight_name", "]", ")", "arg_params", "[", "weight_name", "]", "=", "mx", ".", "nd", ".", "zeros", "(", "wmat", ".", "shape", ")", "arg_params", "[", "weight_name", "]", "[", ":", "]", "=", "wmat", "if", "first_conv", "and", "(", "layer_type", "==", "'Convolution'", "or", "layer_type", "==", "4", ")", ":", "first_conv", "=", "False", "elif", "layer_type", "==", "'Scale'", ":", "if", "'scale'", "in", "layer_name", ":", "bn_name", "=", "layer_name", ".", "replace", "(", "'scale'", ",", "'bn'", ")", "elif", "'sc'", "in", "layer_name", ":", "bn_name", "=", "layer_name", ".", "replace", "(", "'sc'", ",", "'bn'", ")", "else", ":", "assert", "False", ",", "'Unknown name convention for bn/scale'", "gamma", "=", "np", ".", "array", "(", "layer_blobs", "[", "0", "]", ".", "data", ")", "beta", "=", "np", ".", "array", "(", "layer_blobs", "[", "1", "]", ".", "data", ")", "# beta = np.expand_dims(beta, 1)", "beta_name", "=", "'{}_beta'", ".", "format", "(", "bn_name", ")", "gamma_name", "=", "'{}_gamma'", ".", "format", "(", "bn_name", ")", "beta", "=", "beta", ".", "reshape", "(", "arg_shape_dic", "[", "beta_name", "]", ")", "gamma", "=", "gamma", ".", "reshape", "(", "arg_shape_dic", "[", "gamma_name", "]", ")", "arg_params", "[", "beta_name", "]", "=", "mx", ".", "nd", ".", "zeros", "(", "beta", ".", "shape", ")", "arg_params", "[", "gamma_name", "]", "=", "mx", ".", "nd", ".", "zeros", "(", "gamma", ".", "shape", ")", "arg_params", "[", "beta_name", "]", "[", ":", "]", "=", "beta", "arg_params", "[", "gamma_name", "]", "[", ":", "]", "=", "gamma", "assert", "gamma", ".", "flags", "[", "'C_CONTIGUOUS'", "]", "is", "True", "assert", "beta", ".", "flags", "[", "'C_CONTIGUOUS'", "]", "is", "True", "print", "(", "'converting scale layer, beta shape = {}, gamma shape = {}'", ".", "format", "(", "beta", ".", "shape", ",", "gamma", ".", "shape", ")", ")", "elif", "layer_type", "==", "'BatchNorm'", ":", "bn_name", "=", "layer_name", "mean", "=", "np", ".", "array", "(", "layer_blobs", "[", "0", "]", ".", "data", ")", "var", "=", "np", ".", "array", "(", "layer_blobs", "[", "1", "]", ".", "data", ")", "rescale_factor", "=", "layer_blobs", "[", "2", "]", ".", "data", "[", "0", "]", "if", "rescale_factor", "!=", "0", ":", "rescale_factor", "=", "1", "/", "rescale_factor", "mean_name", "=", "'{}_moving_mean'", ".", "format", "(", "bn_name", ")", "var_name", "=", "'{}_moving_var'", ".", "format", "(", "bn_name", ")", "mean", "=", "mean", ".", "reshape", "(", "aux_shape_dic", "[", "mean_name", "]", ")", "var", "=", "var", ".", "reshape", "(", "aux_shape_dic", "[", "var_name", "]", ")", "aux_params", "[", "mean_name", "]", "=", "mx", ".", "nd", ".", "zeros", "(", "mean", ".", "shape", ")", "aux_params", "[", "var_name", "]", "=", "mx", ".", "nd", ".", "zeros", "(", "var", ".", "shape", ")", "# Get the original epsilon", "for", "idx", ",", "layer", "in", "enumerate", "(", "layers_proto", ")", ":", "if", "layer", ".", "name", "==", "bn_name", ":", "bn_index", "=", "idx", "eps_caffe", "=", "layers_proto", "[", "bn_index", "]", ".", "batch_norm_param", ".", "eps", "# Compensate for the epsilon shift performed in convert_symbol", "eps_symbol", "=", "float", "(", "sym", ".", "attr_dict", "(", ")", "[", "bn_name", "+", "'_moving_mean'", "]", "[", "'eps'", "]", ")", "eps_correction", "=", "eps_caffe", "-", "eps_symbol", "# Fill parameters", "aux_params", "[", "mean_name", "]", "[", ":", "]", "=", "mean", "*", "rescale_factor", "aux_params", "[", "var_name", "]", "[", ":", "]", "=", "var", "*", "rescale_factor", "+", "eps_correction", "assert", "var", ".", "flags", "[", "'C_CONTIGUOUS'", "]", "is", "True", "assert", "mean", ".", "flags", "[", "'C_CONTIGUOUS'", "]", "is", "True", "print", "(", "'converting batchnorm layer, mean shape = {}, var shape = {}'", ".", "format", "(", "mean", ".", "shape", ",", "var", ".", "shape", ")", ")", "fix_gamma", "=", "layers_proto", "[", "bn_index", "+", "1", "]", ".", "type", "!=", "'Scale'", "if", "fix_gamma", ":", "gamma_name", "=", "'{}_gamma'", ".", "format", "(", "bn_name", ")", "gamma", "=", "np", ".", "array", "(", "np", ".", "ones", "(", "arg_shape_dic", "[", "gamma_name", "]", ")", ")", "beta_name", "=", "'{}_beta'", ".", "format", "(", "bn_name", ")", "beta", "=", "np", ".", "array", "(", "np", ".", "zeros", "(", "arg_shape_dic", "[", "beta_name", "]", ")", ")", "arg_params", "[", "beta_name", "]", "=", "mx", ".", "nd", ".", "zeros", "(", "beta", ".", "shape", ")", "arg_params", "[", "gamma_name", "]", "=", "mx", ".", "nd", ".", "zeros", "(", "gamma", ".", "shape", ")", "arg_params", "[", "beta_name", "]", "[", ":", "]", "=", "beta", "arg_params", "[", "gamma_name", "]", "[", ":", "]", "=", "gamma", "assert", "gamma", ".", "flags", "[", "'C_CONTIGUOUS'", "]", "is", "True", "assert", "beta", ".", "flags", "[", "'C_CONTIGUOUS'", "]", "is", "True", "else", ":", "print", "(", "'\\tskipping layer {} of type {}'", ".", "format", "(", "layer_name", ",", "layer_type", ")", ")", "assert", "len", "(", "layer_blobs", ")", "==", "0", "if", "output_prefix", "is", "not", "None", ":", "model", "=", "mx", ".", "mod", ".", "Module", "(", "symbol", "=", "sym", ",", "label_names", "=", "None", ")", "model", ".", "bind", "(", "data_shapes", "=", "[", "(", "'data'", ",", "tuple", "(", "input_dim", ")", ")", "]", ")", "model", ".", "init_params", "(", "arg_params", "=", "arg_params", ",", "aux_params", "=", "aux_params", ")", "model", ".", "save_checkpoint", "(", "output_prefix", ",", "0", ")", "return", "sym", ",", "arg_params", ",", "aux_params", ",", "input_dim" ]
Convert caffe model Parameters ---------- prototxt_fname : str Filename of the prototxt model definition caffemodel_fname : str Filename of the binary caffe model output_prefix : str, optinoal If given, then save the converted MXNet into output_prefx+'.json' and output_prefx+'.params' Returns ------- sym : Symbol Symbol convereted from prototxt arg_params : list of NDArray Argument parameters aux_params : list of NDArray Aux parameters input_dim : tuple Input dimension
[ "Convert", "caffe", "model" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ssd/tools/caffe_converter/convert_model.py#L26-L210
train
apache/incubator-mxnet
tools/caffe_converter/convert_symbol.py
_parse_proto
def _parse_proto(prototxt_fname): """Parse Caffe prototxt into symbol string """ proto = caffe_parser.read_prototxt(prototxt_fname) # process data layer input_name, input_dim, layers = _get_input(proto) # only support single input, so always use `data` as the input data mapping = {input_name: 'data'} need_flatten = {input_name: False} symbol_string = "import mxnet as mx\ndata = mx.symbol.Variable(name='data')\n" flatten_count = 0 output_name = "" prev_name = None _output_name = {} # convert reset layers one by one for i, layer in enumerate(layers): type_string = '' param_string = '' skip_layer = False name = re.sub('[-/]', '_', layer.name) for k in range(len(layer.bottom)): if layer.bottom[k] in _output_name: _output_name[layer.bottom[k]]['count'] = _output_name[layer.bottom[k]]['count']+1 else: _output_name[layer.bottom[k]] = {'count':0} for k in range(len(layer.top)): if layer.top[k] in _output_name: _output_name[layer.top[k]]['count'] = _output_name[layer.top[k]]['count']+1 else: _output_name[layer.top[k]] = {'count':0, 'name':name} if layer.type == 'Convolution' or layer.type == 4: type_string = 'mx.symbol.Convolution' param_string = _convert_conv_param(layer.convolution_param) need_flatten[name] = True if layer.type == 'Deconvolution' or layer.type == 39: type_string = 'mx.symbol.Deconvolution' param_string = _convert_conv_param(layer.convolution_param) need_flatten[name] = True if layer.type == 'Pooling' or layer.type == 17: type_string = 'mx.symbol.Pooling' param_string = _convert_pooling_param(layer.pooling_param) need_flatten[name] = True if layer.type == 'ReLU' or layer.type == 18: type_string = 'mx.symbol.Activation' param_string = "act_type='relu'" param = layer.relu_param if hasattr(param, 'negative_slope'): if param.negative_slope > 0: type_string = 'mx.symbol.LeakyReLU' param_string = "act_type='leaky', slope=%f" % param.negative_slope need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'TanH' or layer.type == 23: type_string = 'mx.symbol.Activation' param_string = "act_type='tanh'" need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'Sigmoid' or layer.type == 19: type_string = 'mx.symbol.Activation' param_string = "act_type='sigmoid'" need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'LRN' or layer.type == 15: type_string = 'mx.symbol.LRN' param = layer.lrn_param param_string = "alpha=%f, beta=%f, knorm=%f, nsize=%d" % ( param.alpha, param.beta, param.k, param.local_size) need_flatten[name] = True if layer.type == 'InnerProduct' or layer.type == 14: type_string = 'mx.symbol.FullyConnected' param = layer.inner_product_param param_string = "num_hidden=%d, no_bias=%s" % ( param.num_output, not param.bias_term) need_flatten[name] = False if layer.type == 'Dropout' or layer.type == 6: type_string = 'mx.symbol.Dropout' param = layer.dropout_param param_string = "p=%f" % param.dropout_ratio need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'Softmax' or layer.type == 20: type_string = 'mx.symbol.SoftmaxOutput' if layer.type == 'Flatten' or layer.type == 8: type_string = 'mx.symbol.Flatten' need_flatten[name] = False if layer.type == 'Split' or layer.type == 22: type_string = 'split' # will process later if layer.type == 'Concat' or layer.type == 3: type_string = 'mx.symbol.Concat' need_flatten[name] = True if layer.type == 'Crop': type_string = 'mx.symbol.Crop' need_flatten[name] = True param_string = 'center_crop=True' if layer.type == 'BatchNorm': type_string = 'mx.symbol.BatchNorm' param = layer.batch_norm_param # CuDNN requires eps to be greater than 1e-05 # We compensate for this change in convert_model epsilon = param.eps if (epsilon <= 1e-05): epsilon = 1e-04 # if next layer is scale, don't fix gamma fix_gamma = layers[i+1].type != 'Scale' param_string = 'use_global_stats=%s, fix_gamma=%s, eps=%f' % ( param.use_global_stats, fix_gamma, epsilon) need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'Scale': assert layers[i-1].type == 'BatchNorm' need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] skip_layer = True prev_name = re.sub('[-/]', '_', layers[i-1].name) if layer.type == 'PReLU': type_string = 'mx.symbol.LeakyReLU' param = layer.prelu_param param_string = "act_type='prelu', slope=%f" % param.filler.value need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'Eltwise': type_string = 'mx.symbol.broadcast_add' param = layer.eltwise_param param_string = "" need_flatten[name] = False if layer.type == 'Reshape': type_string = 'mx.symbol.Reshape' need_flatten[name] = False param = layer.reshape_param param_string = "shape=(%s)" % (','.join(param.shape.dim),) if layer.type == 'AbsVal': type_string = 'mx.symbol.abs' need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if skip_layer: assert len(layer.bottom) == 1 symbol_string += "%s = %s\n" % (name, prev_name) elif type_string == '': raise ValueError('Unknown layer %s!' % layer.type) elif type_string != 'split': bottom = layer.bottom if param_string != "": param_string = ", " + param_string if len(bottom) == 1: if need_flatten[mapping[bottom[0]]] and type_string == 'mx.symbol.FullyConnected': flatten_name = "flatten_%d" % flatten_count symbol_string += "%s=mx.symbol.Flatten(name='%s', data=%s)\n" % ( flatten_name, flatten_name, mapping[bottom[0]]) flatten_count += 1 need_flatten[flatten_name] = False bottom[0] = flatten_name mapping[bottom[0]] = bottom[0] symbol_string += "%s = %s(name='%s', data=%s %s)\n" % ( name, type_string, name, mapping[bottom[0]], param_string) else: if layer.type == 'Eltwise' and param.operation == 1 and len(param.coeff) > 0: symbol_string += "%s = " % name symbol_string += " + ".join(["%s * %s" % ( mapping[bottom[i]], param.coeff[i]) for i in range(len(param.coeff))]) symbol_string += "\n" else: symbol_string += "%s = %s(name='%s', *[%s] %s)\n" % ( name, type_string, name, ','.join( [mapping[x] for x in bottom]), param_string) for j in range(len(layer.top)): mapping[layer.top[j]] = name output_name = name output_name = [] for i in _output_name: if 'name' in _output_name[i] and _output_name[i]['count'] == 0: output_name.append(_output_name[i]['name']) return symbol_string, output_name, input_dim
python
def _parse_proto(prototxt_fname): """Parse Caffe prototxt into symbol string """ proto = caffe_parser.read_prototxt(prototxt_fname) # process data layer input_name, input_dim, layers = _get_input(proto) # only support single input, so always use `data` as the input data mapping = {input_name: 'data'} need_flatten = {input_name: False} symbol_string = "import mxnet as mx\ndata = mx.symbol.Variable(name='data')\n" flatten_count = 0 output_name = "" prev_name = None _output_name = {} # convert reset layers one by one for i, layer in enumerate(layers): type_string = '' param_string = '' skip_layer = False name = re.sub('[-/]', '_', layer.name) for k in range(len(layer.bottom)): if layer.bottom[k] in _output_name: _output_name[layer.bottom[k]]['count'] = _output_name[layer.bottom[k]]['count']+1 else: _output_name[layer.bottom[k]] = {'count':0} for k in range(len(layer.top)): if layer.top[k] in _output_name: _output_name[layer.top[k]]['count'] = _output_name[layer.top[k]]['count']+1 else: _output_name[layer.top[k]] = {'count':0, 'name':name} if layer.type == 'Convolution' or layer.type == 4: type_string = 'mx.symbol.Convolution' param_string = _convert_conv_param(layer.convolution_param) need_flatten[name] = True if layer.type == 'Deconvolution' or layer.type == 39: type_string = 'mx.symbol.Deconvolution' param_string = _convert_conv_param(layer.convolution_param) need_flatten[name] = True if layer.type == 'Pooling' or layer.type == 17: type_string = 'mx.symbol.Pooling' param_string = _convert_pooling_param(layer.pooling_param) need_flatten[name] = True if layer.type == 'ReLU' or layer.type == 18: type_string = 'mx.symbol.Activation' param_string = "act_type='relu'" param = layer.relu_param if hasattr(param, 'negative_slope'): if param.negative_slope > 0: type_string = 'mx.symbol.LeakyReLU' param_string = "act_type='leaky', slope=%f" % param.negative_slope need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'TanH' or layer.type == 23: type_string = 'mx.symbol.Activation' param_string = "act_type='tanh'" need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'Sigmoid' or layer.type == 19: type_string = 'mx.symbol.Activation' param_string = "act_type='sigmoid'" need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'LRN' or layer.type == 15: type_string = 'mx.symbol.LRN' param = layer.lrn_param param_string = "alpha=%f, beta=%f, knorm=%f, nsize=%d" % ( param.alpha, param.beta, param.k, param.local_size) need_flatten[name] = True if layer.type == 'InnerProduct' or layer.type == 14: type_string = 'mx.symbol.FullyConnected' param = layer.inner_product_param param_string = "num_hidden=%d, no_bias=%s" % ( param.num_output, not param.bias_term) need_flatten[name] = False if layer.type == 'Dropout' or layer.type == 6: type_string = 'mx.symbol.Dropout' param = layer.dropout_param param_string = "p=%f" % param.dropout_ratio need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'Softmax' or layer.type == 20: type_string = 'mx.symbol.SoftmaxOutput' if layer.type == 'Flatten' or layer.type == 8: type_string = 'mx.symbol.Flatten' need_flatten[name] = False if layer.type == 'Split' or layer.type == 22: type_string = 'split' # will process later if layer.type == 'Concat' or layer.type == 3: type_string = 'mx.symbol.Concat' need_flatten[name] = True if layer.type == 'Crop': type_string = 'mx.symbol.Crop' need_flatten[name] = True param_string = 'center_crop=True' if layer.type == 'BatchNorm': type_string = 'mx.symbol.BatchNorm' param = layer.batch_norm_param # CuDNN requires eps to be greater than 1e-05 # We compensate for this change in convert_model epsilon = param.eps if (epsilon <= 1e-05): epsilon = 1e-04 # if next layer is scale, don't fix gamma fix_gamma = layers[i+1].type != 'Scale' param_string = 'use_global_stats=%s, fix_gamma=%s, eps=%f' % ( param.use_global_stats, fix_gamma, epsilon) need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'Scale': assert layers[i-1].type == 'BatchNorm' need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] skip_layer = True prev_name = re.sub('[-/]', '_', layers[i-1].name) if layer.type == 'PReLU': type_string = 'mx.symbol.LeakyReLU' param = layer.prelu_param param_string = "act_type='prelu', slope=%f" % param.filler.value need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if layer.type == 'Eltwise': type_string = 'mx.symbol.broadcast_add' param = layer.eltwise_param param_string = "" need_flatten[name] = False if layer.type == 'Reshape': type_string = 'mx.symbol.Reshape' need_flatten[name] = False param = layer.reshape_param param_string = "shape=(%s)" % (','.join(param.shape.dim),) if layer.type == 'AbsVal': type_string = 'mx.symbol.abs' need_flatten[name] = need_flatten[mapping[layer.bottom[0]]] if skip_layer: assert len(layer.bottom) == 1 symbol_string += "%s = %s\n" % (name, prev_name) elif type_string == '': raise ValueError('Unknown layer %s!' % layer.type) elif type_string != 'split': bottom = layer.bottom if param_string != "": param_string = ", " + param_string if len(bottom) == 1: if need_flatten[mapping[bottom[0]]] and type_string == 'mx.symbol.FullyConnected': flatten_name = "flatten_%d" % flatten_count symbol_string += "%s=mx.symbol.Flatten(name='%s', data=%s)\n" % ( flatten_name, flatten_name, mapping[bottom[0]]) flatten_count += 1 need_flatten[flatten_name] = False bottom[0] = flatten_name mapping[bottom[0]] = bottom[0] symbol_string += "%s = %s(name='%s', data=%s %s)\n" % ( name, type_string, name, mapping[bottom[0]], param_string) else: if layer.type == 'Eltwise' and param.operation == 1 and len(param.coeff) > 0: symbol_string += "%s = " % name symbol_string += " + ".join(["%s * %s" % ( mapping[bottom[i]], param.coeff[i]) for i in range(len(param.coeff))]) symbol_string += "\n" else: symbol_string += "%s = %s(name='%s', *[%s] %s)\n" % ( name, type_string, name, ','.join( [mapping[x] for x in bottom]), param_string) for j in range(len(layer.top)): mapping[layer.top[j]] = name output_name = name output_name = [] for i in _output_name: if 'name' in _output_name[i] and _output_name[i]['count'] == 0: output_name.append(_output_name[i]['name']) return symbol_string, output_name, input_dim
[ "def", "_parse_proto", "(", "prototxt_fname", ")", ":", "proto", "=", "caffe_parser", ".", "read_prototxt", "(", "prototxt_fname", ")", "# process data layer", "input_name", ",", "input_dim", ",", "layers", "=", "_get_input", "(", "proto", ")", "# only support single input, so always use `data` as the input data", "mapping", "=", "{", "input_name", ":", "'data'", "}", "need_flatten", "=", "{", "input_name", ":", "False", "}", "symbol_string", "=", "\"import mxnet as mx\\ndata = mx.symbol.Variable(name='data')\\n\"", "flatten_count", "=", "0", "output_name", "=", "\"\"", "prev_name", "=", "None", "_output_name", "=", "{", "}", "# convert reset layers one by one", "for", "i", ",", "layer", "in", "enumerate", "(", "layers", ")", ":", "type_string", "=", "''", "param_string", "=", "''", "skip_layer", "=", "False", "name", "=", "re", ".", "sub", "(", "'[-/]'", ",", "'_'", ",", "layer", ".", "name", ")", "for", "k", "in", "range", "(", "len", "(", "layer", ".", "bottom", ")", ")", ":", "if", "layer", ".", "bottom", "[", "k", "]", "in", "_output_name", ":", "_output_name", "[", "layer", ".", "bottom", "[", "k", "]", "]", "[", "'count'", "]", "=", "_output_name", "[", "layer", ".", "bottom", "[", "k", "]", "]", "[", "'count'", "]", "+", "1", "else", ":", "_output_name", "[", "layer", ".", "bottom", "[", "k", "]", "]", "=", "{", "'count'", ":", "0", "}", "for", "k", "in", "range", "(", "len", "(", "layer", ".", "top", ")", ")", ":", "if", "layer", ".", "top", "[", "k", "]", "in", "_output_name", ":", "_output_name", "[", "layer", ".", "top", "[", "k", "]", "]", "[", "'count'", "]", "=", "_output_name", "[", "layer", ".", "top", "[", "k", "]", "]", "[", "'count'", "]", "+", "1", "else", ":", "_output_name", "[", "layer", ".", "top", "[", "k", "]", "]", "=", "{", "'count'", ":", "0", ",", "'name'", ":", "name", "}", "if", "layer", ".", "type", "==", "'Convolution'", "or", "layer", ".", "type", "==", "4", ":", "type_string", "=", "'mx.symbol.Convolution'", "param_string", "=", "_convert_conv_param", "(", "layer", ".", "convolution_param", ")", "need_flatten", "[", "name", "]", "=", "True", "if", "layer", ".", "type", "==", "'Deconvolution'", "or", "layer", ".", "type", "==", "39", ":", "type_string", "=", "'mx.symbol.Deconvolution'", "param_string", "=", "_convert_conv_param", "(", "layer", ".", "convolution_param", ")", "need_flatten", "[", "name", "]", "=", "True", "if", "layer", ".", "type", "==", "'Pooling'", "or", "layer", ".", "type", "==", "17", ":", "type_string", "=", "'mx.symbol.Pooling'", "param_string", "=", "_convert_pooling_param", "(", "layer", ".", "pooling_param", ")", "need_flatten", "[", "name", "]", "=", "True", "if", "layer", ".", "type", "==", "'ReLU'", "or", "layer", ".", "type", "==", "18", ":", "type_string", "=", "'mx.symbol.Activation'", "param_string", "=", "\"act_type='relu'\"", "param", "=", "layer", ".", "relu_param", "if", "hasattr", "(", "param", ",", "'negative_slope'", ")", ":", "if", "param", ".", "negative_slope", ">", "0", ":", "type_string", "=", "'mx.symbol.LeakyReLU'", "param_string", "=", "\"act_type='leaky', slope=%f\"", "%", "param", ".", "negative_slope", "need_flatten", "[", "name", "]", "=", "need_flatten", "[", "mapping", "[", "layer", ".", "bottom", "[", "0", "]", "]", "]", "if", "layer", ".", "type", "==", "'TanH'", "or", "layer", ".", "type", "==", "23", ":", "type_string", "=", "'mx.symbol.Activation'", "param_string", "=", "\"act_type='tanh'\"", "need_flatten", "[", "name", "]", "=", "need_flatten", "[", "mapping", "[", "layer", ".", "bottom", "[", "0", "]", "]", "]", "if", "layer", ".", "type", "==", "'Sigmoid'", "or", "layer", ".", "type", "==", "19", ":", "type_string", "=", "'mx.symbol.Activation'", "param_string", "=", "\"act_type='sigmoid'\"", "need_flatten", "[", "name", "]", "=", "need_flatten", "[", "mapping", "[", "layer", ".", "bottom", "[", "0", "]", "]", "]", "if", "layer", ".", "type", "==", "'LRN'", "or", "layer", ".", "type", "==", "15", ":", "type_string", "=", "'mx.symbol.LRN'", "param", "=", "layer", ".", "lrn_param", "param_string", "=", "\"alpha=%f, beta=%f, knorm=%f, nsize=%d\"", "%", "(", "param", ".", "alpha", ",", "param", ".", "beta", ",", "param", ".", "k", ",", "param", ".", "local_size", ")", "need_flatten", "[", "name", "]", "=", "True", "if", "layer", ".", "type", "==", "'InnerProduct'", "or", "layer", ".", "type", "==", "14", ":", "type_string", "=", "'mx.symbol.FullyConnected'", "param", "=", "layer", ".", "inner_product_param", "param_string", "=", "\"num_hidden=%d, no_bias=%s\"", "%", "(", "param", ".", "num_output", ",", "not", "param", ".", "bias_term", ")", "need_flatten", "[", "name", "]", "=", "False", "if", "layer", ".", "type", "==", "'Dropout'", "or", "layer", ".", "type", "==", "6", ":", "type_string", "=", "'mx.symbol.Dropout'", "param", "=", "layer", ".", "dropout_param", "param_string", "=", "\"p=%f\"", "%", "param", ".", "dropout_ratio", "need_flatten", "[", "name", "]", "=", "need_flatten", "[", "mapping", "[", "layer", ".", "bottom", "[", "0", "]", "]", "]", "if", "layer", ".", "type", "==", "'Softmax'", "or", "layer", ".", "type", "==", "20", ":", "type_string", "=", "'mx.symbol.SoftmaxOutput'", "if", "layer", ".", "type", "==", "'Flatten'", "or", "layer", ".", "type", "==", "8", ":", "type_string", "=", "'mx.symbol.Flatten'", "need_flatten", "[", "name", "]", "=", "False", "if", "layer", ".", "type", "==", "'Split'", "or", "layer", ".", "type", "==", "22", ":", "type_string", "=", "'split'", "# will process later", "if", "layer", ".", "type", "==", "'Concat'", "or", "layer", ".", "type", "==", "3", ":", "type_string", "=", "'mx.symbol.Concat'", "need_flatten", "[", "name", "]", "=", "True", "if", "layer", ".", "type", "==", "'Crop'", ":", "type_string", "=", "'mx.symbol.Crop'", "need_flatten", "[", "name", "]", "=", "True", "param_string", "=", "'center_crop=True'", "if", "layer", ".", "type", "==", "'BatchNorm'", ":", "type_string", "=", "'mx.symbol.BatchNorm'", "param", "=", "layer", ".", "batch_norm_param", "# CuDNN requires eps to be greater than 1e-05", "# We compensate for this change in convert_model", "epsilon", "=", "param", ".", "eps", "if", "(", "epsilon", "<=", "1e-05", ")", ":", "epsilon", "=", "1e-04", "# if next layer is scale, don't fix gamma", "fix_gamma", "=", "layers", "[", "i", "+", "1", "]", ".", "type", "!=", "'Scale'", "param_string", "=", "'use_global_stats=%s, fix_gamma=%s, eps=%f'", "%", "(", "param", ".", "use_global_stats", ",", "fix_gamma", ",", "epsilon", ")", "need_flatten", "[", "name", "]", "=", "need_flatten", "[", "mapping", "[", "layer", ".", "bottom", "[", "0", "]", "]", "]", "if", "layer", ".", "type", "==", "'Scale'", ":", "assert", "layers", "[", "i", "-", "1", "]", ".", "type", "==", "'BatchNorm'", "need_flatten", "[", "name", "]", "=", "need_flatten", "[", "mapping", "[", "layer", ".", "bottom", "[", "0", "]", "]", "]", "skip_layer", "=", "True", "prev_name", "=", "re", ".", "sub", "(", "'[-/]'", ",", "'_'", ",", "layers", "[", "i", "-", "1", "]", ".", "name", ")", "if", "layer", ".", "type", "==", "'PReLU'", ":", "type_string", "=", "'mx.symbol.LeakyReLU'", "param", "=", "layer", ".", "prelu_param", "param_string", "=", "\"act_type='prelu', slope=%f\"", "%", "param", ".", "filler", ".", "value", "need_flatten", "[", "name", "]", "=", "need_flatten", "[", "mapping", "[", "layer", ".", "bottom", "[", "0", "]", "]", "]", "if", "layer", ".", "type", "==", "'Eltwise'", ":", "type_string", "=", "'mx.symbol.broadcast_add'", "param", "=", "layer", ".", "eltwise_param", "param_string", "=", "\"\"", "need_flatten", "[", "name", "]", "=", "False", "if", "layer", ".", "type", "==", "'Reshape'", ":", "type_string", "=", "'mx.symbol.Reshape'", "need_flatten", "[", "name", "]", "=", "False", "param", "=", "layer", ".", "reshape_param", "param_string", "=", "\"shape=(%s)\"", "%", "(", "','", ".", "join", "(", "param", ".", "shape", ".", "dim", ")", ",", ")", "if", "layer", ".", "type", "==", "'AbsVal'", ":", "type_string", "=", "'mx.symbol.abs'", "need_flatten", "[", "name", "]", "=", "need_flatten", "[", "mapping", "[", "layer", ".", "bottom", "[", "0", "]", "]", "]", "if", "skip_layer", ":", "assert", "len", "(", "layer", ".", "bottom", ")", "==", "1", "symbol_string", "+=", "\"%s = %s\\n\"", "%", "(", "name", ",", "prev_name", ")", "elif", "type_string", "==", "''", ":", "raise", "ValueError", "(", "'Unknown layer %s!'", "%", "layer", ".", "type", ")", "elif", "type_string", "!=", "'split'", ":", "bottom", "=", "layer", ".", "bottom", "if", "param_string", "!=", "\"\"", ":", "param_string", "=", "\", \"", "+", "param_string", "if", "len", "(", "bottom", ")", "==", "1", ":", "if", "need_flatten", "[", "mapping", "[", "bottom", "[", "0", "]", "]", "]", "and", "type_string", "==", "'mx.symbol.FullyConnected'", ":", "flatten_name", "=", "\"flatten_%d\"", "%", "flatten_count", "symbol_string", "+=", "\"%s=mx.symbol.Flatten(name='%s', data=%s)\\n\"", "%", "(", "flatten_name", ",", "flatten_name", ",", "mapping", "[", "bottom", "[", "0", "]", "]", ")", "flatten_count", "+=", "1", "need_flatten", "[", "flatten_name", "]", "=", "False", "bottom", "[", "0", "]", "=", "flatten_name", "mapping", "[", "bottom", "[", "0", "]", "]", "=", "bottom", "[", "0", "]", "symbol_string", "+=", "\"%s = %s(name='%s', data=%s %s)\\n\"", "%", "(", "name", ",", "type_string", ",", "name", ",", "mapping", "[", "bottom", "[", "0", "]", "]", ",", "param_string", ")", "else", ":", "if", "layer", ".", "type", "==", "'Eltwise'", "and", "param", ".", "operation", "==", "1", "and", "len", "(", "param", ".", "coeff", ")", ">", "0", ":", "symbol_string", "+=", "\"%s = \"", "%", "name", "symbol_string", "+=", "\" + \"", ".", "join", "(", "[", "\"%s * %s\"", "%", "(", "mapping", "[", "bottom", "[", "i", "]", "]", ",", "param", ".", "coeff", "[", "i", "]", ")", "for", "i", "in", "range", "(", "len", "(", "param", ".", "coeff", ")", ")", "]", ")", "symbol_string", "+=", "\"\\n\"", "else", ":", "symbol_string", "+=", "\"%s = %s(name='%s', *[%s] %s)\\n\"", "%", "(", "name", ",", "type_string", ",", "name", ",", "','", ".", "join", "(", "[", "mapping", "[", "x", "]", "for", "x", "in", "bottom", "]", ")", ",", "param_string", ")", "for", "j", "in", "range", "(", "len", "(", "layer", ".", "top", ")", ")", ":", "mapping", "[", "layer", ".", "top", "[", "j", "]", "]", "=", "name", "output_name", "=", "name", "output_name", "=", "[", "]", "for", "i", "in", "_output_name", ":", "if", "'name'", "in", "_output_name", "[", "i", "]", "and", "_output_name", "[", "i", "]", "[", "'count'", "]", "==", "0", ":", "output_name", ".", "append", "(", "_output_name", "[", "i", "]", "[", "'name'", "]", ")", "return", "symbol_string", ",", "output_name", ",", "input_dim" ]
Parse Caffe prototxt into symbol string
[ "Parse", "Caffe", "prototxt", "into", "symbol", "string" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/tools/caffe_converter/convert_symbol.py#L127-L295
train
apache/incubator-mxnet
tools/caffe_converter/convert_symbol.py
convert_symbol
def convert_symbol(prototxt_fname): """Convert caffe model definition into Symbol Parameters ---------- prototxt_fname : str Filename of the prototxt file Returns ------- Symbol Converted Symbol tuple Input shape """ sym, output_name, input_dim = _parse_proto(prototxt_fname) exec(sym) # pylint: disable=exec-used _locals = locals() ret = [] for i in output_name: exec("ret = " + i, globals(), _locals) # pylint: disable=exec-used ret.append(_locals['ret']) ret = mx.sym.Group(ret) return ret, input_dim
python
def convert_symbol(prototxt_fname): """Convert caffe model definition into Symbol Parameters ---------- prototxt_fname : str Filename of the prototxt file Returns ------- Symbol Converted Symbol tuple Input shape """ sym, output_name, input_dim = _parse_proto(prototxt_fname) exec(sym) # pylint: disable=exec-used _locals = locals() ret = [] for i in output_name: exec("ret = " + i, globals(), _locals) # pylint: disable=exec-used ret.append(_locals['ret']) ret = mx.sym.Group(ret) return ret, input_dim
[ "def", "convert_symbol", "(", "prototxt_fname", ")", ":", "sym", ",", "output_name", ",", "input_dim", "=", "_parse_proto", "(", "prototxt_fname", ")", "exec", "(", "sym", ")", "# pylint: disable=exec-used", "_locals", "=", "locals", "(", ")", "ret", "=", "[", "]", "for", "i", "in", "output_name", ":", "exec", "(", "\"ret = \"", "+", "i", ",", "globals", "(", ")", ",", "_locals", ")", "# pylint: disable=exec-used", "ret", ".", "append", "(", "_locals", "[", "'ret'", "]", ")", "ret", "=", "mx", ".", "sym", ".", "Group", "(", "ret", ")", "return", "ret", ",", "input_dim" ]
Convert caffe model definition into Symbol Parameters ---------- prototxt_fname : str Filename of the prototxt file Returns ------- Symbol Converted Symbol tuple Input shape
[ "Convert", "caffe", "model", "definition", "into", "Symbol" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/tools/caffe_converter/convert_symbol.py#L297-L320
train
apache/incubator-mxnet
python/mxnet/gluon/model_zoo/vision/vgg.py
get_vgg
def get_vgg(num_layers, pretrained=False, ctx=cpu(), root=os.path.join(base.data_dir(), 'models'), **kwargs): r"""VGG model from the `"Very Deep Convolutional Networks for Large-Scale Image Recognition" <https://arxiv.org/abs/1409.1556>`_ paper. Parameters ---------- num_layers : int Number of layers for the variant of densenet. Options are 11, 13, 16, 19. pretrained : bool, default False Whether to load the pretrained weights for model. ctx : Context, default CPU The context in which to load the pretrained weights. root : str, default $MXNET_HOME/models Location for keeping the model parameters. """ layers, filters = vgg_spec[num_layers] net = VGG(layers, filters, **kwargs) if pretrained: from ..model_store import get_model_file batch_norm_suffix = '_bn' if kwargs.get('batch_norm') else '' net.load_parameters(get_model_file('vgg%d%s'%(num_layers, batch_norm_suffix), root=root), ctx=ctx) return net
python
def get_vgg(num_layers, pretrained=False, ctx=cpu(), root=os.path.join(base.data_dir(), 'models'), **kwargs): r"""VGG model from the `"Very Deep Convolutional Networks for Large-Scale Image Recognition" <https://arxiv.org/abs/1409.1556>`_ paper. Parameters ---------- num_layers : int Number of layers for the variant of densenet. Options are 11, 13, 16, 19. pretrained : bool, default False Whether to load the pretrained weights for model. ctx : Context, default CPU The context in which to load the pretrained weights. root : str, default $MXNET_HOME/models Location for keeping the model parameters. """ layers, filters = vgg_spec[num_layers] net = VGG(layers, filters, **kwargs) if pretrained: from ..model_store import get_model_file batch_norm_suffix = '_bn' if kwargs.get('batch_norm') else '' net.load_parameters(get_model_file('vgg%d%s'%(num_layers, batch_norm_suffix), root=root), ctx=ctx) return net
[ "def", "get_vgg", "(", "num_layers", ",", "pretrained", "=", "False", ",", "ctx", "=", "cpu", "(", ")", ",", "root", "=", "os", ".", "path", ".", "join", "(", "base", ".", "data_dir", "(", ")", ",", "'models'", ")", ",", "*", "*", "kwargs", ")", ":", "layers", ",", "filters", "=", "vgg_spec", "[", "num_layers", "]", "net", "=", "VGG", "(", "layers", ",", "filters", ",", "*", "*", "kwargs", ")", "if", "pretrained", ":", "from", ".", ".", "model_store", "import", "get_model_file", "batch_norm_suffix", "=", "'_bn'", "if", "kwargs", ".", "get", "(", "'batch_norm'", ")", "else", "''", "net", ".", "load_parameters", "(", "get_model_file", "(", "'vgg%d%s'", "%", "(", "num_layers", ",", "batch_norm_suffix", ")", ",", "root", "=", "root", ")", ",", "ctx", "=", "ctx", ")", "return", "net" ]
r"""VGG model from the `"Very Deep Convolutional Networks for Large-Scale Image Recognition" <https://arxiv.org/abs/1409.1556>`_ paper. Parameters ---------- num_layers : int Number of layers for the variant of densenet. Options are 11, 13, 16, 19. pretrained : bool, default False Whether to load the pretrained weights for model. ctx : Context, default CPU The context in which to load the pretrained weights. root : str, default $MXNET_HOME/models Location for keeping the model parameters.
[ "r", "VGG", "model", "from", "the", "Very", "Deep", "Convolutional", "Networks", "for", "Large", "-", "Scale", "Image", "Recognition", "<https", ":", "//", "arxiv", ".", "org", "/", "abs", "/", "1409", ".", "1556", ">", "_", "paper", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/model_zoo/vision/vgg.py#L97-L120
train
apache/incubator-mxnet
example/profiler/profiler_ndarray.py
check_with_uniform
def check_with_uniform(uf, arg_shapes, dim=None, npuf=None, rmin=-10, type_list=[np.float32]): """check function consistency with uniform random numbers""" if isinstance(arg_shapes, int): assert dim shape = tuple(np.random.randint(1, int(1000**(1.0/dim)), size=dim)) arg_shapes = [shape] * arg_shapes for dtype in type_list: ndarray_arg = [] numpy_arg = [] for s in arg_shapes: npy = np.random.uniform(rmin, 10, s).astype(dtype) narr = mx.nd.array(npy, dtype=dtype) ndarray_arg.append(narr) numpy_arg.append(npy) out1 = uf(*ndarray_arg) if npuf is None: out2 = uf(*numpy_arg).astype(dtype) else: out2 = npuf(*numpy_arg).astype(dtype) assert out1.shape == out2.shape if isinstance(out1, mx.nd.NDArray): out1 = out1.asnumpy() if dtype == np.float16: assert reldiff(out1, out2) < 2e-3 else: assert reldiff(out1, out2) < 1e-6
python
def check_with_uniform(uf, arg_shapes, dim=None, npuf=None, rmin=-10, type_list=[np.float32]): """check function consistency with uniform random numbers""" if isinstance(arg_shapes, int): assert dim shape = tuple(np.random.randint(1, int(1000**(1.0/dim)), size=dim)) arg_shapes = [shape] * arg_shapes for dtype in type_list: ndarray_arg = [] numpy_arg = [] for s in arg_shapes: npy = np.random.uniform(rmin, 10, s).astype(dtype) narr = mx.nd.array(npy, dtype=dtype) ndarray_arg.append(narr) numpy_arg.append(npy) out1 = uf(*ndarray_arg) if npuf is None: out2 = uf(*numpy_arg).astype(dtype) else: out2 = npuf(*numpy_arg).astype(dtype) assert out1.shape == out2.shape if isinstance(out1, mx.nd.NDArray): out1 = out1.asnumpy() if dtype == np.float16: assert reldiff(out1, out2) < 2e-3 else: assert reldiff(out1, out2) < 1e-6
[ "def", "check_with_uniform", "(", "uf", ",", "arg_shapes", ",", "dim", "=", "None", ",", "npuf", "=", "None", ",", "rmin", "=", "-", "10", ",", "type_list", "=", "[", "np", ".", "float32", "]", ")", ":", "if", "isinstance", "(", "arg_shapes", ",", "int", ")", ":", "assert", "dim", "shape", "=", "tuple", "(", "np", ".", "random", ".", "randint", "(", "1", ",", "int", "(", "1000", "**", "(", "1.0", "/", "dim", ")", ")", ",", "size", "=", "dim", ")", ")", "arg_shapes", "=", "[", "shape", "]", "*", "arg_shapes", "for", "dtype", "in", "type_list", ":", "ndarray_arg", "=", "[", "]", "numpy_arg", "=", "[", "]", "for", "s", "in", "arg_shapes", ":", "npy", "=", "np", ".", "random", ".", "uniform", "(", "rmin", ",", "10", ",", "s", ")", ".", "astype", "(", "dtype", ")", "narr", "=", "mx", ".", "nd", ".", "array", "(", "npy", ",", "dtype", "=", "dtype", ")", "ndarray_arg", ".", "append", "(", "narr", ")", "numpy_arg", ".", "append", "(", "npy", ")", "out1", "=", "uf", "(", "*", "ndarray_arg", ")", "if", "npuf", "is", "None", ":", "out2", "=", "uf", "(", "*", "numpy_arg", ")", ".", "astype", "(", "dtype", ")", "else", ":", "out2", "=", "npuf", "(", "*", "numpy_arg", ")", ".", "astype", "(", "dtype", ")", "assert", "out1", ".", "shape", "==", "out2", ".", "shape", "if", "isinstance", "(", "out1", ",", "mx", ".", "nd", ".", "NDArray", ")", ":", "out1", "=", "out1", ".", "asnumpy", "(", ")", "if", "dtype", "==", "np", ".", "float16", ":", "assert", "reldiff", "(", "out1", ",", "out2", ")", "<", "2e-3", "else", ":", "assert", "reldiff", "(", "out1", ",", "out2", ")", "<", "1e-6" ]
check function consistency with uniform random numbers
[ "check", "function", "consistency", "with", "uniform", "random", "numbers" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/profiler/profiler_ndarray.py#L51-L77
train
apache/incubator-mxnet
example/rcnn/symimdb/imdb.py
IMDB.filter_roidb
def filter_roidb(self): """Remove images without usable rois""" num_roidb = len(self._roidb) self._roidb = [roi_rec for roi_rec in self._roidb if len(roi_rec['gt_classes'])] num_after = len(self._roidb) logger.info('filter roidb: {} -> {}'.format(num_roidb, num_after))
python
def filter_roidb(self): """Remove images without usable rois""" num_roidb = len(self._roidb) self._roidb = [roi_rec for roi_rec in self._roidb if len(roi_rec['gt_classes'])] num_after = len(self._roidb) logger.info('filter roidb: {} -> {}'.format(num_roidb, num_after))
[ "def", "filter_roidb", "(", "self", ")", ":", "num_roidb", "=", "len", "(", "self", ".", "_roidb", ")", "self", ".", "_roidb", "=", "[", "roi_rec", "for", "roi_rec", "in", "self", ".", "_roidb", "if", "len", "(", "roi_rec", "[", "'gt_classes'", "]", ")", "]", "num_after", "=", "len", "(", "self", ".", "_roidb", ")", "logger", ".", "info", "(", "'filter roidb: {} -> {}'", ".", "format", "(", "num_roidb", ",", "num_after", ")", ")" ]
Remove images without usable rois
[ "Remove", "images", "without", "usable", "rois" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/rcnn/symimdb/imdb.py#L76-L81
train
apache/incubator-mxnet
example/rcnn/symimdb/imdb.py
IMDB.append_flipped_images
def append_flipped_images(self): """Only flip boxes coordinates, images will be flipped when loading into network""" logger.info('%s append flipped images to roidb' % self._name) roidb_flipped = [] for roi_rec in self._roidb: boxes = roi_rec['boxes'].copy() oldx1 = boxes[:, 0].copy() oldx2 = boxes[:, 2].copy() boxes[:, 0] = roi_rec['width'] - oldx2 - 1 boxes[:, 2] = roi_rec['width'] - oldx1 - 1 assert (boxes[:, 2] >= boxes[:, 0]).all() roi_rec_flipped = roi_rec.copy() roi_rec_flipped['boxes'] = boxes roi_rec_flipped['flipped'] = True roidb_flipped.append(roi_rec_flipped) self._roidb.extend(roidb_flipped)
python
def append_flipped_images(self): """Only flip boxes coordinates, images will be flipped when loading into network""" logger.info('%s append flipped images to roidb' % self._name) roidb_flipped = [] for roi_rec in self._roidb: boxes = roi_rec['boxes'].copy() oldx1 = boxes[:, 0].copy() oldx2 = boxes[:, 2].copy() boxes[:, 0] = roi_rec['width'] - oldx2 - 1 boxes[:, 2] = roi_rec['width'] - oldx1 - 1 assert (boxes[:, 2] >= boxes[:, 0]).all() roi_rec_flipped = roi_rec.copy() roi_rec_flipped['boxes'] = boxes roi_rec_flipped['flipped'] = True roidb_flipped.append(roi_rec_flipped) self._roidb.extend(roidb_flipped)
[ "def", "append_flipped_images", "(", "self", ")", ":", "logger", ".", "info", "(", "'%s append flipped images to roidb'", "%", "self", ".", "_name", ")", "roidb_flipped", "=", "[", "]", "for", "roi_rec", "in", "self", ".", "_roidb", ":", "boxes", "=", "roi_rec", "[", "'boxes'", "]", ".", "copy", "(", ")", "oldx1", "=", "boxes", "[", ":", ",", "0", "]", ".", "copy", "(", ")", "oldx2", "=", "boxes", "[", ":", ",", "2", "]", ".", "copy", "(", ")", "boxes", "[", ":", ",", "0", "]", "=", "roi_rec", "[", "'width'", "]", "-", "oldx2", "-", "1", "boxes", "[", ":", ",", "2", "]", "=", "roi_rec", "[", "'width'", "]", "-", "oldx1", "-", "1", "assert", "(", "boxes", "[", ":", ",", "2", "]", ">=", "boxes", "[", ":", ",", "0", "]", ")", ".", "all", "(", ")", "roi_rec_flipped", "=", "roi_rec", ".", "copy", "(", ")", "roi_rec_flipped", "[", "'boxes'", "]", "=", "boxes", "roi_rec_flipped", "[", "'flipped'", "]", "=", "True", "roidb_flipped", ".", "append", "(", "roi_rec_flipped", ")", "self", ".", "_roidb", ".", "extend", "(", "roidb_flipped", ")" ]
Only flip boxes coordinates, images will be flipped when loading into network
[ "Only", "flip", "boxes", "coordinates", "images", "will", "be", "flipped", "when", "loading", "into", "network" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/rcnn/symimdb/imdb.py#L83-L98
train
apache/incubator-mxnet
python/mxnet/gluon/model_zoo/model_store.py
get_model_file
def get_model_file(name, root=os.path.join(base.data_dir(), 'models')): r"""Return location for the pretrained on local file system. This function will download from online model zoo when model cannot be found or has mismatch. The root directory will be created if it doesn't exist. Parameters ---------- name : str Name of the model. root : str, default $MXNET_HOME/models Location for keeping the model parameters. Returns ------- file_path Path to the requested pretrained model file. """ file_name = '{name}-{short_hash}'.format(name=name, short_hash=short_hash(name)) root = os.path.expanduser(root) file_path = os.path.join(root, file_name+'.params') sha1_hash = _model_sha1[name] if os.path.exists(file_path): if check_sha1(file_path, sha1_hash): return file_path else: logging.warning('Mismatch in the content of model file detected. Downloading again.') else: logging.info('Model file not found. Downloading to %s.', file_path) util.makedirs(root) zip_file_path = os.path.join(root, file_name+'.zip') repo_url = os.environ.get('MXNET_GLUON_REPO', apache_repo_url) if repo_url[-1] != '/': repo_url = repo_url + '/' download(_url_format.format(repo_url=repo_url, file_name=file_name), path=zip_file_path, overwrite=True) with zipfile.ZipFile(zip_file_path) as zf: zf.extractall(root) os.remove(zip_file_path) if check_sha1(file_path, sha1_hash): return file_path else: raise ValueError('Downloaded file has different hash. Please try again.')
python
def get_model_file(name, root=os.path.join(base.data_dir(), 'models')): r"""Return location for the pretrained on local file system. This function will download from online model zoo when model cannot be found or has mismatch. The root directory will be created if it doesn't exist. Parameters ---------- name : str Name of the model. root : str, default $MXNET_HOME/models Location for keeping the model parameters. Returns ------- file_path Path to the requested pretrained model file. """ file_name = '{name}-{short_hash}'.format(name=name, short_hash=short_hash(name)) root = os.path.expanduser(root) file_path = os.path.join(root, file_name+'.params') sha1_hash = _model_sha1[name] if os.path.exists(file_path): if check_sha1(file_path, sha1_hash): return file_path else: logging.warning('Mismatch in the content of model file detected. Downloading again.') else: logging.info('Model file not found. Downloading to %s.', file_path) util.makedirs(root) zip_file_path = os.path.join(root, file_name+'.zip') repo_url = os.environ.get('MXNET_GLUON_REPO', apache_repo_url) if repo_url[-1] != '/': repo_url = repo_url + '/' download(_url_format.format(repo_url=repo_url, file_name=file_name), path=zip_file_path, overwrite=True) with zipfile.ZipFile(zip_file_path) as zf: zf.extractall(root) os.remove(zip_file_path) if check_sha1(file_path, sha1_hash): return file_path else: raise ValueError('Downloaded file has different hash. Please try again.')
[ "def", "get_model_file", "(", "name", ",", "root", "=", "os", ".", "path", ".", "join", "(", "base", ".", "data_dir", "(", ")", ",", "'models'", ")", ")", ":", "file_name", "=", "'{name}-{short_hash}'", ".", "format", "(", "name", "=", "name", ",", "short_hash", "=", "short_hash", "(", "name", ")", ")", "root", "=", "os", ".", "path", ".", "expanduser", "(", "root", ")", "file_path", "=", "os", ".", "path", ".", "join", "(", "root", ",", "file_name", "+", "'.params'", ")", "sha1_hash", "=", "_model_sha1", "[", "name", "]", "if", "os", ".", "path", ".", "exists", "(", "file_path", ")", ":", "if", "check_sha1", "(", "file_path", ",", "sha1_hash", ")", ":", "return", "file_path", "else", ":", "logging", ".", "warning", "(", "'Mismatch in the content of model file detected. Downloading again.'", ")", "else", ":", "logging", ".", "info", "(", "'Model file not found. Downloading to %s.'", ",", "file_path", ")", "util", ".", "makedirs", "(", "root", ")", "zip_file_path", "=", "os", ".", "path", ".", "join", "(", "root", ",", "file_name", "+", "'.zip'", ")", "repo_url", "=", "os", ".", "environ", ".", "get", "(", "'MXNET_GLUON_REPO'", ",", "apache_repo_url", ")", "if", "repo_url", "[", "-", "1", "]", "!=", "'/'", ":", "repo_url", "=", "repo_url", "+", "'/'", "download", "(", "_url_format", ".", "format", "(", "repo_url", "=", "repo_url", ",", "file_name", "=", "file_name", ")", ",", "path", "=", "zip_file_path", ",", "overwrite", "=", "True", ")", "with", "zipfile", ".", "ZipFile", "(", "zip_file_path", ")", "as", "zf", ":", "zf", ".", "extractall", "(", "root", ")", "os", ".", "remove", "(", "zip_file_path", ")", "if", "check_sha1", "(", "file_path", ",", "sha1_hash", ")", ":", "return", "file_path", "else", ":", "raise", "ValueError", "(", "'Downloaded file has different hash. Please try again.'", ")" ]
r"""Return location for the pretrained on local file system. This function will download from online model zoo when model cannot be found or has mismatch. The root directory will be created if it doesn't exist. Parameters ---------- name : str Name of the model. root : str, default $MXNET_HOME/models Location for keeping the model parameters. Returns ------- file_path Path to the requested pretrained model file.
[ "r", "Return", "location", "for", "the", "pretrained", "on", "local", "file", "system", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/model_zoo/model_store.py#L73-L120
train
apache/incubator-mxnet
python/mxnet/gluon/model_zoo/model_store.py
purge
def purge(root=os.path.join(base.data_dir(), 'models')): r"""Purge all pretrained model files in local file store. Parameters ---------- root : str, default '$MXNET_HOME/models' Location for keeping the model parameters. """ root = os.path.expanduser(root) files = os.listdir(root) for f in files: if f.endswith(".params"): os.remove(os.path.join(root, f))
python
def purge(root=os.path.join(base.data_dir(), 'models')): r"""Purge all pretrained model files in local file store. Parameters ---------- root : str, default '$MXNET_HOME/models' Location for keeping the model parameters. """ root = os.path.expanduser(root) files = os.listdir(root) for f in files: if f.endswith(".params"): os.remove(os.path.join(root, f))
[ "def", "purge", "(", "root", "=", "os", ".", "path", ".", "join", "(", "base", ".", "data_dir", "(", ")", ",", "'models'", ")", ")", ":", "root", "=", "os", ".", "path", ".", "expanduser", "(", "root", ")", "files", "=", "os", ".", "listdir", "(", "root", ")", "for", "f", "in", "files", ":", "if", "f", ".", "endswith", "(", "\".params\"", ")", ":", "os", ".", "remove", "(", "os", ".", "path", ".", "join", "(", "root", ",", "f", ")", ")" ]
r"""Purge all pretrained model files in local file store. Parameters ---------- root : str, default '$MXNET_HOME/models' Location for keeping the model parameters.
[ "r", "Purge", "all", "pretrained", "model", "files", "in", "local", "file", "store", "." ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/python/mxnet/gluon/model_zoo/model_store.py#L122-L134
train
apache/incubator-mxnet
example/ssd/dataset/mscoco.py
Coco.image_path_from_index
def image_path_from_index(self, index): """ given image index, find out full path Parameters: ---------- index: int index of a specific image Returns: ---------- full path of this image """ assert self.image_set_index is not None, "Dataset not initialized" name = self.image_set_index[index] image_file = os.path.join(self.image_dir, 'images', name) assert os.path.isfile(image_file), 'Path does not exist: {}'.format(image_file) return image_file
python
def image_path_from_index(self, index): """ given image index, find out full path Parameters: ---------- index: int index of a specific image Returns: ---------- full path of this image """ assert self.image_set_index is not None, "Dataset not initialized" name = self.image_set_index[index] image_file = os.path.join(self.image_dir, 'images', name) assert os.path.isfile(image_file), 'Path does not exist: {}'.format(image_file) return image_file
[ "def", "image_path_from_index", "(", "self", ",", "index", ")", ":", "assert", "self", ".", "image_set_index", "is", "not", "None", ",", "\"Dataset not initialized\"", "name", "=", "self", ".", "image_set_index", "[", "index", "]", "image_file", "=", "os", ".", "path", ".", "join", "(", "self", ".", "image_dir", ",", "'images'", ",", "name", ")", "assert", "os", ".", "path", ".", "isfile", "(", "image_file", ")", ",", "'Path does not exist: {}'", ".", "format", "(", "image_file", ")", "return", "image_file" ]
given image index, find out full path Parameters: ---------- index: int index of a specific image Returns: ---------- full path of this image
[ "given", "image", "index", "find", "out", "full", "path" ]
1af29e9c060a4c7d60eeaacba32afdb9a7775ba7
https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/ssd/dataset/mscoco.py#L52-L68
train