| // Copyright (c) Facebook, Inc. and its affiliates. | |
| // All rights reserved. | |
| // | |
| // Copyright 2019 Google LLC | |
| // | |
| // This source code is licensed under the BSD-style license found in the | |
| // LICENSE file in the root directory of this source tree. | |
| extern "C" { | |
| /// The number of bytes XNNPACK may read beyond array bounds. | |
| /// The caller must allocate at least this many extra bytes after the tensor data passed to XNNPACK. | |
| /// | |
| /// Note: XNNPACK reads, but never writes beyond array bounds. | |
| /// Maximum number of dimensions in tensor shape. | |
| /// A value ID that cannot be valid. | |
| /// Allow sparse inference in a Runtime. | |
| /// | |
| /// Note: this flag is a hint to XNNPACK that it should consider sparse inference, but does not guarantee it. | |
| /// Allow IEEE FP16 inference in a Runtime. | |
| /// | |
| /// Note: this flag hints XNNPACK to consider IEEE FP16 inference, but does not guarantee it. | |
| /// Force IEEE FP16 inference in a Runtime, and fail if FP16 inference is not possible. | |
| /// | |
| /// Note: this flag guarantees that XNNPACK will use IEEE FP16 inference, or fail to create the Runtime object. | |
| /// Warning: on x86 systems FP16 computations will be emulated at a substantial performance cost. | |
| /// Enable timing of each operator's runtime. | |
| /// Enable the just-in-time compiler. | |
| /// The convolution operator represents a depthwise convolution, and use HWGo layout for filters. | |
| /// Assume transposed weights in a fully connected operator. | |
| /// The operator assumes NHWC layout for the input, regardless of the output layout. | |
| /// Match "SAME" padding in TensorFlow. Exact padding values are computed dynamically depending on input size. | |
| /// Assume transposed weights in a batch matrix multiply operator. | |
| /// Assume transposed input in a batch matrix multiply operator. | |
| /// Implicitly flatten and reshape input of a Fully Connected operator into a 2D tensor. | |
| /// Match behaviour of TensorFlow 1.x. | |
| /// Static weights of the FP16 operator are in FP32 format. | |
| /// Static biases of the FP16 operator are in FP32 format. | |
| /// Align corners of input and output images in resize operations. | |
| /// Yield worker threads of the thread pool to the system scheduler after the inference. | |
| /// Use transient indirection buffer to reduce memory footprint | |
| /// Retain reduced dimensions with length 1. | |
| // Next unused flag value: 0x00000100. | |
| /// The number of entries in an array of xnn_quantization_params that XNNPACK may read beyond array bounds. | |
| /// The caller must allocate at least this many extra xnn_quantization_params before passing the array to XNNPACK. | |
| /// | |
| /// Note: XNNPACK reads, but never writes beyond array bounds. | |
| /// The minimum blocksize for blockwise quantized operators. | |
| struct xnn_quantization_params { | |
| int32_t zero_point; | |
| float scale; | |
| }; | |
| /// Status code for any XNNPACK function call. | |
| enum xnn_status { | |
| /// The call succeeded, and all output arguments now contain valid data. | |
| xnn_status_success = 0, | |
| xnn_status_uninitialized = 1, | |
| xnn_status_invalid_parameter = 2, | |
| xnn_status_invalid_state = 3, | |
| xnn_status_unsupported_parameter = 4, | |
| xnn_status_unsupported_hardware = 5, | |
| xnn_status_out_of_memory = 6, | |
| xnn_status_reallocation_required = 7, | |
| xnn_status_deprecated = 8, | |
| }; | |
| struct xnn_allocator { | |
| /// User-specified pointer that will be passed as-is to all functions in this structure. | |
| void* context; | |
| /// Pointer to a function to be called for general memory allocation. | |
| /// | |
| /// @param context - The user-specified pointer from xnn_allocator structure. | |
| /// @param size - The size of the memory block to allocate, in bytes. | |
| /// | |
| /// @returns Pointer to the allocated memory block of at least @ref size bytes. | |
| /// If allocation fails, the function must return NULL. | |
| void* (*allocate)(void* context, size_t size); | |
| /// Pointer to a function to be called for general memory re-allocation, i.e. to increase or shrink a previously | |
| /// allocated memory block. The content of the old memory block is copied to the new memory block. | |
| /// | |
| /// @param context - The user-specified pointer from xnn_allocator structure. | |
| /// @param pointer - Pointer to a memory block allocated by @ref allocate or @ref reallocate functions. Can be NULL. | |
| /// If the pointer is NULL, the @ref reallocate call is equivalent to an @ref allocate call. | |
| /// @param size - The new size of the memory block to allocate, in bytes. | |
| /// | |
| /// @returns Pointer to the newly allocated memory block of at least @ref size bytes with the content of the previous | |
| /// memory block. | |
| /// If allocation fails, the function must return NULL, but must not release the previous memory block. | |
| void* (*reallocate)(void* context, void* pointer, size_t size); | |
| /// Pointer to a function to be called for general memory de-allocation. | |
| /// | |
| /// @param context - The user-specified pointer from xnn_allocator structure. | |
| /// @param pointer - Pointer to a memory block allocated by @ref allocate or @ref reallocate functions. Can be NULL. | |
| /// If the pointer is NULL, the @ref deallocate call is a no-op. | |
| void (*deallocate)(void* context, void* pointer); | |
| /// Pointer to a function to be called for aligned memory allocation. | |
| /// | |
| /// @param context - The user-specified pointer from xnn_allocator structure. | |
| /// @param alignment - The alignment of the memory block to allocate, in bytes. Alignment is always a power-of-2. | |
| /// @param size - The size of the memory block to allocate, in bytes. | |
| /// | |
| /// @returns Pointer to the allocated memory block of at least @ref size bytes. | |
| /// If allocation fails, the function must return NULL. | |
| void* (*aligned_allocate)(void* context, size_t alignment, size_t size); | |
| /// Pointer to a function to be called for aligned memory deallocation. | |
| /// | |
| /// @param context - The user-specified pointer from xnn_allocator structure. | |
| /// @param pointer - Pointer to a memory block allocated by @ref aligned_allocate function. Can be NULL. | |
| /// If the pointer is NULL, the @ref aligned_deallocate call is a no-op. | |
| void (*aligned_deallocate)(void* context, void* pointer); | |
| }; | |
| /// Initialize XNNPACK library. | |
| /// | |
| /// XNNPACK must be successfully initialized before use. During initialization, XNNPACK populates internal structures | |
| /// depending on the host processor. Initialization can be time-consuming. | |
| /// | |
| /// @param[in] allocator - structure with function pointers to be use for memory allocation and de-allocation. | |
| /// If this argument is NULL, system-provided memory management functions (e.g. malloc/free) | |
| /// will be used. | |
| /// | |
| /// @retval xnn_status_success - XNNPACK is successfully initialized and ready to use. | |
| /// @retval xnn_status_out_of_memory - initialization failed due to out-of-memory condition. | |
| /// @retval xnn_status_unsupported_hardware - initialization failed because the host processor does not satisfy the | |
| /// minimum hardware requirements for XNNPACK. E.g. this may happen on x86 | |
| /// processors without SSE2 extension, or on 32-bit ARM processors without | |
| /// the NEON SIMD extension. | |
| enum xnn_status xnn_initialize(const struct xnn_allocator* allocator); | |
| /// Deinitialize XNNPACK library. | |
| /// | |
| /// To avoid memory and resource leaks, users must call xnn_deinitialize once for each successful xnn_initialize call. | |
| /// | |
| /// @retval xnn_status_success - deinitialization call succeeded. | |
| enum xnn_status xnn_deinitialize(void); | |
| /// Get the microkernel implementation build identifier's data. | |
| /// | |
| /// That identifier will be unique for the current set of microkernels implementations. | |
| /// | |
| /// @returns A pointer to the current identifier's data. | |
| const void* xnn_experimental_get_build_identifier_data(); | |
| /// Get the microkernel implementation build identifier's data size. | |
| /// | |
| /// @returns The size in bytes of the identifier's data. | |
| size_t xnn_experimental_get_build_identifier_size(); | |
| /// Check whether the given data matches this version's identifier. | |
| /// | |
| /// @returns The size in bytes of the identifier's data. | |
| bool xnn_experimental_check_build_identifier(const void* data, size_t size); | |
| /// Subgraph is an abstract representation of a neural network model. | |
| /// Subgraph objects are used to define Values (tensors) and Nodes (operators) comprising the model. | |
| typedef struct xnn_subgraph* xnn_subgraph_t; | |
| /// Create a empty Subgraph object. | |
| /// | |
| /// @param external_value_ids - number of Value IDs to reserve for communication with external graph representation. | |
| /// The Subgraph object would avoid creating internal Value IDs in the | |
| /// [0, reserved_value_ids-1] range. | |
| /// @param flags - binary features of the subgraph. No supported flags are currently defined. | |
| /// @param subgraph_out - pointer to the variable that will be initialized with a handle to the Subgraph object upon | |
| /// successful return. | |
| enum xnn_status xnn_create_subgraph( | |
| uint32_t external_value_ids, | |
| uint32_t flags, | |
| xnn_subgraph_t* subgraph_out); | |
| /// Destroy a Subgraph object, as well as Values, and Nodes associated with the subgraph. | |
| /// | |
| /// @param subgraph - the Subgraph object to destroy. | |
| enum xnn_status xnn_delete_subgraph( | |
| xnn_subgraph_t subgraph); | |
| /// Type of elements in a Value object. | |
| enum xnn_datatype { | |
| /// Invalid data type. Valid Values never have this datatype. | |
| xnn_datatype_invalid = 0, | |
| /// IEEE754 single-precision floating-point. | |
| xnn_datatype_fp32 = 1, | |
| /// IEEE754 half-precision floating-point. | |
| xnn_datatype_fp16 = 2, | |
| /// Quantized 8-bit signed integer with shared per-Value quantization | |
| /// parameters. | |
| xnn_datatype_qint8 = 3, | |
| /// Quantized 8-bit unsigned integer with shared per-Value quantization | |
| /// parameters. | |
| xnn_datatype_quint8 = 4, | |
| /// Quantized 32-bit signed integer with shared per-Value quantization | |
| /// parameters. | |
| xnn_datatype_qint32 = 5, | |
| /// Quantized 8-bit signed integer with shared per-channel quantization | |
| /// parameters. | |
| xnn_datatype_qcint8 = 6, | |
| /// Quantized 32-bit signed integer with shared per-channel quantization | |
| /// parameters. | |
| xnn_datatype_qcint32 = 7, | |
| /// Quantized 4-bit signed integer with shared per-channel quantization | |
| /// parameters. | |
| xnn_datatype_qcint4 = 8, | |
| /// Dynamically quantized 8-bit signed integer with per-batch quantization | |
| /// parameters. | |
| xnn_datatype_qdint8 = 9, | |
| /// Dynamically quantized 8-bit signed integers packed with their per-row | |
| /// quantization parameters. | |
| xnn_datatype_qpint8 = 10, | |
| /// 32-bit signed integers. | |
| xnn_datatype_int32 = 11, | |
| /// Quantized 4-bit signed integer with shared per-channel-block quantization | |
| /// parameters. | |
| xnn_datatype_qbint4 = 12, | |
| /// IEEE754 single-precision packed floating-point. | |
| xnn_datatype_pfp32 = 13, | |
| /// BFloat16, i.e. the upper 16 bits of a float32. | |
| xnn_datatype_bf16 = 14, | |
| /// Dynamically quantized 8-bit unsigned integer with per-batch quantization | |
| /// parameters. | |
| xnn_datatype_qduint8 = 15, | |
| }; | |
| /// Define a tensor-type Value and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Value. | |
| /// @param datatype - type of the tensor elements. | |
| /// @param num_dims - number of dimensions in the shape. | |
| /// @param dims - pointer to an array of @a num_dims shape dimensions. If num_dims is 0, this pointer can be NULL. | |
| /// XNNPACK does not keep any pointers to this array after the function returns. | |
| /// @param data - pointer to static data used for tensor initialization. If the tensor is not statically initialized, | |
| /// this pointer must be is NULL. If non-NULL, the life-time of the static data must exceed the life-time | |
| /// of the Subgraph object, and of any Runtime objects created from the Subgraph. | |
| /// @param external_id - external ID for the Value. The ID must be within the range of reversed Value IDs specified on | |
| /// the Subgraph creation. If the external ID is XNN_INVALID_VALUE_ID, an internal ID will be | |
| /// created for the Value. | |
| /// @param flags - binary features of the Value. Supported values are any combination of XNN_VALUE_FLAG_EXTERNAL_INPUT | |
| /// and XNN_VALUE_FLAG_EXTERNAL_OUTPUT. | |
| /// @param id_out - pointer to the variable that will be initialized with the Value ID upon successful return. If a | |
| /// valid @a external_id was provided, the variable will be initialized with the @a external_id value. | |
| enum xnn_status xnn_define_tensor_value( | |
| xnn_subgraph_t subgraph, | |
| enum xnn_datatype datatype, | |
| size_t num_dims, | |
| const size_t* dims, | |
| const void* data, | |
| uint32_t external_id, | |
| uint32_t flags, | |
| uint32_t* id_out); | |
| /// Define a quantized tensor-type Value and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Value. | |
| /// @param datatype - type of the tensor elements. | |
| /// @param zero_point - offset from zero to subtract from the quantized elements in the Value. | |
| /// @param scale - multiplication factor to convert quantized elements to real representation. | |
| /// @param num_dims - number of dimensions in the shape. | |
| /// @param dims - pointer to an array of @a num_dims shape dimensions. If num_dims is 0, this pointer can be NULL. | |
| /// XNNPACK does not keep any pointers to this array after the function returns. | |
| /// @param data - pointer to static data used for tensor initialization. If the tensor is not statically initialized, | |
| /// this pointer must be is NULL. If non-NULL, the life-time of the static data must exceed the life-time | |
| /// of the Subgraph object, and of any Runtime objects created from the Subgraph. | |
| /// @param external_id - external ID for the Value. The ID must be within the range of reversed Value IDs specified on | |
| /// the Subgraph creation. If the external ID is XNN_INVALID_VALUE_ID, an internal ID will be | |
| /// created for the Value. | |
| /// @param flags - binary features of the Value. Supported values are any combination of XNN_VALUE_FLAG_EXTERNAL_INPUT | |
| /// and XNN_VALUE_FLAG_EXTERNAL_OUTPUT. | |
| /// @param id_out - pointer to the variable that will be initialized with the Value ID upon successful return. If a | |
| /// valid @a external_id was provided, the variable will be initialized with the @a external_id value. | |
| enum xnn_status xnn_define_quantized_tensor_value( | |
| xnn_subgraph_t subgraph, | |
| enum xnn_datatype datatype, | |
| int32_t zero_point, | |
| float scale, | |
| size_t num_dims, | |
| const size_t* dims, | |
| const void* data, | |
| uint32_t external_id, | |
| uint32_t flags, | |
| uint32_t* id_out); | |
| enum xnn_status xnn_define_channelwise_quantized_tensor_value( | |
| xnn_subgraph_t subgraph, | |
| enum xnn_datatype datatype, | |
| const float* scale, | |
| size_t num_dims, | |
| size_t channel_dim, | |
| const size_t* dims, | |
| const void* data, | |
| uint32_t external_id, | |
| uint32_t flags, | |
| uint32_t* id_out); | |
| /// Validate the dimensions, channel_dim, zero point, datatype, and scale of a quantized tensor-type. | |
| /// | |
| /// @param datatype - type of the tensor elements. | |
| /// @param zero_point - offset from zero to subtract from the quantized elements in the Value. | |
| /// @param scale - multiplication factor to convert quantized elements to real representation. | |
| /// @param num_dims - number of dimensions in the shape. | |
| /// @param dims - pointer to an array of @a num_dims shape dimensions. If num_dims is 0, this pointer can be NULL. | |
| /// XNNPACK does not keep any pointers to this array after the function returns. | |
| enum xnn_status xnn_validate_quantized_tensor( | |
| enum xnn_datatype datatype, | |
| int32_t zero_point, | |
| float scale, | |
| size_t num_dims, | |
| const size_t* dims); | |
| /// Validate the dimensions, channel_dim, zero point, datatype, and scales of a channelwise quantized tensor-type. | |
| /// | |
| /// @param datatype - type of the tensor elements. | |
| /// @param zero_point - offset from zero to subtract from the quantized elements in the Value. | |
| /// @param scale - per-channel multiplication factors to convert quantized elements to real representation. | |
| /// @param num_dims - number of dimensions in the shape. | |
| /// @param channel_dim - index of the channel dimension in the tensor with per-channel quantization parameters. | |
| /// Typically this is the first dimension (dimension #0) of the filter tensors in the Convolution, | |
| /// Deconvolution, and Fully Connected operators and the last dimension of the filter tensors in | |
| /// the Depthwise Convolution operators. | |
| /// @param dims - pointer to an array of @a num_dims shape dimensions. If num_dims is 0, this pointer can be NULL. | |
| /// XNNPACK does not keep any pointers to this array after the function returns. | |
| enum xnn_status xnn_validate_channelwise_quantized_tensor( | |
| enum xnn_datatype datatype, | |
| int32_t zero_point, | |
| const float* scale, | |
| size_t num_dims, | |
| size_t channel_dim, | |
| const size_t* dims); | |
| /// Define a channelwise quantized tensor-type Value and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Value. | |
| /// @param datatype - type of the tensor elements. | |
| /// @param zero_point - offset from zero to subtract from the quantized elements in the Value. | |
| /// @param scale - per-channel multiplication factors to convert quantized elements to real representation. | |
| /// @param num_dims - number of dimensions in the shape. | |
| /// @param channel_dim - index of the channel dimension in the tensor with per-channel quantization parameters. | |
| /// Typically this is the first dimension (dimension #0) of the filter tensors in the Convolution, | |
| /// Deconvolution, and Fully Connected operators and the last dimension of the filter tensors in | |
| /// the Depthwise Convolution operators. | |
| /// @param dims - pointer to an array of @a num_dims shape dimensions. If num_dims is 0, this pointer can be NULL. | |
| /// XNNPACK does not keep any pointers to this array after the function returns. | |
| /// @param data - pointer to static data used for tensor initialization. If the tensor is not statically initialized, | |
| /// this pointer must be is NULL. If non-NULL, the life-time of the static data must exceed the life-time | |
| /// of the Subgraph object, and of any Runtime objects created from the Subgraph. | |
| /// @param external_id - external ID for the Value. The ID must be within the range of reversed Value IDs specified on | |
| /// the Subgraph creation. If the external ID is XNN_INVALID_VALUE_ID, an internal ID will be | |
| /// created for the Value. | |
| /// @param flags - binary features of the Value. Supported values are any combination of XNN_VALUE_FLAG_EXTERNAL_INPUT | |
| /// and XNN_VALUE_FLAG_EXTERNAL_OUTPUT. | |
| /// @param id_out - pointer to the variable that will be initialized with the Value ID upon successful return. If a | |
| /// valid @a external_id was provided, the variable will be initialized with the @a external_id value. | |
| enum xnn_status xnn_define_channelwise_quantized_tensor_value_v2( | |
| xnn_subgraph_t subgraph, | |
| enum xnn_datatype datatype, | |
| int32_t zero_point, | |
| const float* scale, | |
| size_t num_dims, | |
| size_t channel_dim, | |
| const size_t* dims, | |
| const void* data, | |
| uint32_t external_id, | |
| uint32_t flags, | |
| uint32_t* id_out); | |
| /// Define a blockwise quantized tensor-type Value and add it to a Subgraph. | |
| /// @param block_size - size of a block in the tensor with blockwise quantization parameters. Block is defined as | |
| /// number of input channel element per output channel. | |
| /// For Fully connected operators with 2d filters of size [output_channels, input_channels], | |
| /// expecting number of scale values to be = output_channels * (input_channels / block_size). | |
| enum xnn_status xnn_define_blockwise_quantized_tensor_value( | |
| xnn_subgraph_t subgraph, | |
| enum xnn_datatype datatype, | |
| int32_t zero_point, | |
| const uint16_t* scale, | |
| size_t num_dims, | |
| size_t channel_dim, | |
| size_t block_size, | |
| const size_t* dims, | |
| const void* data, | |
| uint32_t external_id, | |
| uint32_t flags, | |
| uint32_t* id_out); | |
| /// Define a dynamically quantized tensor-type Value and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Value. | |
| /// @param datatype - type of the tensor elements. | |
| /// @param num_dims - number of dimensions in the shape. | |
| /// @param num_non_batch_dims - number of non-batch dimensions in the shape. The leading (num_dims - num_non_batch_dims) | |
| /// dimensions will be flattened and treated as batch size. A set of quantization parameters | |
| /// will be calculated for each batch element. | |
| /// @param dims - pointer to an array of @a num_dims shape dimensions. If num_dims is 0, this pointer can be NULL. | |
| /// XNNPACK does not keep any pointers to this array after the function returns. | |
| /// @param external_id - external ID for the Value. The ID must be within the range of reversed Value IDs specified on | |
| /// the Subgraph creation. If the external ID is XNN_INVALID_VALUE_ID, an internal ID will be | |
| /// created for the Value. | |
| /// @param flags - binary features of the Value. No supported flags are currently defined. | |
| /// @param id_out - pointer to the variable that will be initialized with the Value ID upon successful return. If a | |
| /// valid @a external_id was provided, the variable will be initialized with the @a external_id value. | |
| enum xnn_status xnn_define_dynamically_quantized_tensor_value( | |
| xnn_subgraph_t subgraph, | |
| enum xnn_datatype datatype, | |
| size_t num_dims, | |
| size_t num_nonbatch_dims, | |
| const size_t* dims, | |
| uint32_t external_id, | |
| uint32_t flags, | |
| uint32_t* id_out); | |
| /// Type of unary operation | |
| enum xnn_unary_operator { | |
| xnn_unary_invalid = -1, | |
| xnn_unary_convert, | |
| xnn_unary_clamp, | |
| xnn_unary_abs, | |
| xnn_unary_bankers_rounding, | |
| xnn_unary_ceiling, | |
| xnn_unary_elu, | |
| xnn_unary_exp, | |
| xnn_unary_floor, | |
| xnn_unary_gelu, | |
| xnn_unary_hardswish, | |
| xnn_unary_leaky_relu, | |
| xnn_unary_log, | |
| xnn_unary_negate, | |
| xnn_unary_sigmoid, | |
| xnn_unary_square, | |
| xnn_unary_square_root, | |
| xnn_unary_reciprocal_square_root, | |
| xnn_unary_tanh, | |
| // The following operators are experimental and may be removed. | |
| xnn_unary_cube_root, | |
| xnn_unary_cosine, | |
| xnn_unary_sine, | |
| xnn_unary_count_leading_zeros, | |
| xnn_unary_bitwise_not, | |
| xnn_unary_popcount, | |
| xnn_unary_sign, | |
| }; | |
| /// Parameters for xnn_define_unary | |
| union xnn_unary_params { | |
| struct { | |
| /// lower bound for clipping output values. | |
| float min; | |
| /// upper bound for clipping output values. | |
| float max; | |
| } clamp; | |
| struct { | |
| /// scale factor for negative input elements. | |
| float alpha; | |
| } elu; | |
| struct { | |
| /// scale factor for negative input elements. | |
| float negative_slope; | |
| } leaky_relu; | |
| }; | |
| /// Define a unary operator Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param operator - type of unary operator to define. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param params - parameters to be interpreted by the specific operator type. | |
| /// @param flags - binary features of the Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_unary( | |
| xnn_subgraph_t subgraph, | |
| enum xnn_unary_operator type, | |
| const union xnn_unary_params* params, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Convert Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Convert Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_convert( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2D Convolution Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_padding_top - implicit zero-padding above 2D input data. Must be 0 if XNN_FLAG_TENSORFLOW_SAME_PADDING | |
| /// flag is specified. | |
| /// @param input_padding_right - implicit zero-padding to the right of 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param input_padding_bottom - implicit zero-padding below 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param input_padding_left - implicit zero-padding to the left of 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param kernel_height - kernel (filter) height. | |
| /// @param kernel_width - kernel (filter) width. | |
| /// @param subsampling_height - height of subsampling region for convolution output (convolution height stride). | |
| /// @param subsampling_width - width of subsampling region for convolution output (convolution width stride). | |
| /// @param dilation_height - dilation of kernel elements along the height dimension. | |
| /// @param dilation_width - dilation of kernel elements along the width dimension. | |
| /// @param groups - number of convolution groups. | |
| /// @param group_input_channels - number of input channels per group. | |
| /// @param group_output_channels - number of output channels per group. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, IH, IW, groups * group_input_channels] dimensions | |
| /// @param filter_id - Value ID for the filter tensor. The filter tensor must ge a 4D tensor defined in the @a subgraph | |
| /// with [groups * group_output_channels, kernel_height, kernel_width, group_input_channels] | |
| /// dimensions. | |
| /// @param bias_id - Value ID for the bias tensor, or XNN_INVALID_VALUE_ID for a 2D Convolution Node without a bias. If | |
| /// present, the bias tensor must be a 1D tensor defined in the @a subgraph with [groups * | |
| /// group_output_channels] dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, OH, OW, groups * group_output_channels] dimensions. | |
| /// @param flags - binary features of the 2D Convolution Node. The only currently supported values is | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING. | |
| enum xnn_status xnn_define_convolution_2d( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t filter_id, | |
| uint32_t bias_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2D Deconvolution (Transposed Convolution) Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param padding_top - implicit padding above 2D output data. | |
| /// @param padding_right - implicit padding to the right of 2D output data. | |
| /// @param padding_bottom - implicit padding below 2D output data. | |
| /// @param padding_left - implicit padding to the left of 2D output data. | |
| /// @param adjustment_height - additional elements in the bottom of the 2D output data. | |
| /// @param adjustment_width - additional elements to the right of the 2D output data. | |
| /// @param kernel_height - kernel (filter) height. | |
| /// @param kernel_width - kernel (filter) width. | |
| /// @param upsampling_height - height of upsampling region for deconvolution input (deconvolution height stride). | |
| /// @param upsampling_width - width of upsampling region for deconvolution input (deconvolution width stride). | |
| /// @param dilation_height - dilation of kernel elements along the height dimension. | |
| /// @param dilation_width - dilation of kernel elements along the width dimension. | |
| /// @param groups - number of convolution groups. | |
| /// @param group_input_channels - number of input channels per group. | |
| /// @param group_output_channels - number of output channels per group. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, IH, IW, groups * group_input_channels] dimensions | |
| /// @param filter_id - Value ID for the filter tensor. The filter tensor must ge a 4D tensor defined in the @a subgraph | |
| /// with [groups * group_output_channels, kernel_height, kernel_width, group_input_channels] | |
| /// dimensions. | |
| /// @param bias_id - Value ID for the bias tensor, or XNN_INVALID_VALUE_ID for a 2D Convolution Node without a bias. If | |
| /// present, the bias tensor must be a 1D tensor defined in the @a subgraph with | |
| /// [groups * group_output_channels] dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, OH, OW, groups * group_output_channels] dimensions. | |
| /// @param flags - binary features of the 2D Deconvolution Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_deconvolution_2d( | |
| xnn_subgraph_t subgraph, | |
| uint32_t padding_top, | |
| uint32_t padding_right, | |
| uint32_t padding_bottom, | |
| uint32_t padding_left, | |
| uint32_t adjustment_height, | |
| uint32_t adjustment_width, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t upsampling_height, | |
| uint32_t upsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t filter_id, | |
| uint32_t bias_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2D Depthwise Convolution Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_padding_top - implicit zero-padding above 2D input data. Must be 0 if XNN_FLAG_TENSORFLOW_SAME_PADDING | |
| /// flag is specified. | |
| /// @param input_padding_right - implicit zero-padding to the right of 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param input_padding_bottom - implicit zero-padding below 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param input_padding_left - implicit zero-padding to the left of 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param kernel_height - kernel (filter) height. | |
| /// @param kernel_width - kernel (filter) width. | |
| /// @param subsampling_height - height of subsampling region for convolution output (convolution height stride). | |
| /// @param subsampling_width - width of subsampling region for convolution output (convolution width stride). | |
| /// @param dilation_height - dilation of kernel elements along the height dimension. | |
| /// @param dilation_width - dilation of kernel elements along the width dimension. | |
| /// @param depth_multiplier - ratio of output channels to input channels. | |
| /// @param input_channels - number of input channels. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, IH, IW, input_channels] dimensions | |
| /// @param filter_id - Value ID for the filter tensor. The filter tensor must ge a 4D tensor defined in the @a subgraph | |
| /// with [1, kernel_height, kernel_width, input_channels * depth_multiplier] dimensions. | |
| /// @param bias_id - Value ID for the bias tensor, or XNN_INVALID_VALUE_ID for a 2D Depthwise Convolution Node without | |
| /// a bias. If present, the bias tensor must be a 1D tensor defined in the @a subgraph with | |
| /// [input_channels * depth_multiplier] dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, OH, OW, input_channels * depth_multiplier] dimensions. | |
| /// @param flags - binary features of the 2D Depthwise Convolution Node. The only currently supported values is | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING. | |
| enum xnn_status xnn_define_depthwise_convolution_2d( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t depth_multiplier, | |
| size_t input_channels, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t filter_id, | |
| uint32_t bias_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Depth To Space Node 2D and add it to a Subgraph. | |
| /// | |
| /// The Depth To Space 2D Node rearranges data from depth into blocks of spatial data (a reverse transform to | |
| /// Space To Depth). For a given input pixel, an output square of pixels with side @a block_size is formed from values | |
| /// in the corresponding number of its channels. The output depth is therefore @a block_size x @a block_size times | |
| /// smaller than that of the input. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param block_size - the size of the spatial block. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, IH, IW, OC * block_size * block_size] dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, IH * block_size, IW * block_size, OC] dimensions. | |
| /// @param flags - binary features of the input_channels Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_depth_to_space_2d( | |
| xnn_subgraph_t subgraph, | |
| uint32_t block_size, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| enum xnn_status xnn_define_depth_to_space( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t block_size, | |
| uint32_t flags); | |
| /// Define a 1D Global Average Pooling Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a dense tensor with 2 or more dimensions | |
| /// defined in the @a subgraph. Averaging is performed across the second-innermost dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a dense tensor with 2 or more | |
| /// dimensions defined in the @a subgraph. | |
| /// @param flags - binary features of the 1D Global Average Pooling Node. The only currently supported value is | |
| /// XNN_FLAG_KEEP_DIMS. | |
| XNN_DEPRECATED enum xnn_status xnn_define_global_average_pooling_1d( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2D Global Average Pooling Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a dense tensor with 3 or more dimensions | |
| /// defined in the @a subgraph. Averaging is performed across the second- and third-innermost | |
| /// dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a dense tensor with 3 or more | |
| /// dimensions defined in the @a subgraph. | |
| /// @param flags - binary features of the 2D Global Average Pooling Node. The only currently supported value is | |
| /// XNN_FLAG_KEEP_DIMS. | |
| XNN_DEPRECATED enum xnn_status xnn_define_global_average_pooling_2d( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 1D Global Sum Pooling Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a dense tensor with 2 or more dimensions | |
| /// defined in the @a subgraph. Averaging is performed across the second-innermost dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a dense tensor with 2 or more | |
| /// dimensions defined in the @a subgraph. | |
| /// @param flags - binary features of the 1D Global Sum Pooling Node. The only currently supported value is | |
| /// XNN_FLAG_KEEP_DIMS. | |
| XNN_DEPRECATED enum xnn_status xnn_define_global_sum_pooling_1d( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2D Global Sum Pooling Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a dense tensor with 3 or more dimensions | |
| /// defined in the @a subgraph. Averaging is performed across the second- and third-innermost | |
| /// dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a dense tensor with 3 or more | |
| /// dimensions defined in the @a subgraph. | |
| /// @param flags - binary features of the 2D Global Sum Pooling Node. The only currently supported value is | |
| /// XNN_FLAG_KEEP_DIMS. | |
| XNN_DEPRECATED enum xnn_status xnn_define_global_sum_pooling_2d( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2D Average Pooling Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_padding_top - implicit zero-padding above 2D input data. Must be 0 if XNN_FLAG_TENSORFLOW_SAME_PADDING | |
| /// flag is specified. | |
| /// @param input_padding_right - implicit zero-padding to the right of 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param input_padding_bottom - implicit zero-padding below 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param input_padding_left - implicit zero-padding to the left of 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param pooling_height - pooling (kernel) height. | |
| /// @param pooling_width - pooling (kernel) width. | |
| /// @param stride_height - displacing of the pooling window in the vertical dimension of the input pixels corresponding | |
| /// to vertically adjacent output pixels. | |
| /// @param stride_width - displacing of the pooling window in the horizontal dimension of the input pixels corresponding | |
| /// to horizontally adjacent output pixels. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, IH, IW, channels] dimensions | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, OH, OW, channels] dimensions. | |
| /// @param flags - binary features of the 2D Average Pooling Node. The only currently supported values is | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING. | |
| enum xnn_status xnn_define_average_pooling_2d( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Fully Connected Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be an N-dimensional tensor defined in the | |
| /// @a subgraph. If XNN_FLAG_TENSORFLOW_RESHAPE_2D is not specified, the input tensor must be at least | |
| /// 1D and its last dimension must match the last dimension of the filter tensor. In particular, if | |
| /// input is a 2D tensor, it must have [batch_size, input_channels] dimensions. | |
| /// If XNN_FLAG_TENSORFLOW_RESHAPE_2D is specified, the number of elements in the input tensor must be | |
| /// divisible by the input_channels. The tensor will be first flattened into a 1D tensor of | |
| /// [num_input_elements] dimensions, then reshaped into a 2D tensor of | |
| /// [num_input_elements / input_channels, input_channels] dimensions where num_input_elements is the | |
| /// total number of elements in the input tensor. | |
| /// @param filter_id - Value ID for the filter tensor. The filter tensor must a 2D tensor defined in the @a subgraph. | |
| /// If the XNN_FLAG_TRANSPOSE_WEIGHTS flag is not specified, the filter tensor must have | |
| /// [output_channels, input_channels] dimensions. If the XNN_FLAG_TRANSPOSE_WEIGHTS flag is | |
| /// specified, the filter tensor must have [input_channels, output_channels] dimensions. | |
| /// @param bias_id - Value ID for the bias tensor, or XNN_INVALID_VALUE_ID for a Fully Connected Node without a bias. | |
| /// If present, the bias tensor must be a 1D tensor defined in the @a subgraph with [output_channels] | |
| /// dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph. | |
| /// If XNN_FLAG_TENSORFLOW_RESHAPE_2D is not specified, the output tensor must have the same | |
| /// dimensionality as the input tensor, all its dimensions but the last one must match the | |
| /// corresponding dimensions of the input tensor, and the last dimensions of the output tensor must | |
| /// match the first dimension of the filter tensor. In particular, if input is a 2D tensor, output | |
| /// must be a 2D tensor of [batch_size, output_channels] dimensions. | |
| /// If XNN_FLAG_TENSORFLOW_RESHAPE_2D is specified, output must be a 2D tensor of | |
| /// [num_input_elements / input_channels, output_channels] dimensions where num_input_elements is the | |
| /// total number of elements in the input tensor. | |
| /// @param flags - binary features of the Fully Connected Node. The only currently supported values are | |
| /// XNN_FLAG_TENSORFLOW_RESHAPE_2D and XNN_FLAG_TRANSPOSE_WEIGHTS. | |
| enum xnn_status xnn_define_fully_connected( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t filter_id, | |
| uint32_t bias_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Sparse Fully Connected Node and add it to a Subgraph. | |
| /// | |
| /// This operator is experimental, and will be removed in the future. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be an N-dimensional tensor defined in the | |
| /// @a subgraph. If XNN_FLAG_TENSORFLOW_RESHAPE_2D is not specified, the input tensor must be at least | |
| /// 1D and its last dimension must match the last dimension of the filter tensor. In particular, if | |
| /// input is a 2D tensor, it must have [batch_size, input_channels] dimensions. | |
| /// If XNN_FLAG_TENSORFLOW_RESHAPE_2D is specified, the number of elements in the input tensor must be | |
| /// divisible by the input_channels. The tensor will be first flattened into a 1D tensor of | |
| /// [num_input_elements] dimensions, then reshaped into a 2D tensor of | |
| /// [num_input_elements / input_channels, input_channels] dimensions where num_input_elements is the | |
| /// total number of elements in the input tensor. | |
| /// @param filter_id - Value ID for the filter tensor. The filter tensor must a 2D tensor defined in the @a subgraph. | |
| /// If the XNN_FLAG_TRANSPOSE_WEIGHTS flag is not specified, the filter tensor must have | |
| /// [output_channels, input_channels] dimensions. If the XNN_FLAG_TRANSPOSE_WEIGHTS flag is | |
| /// specified, the filter tensor must have [input_channels, output_channels] dimensions. | |
| /// @param bias_id - Value ID for the bias tensor, or XNN_INVALID_VALUE_ID for a Fully Connected Node without a bias. | |
| /// If present, the bias tensor must be a 1D tensor defined in the @a subgraph with [output_channels] | |
| /// dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph. | |
| /// If XNN_FLAG_TENSORFLOW_RESHAPE_2D is not specified, the output tensor must have the same | |
| /// dimensionality as the input tensor, all its dimensions but the last one must match the | |
| /// corresponding dimensions of the input tensor, and the last dimensions of the output tensor must | |
| /// match the first dimension of the filter tensor. In particular, if input is a 2D tensor, output | |
| /// must be a 2D tensor of [batch_size, output_channels] dimensions. | |
| /// If XNN_FLAG_TENSORFLOW_RESHAPE_2D is specified, output must be a 2D tensor of | |
| /// [num_input_elements / input_channels, output_channels] dimensions where num_input_elements is the | |
| /// total number of elements in the input tensor. | |
| /// @param flags - binary features of the Fully Connected Node. The only currently supported values are | |
| /// XNN_FLAG_TENSORFLOW_RESHAPE_2D and XNN_FLAG_TRANSPOSE_WEIGHTS. | |
| enum xnn_status xnn_define_fully_connected_sparse( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t filter_id, | |
| uint32_t bias_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2D Max Pooling Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_padding_top - implicit zero-padding above 2D input data. Must be 0 if XNN_FLAG_TENSORFLOW_SAME_PADDING | |
| /// flag is specified. | |
| /// @param input_padding_right - implicit zero-padding to the right of 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param input_padding_bottom - implicit zero-padding below 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param input_padding_left - implicit zero-padding to the left of 2D input data. Must be 0 if | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING flag is specified. | |
| /// @param pooling_height - pooling (kernel) height. | |
| /// @param pooling_width - pooling (kernel) width. | |
| /// @param stride_height - displacing of the pooling window in the vertical dimension of the input pixels corresponding | |
| /// to vertically adjacent output pixels. | |
| /// @param stride_width - displacing of the pooling window in the horizontal dimension of the input pixels corresponding | |
| /// to horizontally adjacent output pixels. | |
| /// @param dilation_height - dilation of pooling elements along the height dimension. | |
| /// @param dilation_width - dilation of pooling elements along the width dimension. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, IH, IW, channels] dimensions | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, OH, OW, channels] dimensions. | |
| /// @param flags - binary features of the 2D Max Pooling Node. The only currently supported values is | |
| /// XNN_FLAG_TENSORFLOW_SAME_PADDING. | |
| enum xnn_status xnn_define_max_pooling_2d( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2D ArgMax Pooling Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_padding_top - implicit zero-padding above 2D input data. | |
| /// @param input_padding_right - implicit zero-padding to the right of 2D input data. | |
| /// @param input_padding_bottom - implicit zero-padding below 2D input data. | |
| /// @param input_padding_left - implicit zero-padding to the left of 2D input data. | |
| /// @param pooling_height - pooling (kernel) height. Vertical stride between pooling regions match this value. | |
| /// @param pooling_width - pooling (kernel) width. Horizontal stride between pooling regions match this value. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, IH, IW, channels] dimensions | |
| /// @param output_value_id - Value ID for the output tensor with the maximum values in the pools. The output tensor must | |
| /// be a 4D tensor defined in the @a subgraph with [N, OH, OW, channels] dimensions. | |
| /// @param output_index_id - Value ID for the output tensor with the indexes of the maximum values in the pools. The | |
| /// output tensor must be a 4D tensor defined in the @a subgraph with [N, OH, OW, channels] | |
| /// dimensions. | |
| /// @param flags - binary features of the 2D ArgMax Pooling Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_argmax_pooling_2d( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t input_id, | |
| uint32_t output_value_id, | |
| uint32_t output_index_id, | |
| uint32_t flags); | |
| /// Define a 2D UnPooling Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param padding_top - implicit padding above 2D output data. | |
| /// @param padding_right - implicit padding to the right of 2D output data. | |
| /// @param padding_bottom - implicit padding below 2D output data. | |
| /// @param padding_left - implicit padding to the left of 2D output data. | |
| /// @param pooling_height - height of the pooling window. | |
| /// @param pooling_width - width of the pooling window. | |
| /// @param input_value_id - Value ID for the input tensor with the max-pooling values to invert. The input value tensor | |
| /// must be a 4D tensor defined in the @a subgraph with [N, IH, IW, channels] dimensions. | |
| /// @param input_index_id - Value ID for the input tensor with the indices of the per-pool maximum values produced by | |
| /// a 2D UnPooling Node. The input tensor must be a 4D tensor defined in the @a subgraph with | |
| /// [N, IH, IW, channels] dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, OH, OW, channels] dimensions. | |
| /// @param flags - binary features of the 2D UnPooling Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_unpooling_2d( | |
| xnn_subgraph_t subgraph, | |
| uint32_t padding_top, | |
| uint32_t padding_right, | |
| uint32_t padding_bottom, | |
| uint32_t padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t input_value_id, | |
| uint32_t input_index_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| enum xnn_binary_operator { | |
| xnn_binary_invalid = -1, | |
| xnn_binary_add, | |
| xnn_binary_subtract, | |
| xnn_binary_multiply, | |
| xnn_binary_divide, | |
| xnn_binary_maximum, | |
| xnn_binary_minimum, | |
| xnn_binary_copysign, | |
| xnn_binary_squared_difference, | |
| xnn_binary_prelu, | |
| // The following operators are experimental and may be removed. | |
| xnn_binary_modulus, | |
| xnn_binary_atan2, | |
| xnn_binary_pow, | |
| xnn_binary_bitwise_and, | |
| xnn_binary_bitwise_or, | |
| xnn_binary_bitwise_xor, | |
| xnn_binary_shift_left, | |
| xnn_binary_shift_right_logical, | |
| xnn_binary_shift_right_arithmetic, | |
| }; | |
| struct xnn_binary_params { | |
| /// lower bound for clipping output values. | |
| double output_min; | |
| /// upper bound for clipping output values. | |
| double output_max; | |
| }; | |
| /// Define a 2-Input binary operator Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param type - Type of operator to apply to the two inputs. | |
| /// @param params - Optional parameters for the operator. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the second | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an M-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the first | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a max(N,M)-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the maximum between the corresponding dimension | |
| /// of the two inputs. | |
| /// @param flags - binary features of the Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_binary( | |
| xnn_subgraph_t subgraph, | |
| enum xnn_binary_operator type, | |
| const struct xnn_binary_params* params, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2-Input Add Node and add it to a Subgraph. | |
| /// | |
| /// The 2-Input Add Node computes elementwise addition of two tensor inputs with numpy broadcasting rules. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the second | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an M-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the first | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a max(N,M)-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the maximum between the corresponding dimension | |
| /// of the two inputs. | |
| /// @param flags - binary features of the Add Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_add2( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2-Input Multiply Node and add it to a Subgraph. | |
| /// | |
| /// The 2-Input Multiply Node computes elementwise multiplication of two tensor inputs with numpy broadcasting rules. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the second | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an M-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the first | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a max(N,M)-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the maximum between the corresponding dimension | |
| /// of the two inputs. | |
| /// @param flags - binary features of the Multiply Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_multiply2( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| // Cap operations applied to logits (Q * K) of attention operator. | |
| enum xnn_attention_logits_cap_type { | |
| // No capping. | |
| xnn_attention_logits_cap_type_none = 0, | |
| // Cap the absolute values of logits by tanh: tanh(logits / cap) * cap | |
| xnn_attention_logits_cap_type_tanh | |
| }; | |
| // Params when the cap type is xnn_attention_logits_cap_type_tanh. | |
| struct xnn_attention_logits_cap_tanh_params { | |
| float cap; | |
| }; | |
| /// Define a Scaled Dot-Product Attention Node and add it to a Subgraph. | |
| /// | |
| /// This operator is experimental. | |
| /// | |
| /// The Scaled Dot-Product Attention Node computes a multi-head or multi-query scaled dot attention on the query, key, | |
| /// and value tensors. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param cap_type - type of cap to be applied to the logits. | |
| /// @param cap_params - parameters for the cap. Must be a pointer to xnn_attention_logits_cap_tanh_params if cap_type | |
| /// is xnn_attention_logits_cap_type_tanh. | |
| /// @param query_id - Value ID for the query tensor. The query tensor must be a 3+-dimensional tensor defined in the | |
| /// @a subgraph with the dimensions as [*, H, T, C], where H/T/C are the heads/tokens/channels, and * | |
| /// is the 0 or more dimensions treated as batch size. | |
| /// @param key_id - Value ID for the key tensor. The key tensor must be a 2+--dimensional tensor defined in the | |
| /// @a subgraph. It can have the same number of dimensions as the query, with the dimensions as | |
| /// [*, H, U, C] (multi-head), or have 1 less dimension than the query, with the dimensions as | |
| /// as [*, U, C] (multi-query, number of heads omitted implies single head), where H/U/C are the | |
| /// heads/key_value_tokens/channels, and * is the 0 or more dimensions treated as batch size. These | |
| /// batch size dimensions must be the same as query. | |
| /// @param value_id - Value ID for the value tensor. The value tensor must be a 2+--dimensional tensor defined in the | |
| /// @a subgraph. It can have the same number of dimensions as the query, with the dimensions as | |
| /// [*, H, U, D] (multi-head), or have 1 less dimension than the query, with the dimensions as | |
| /// as [*, U, D] (multi-query, number of heads omitted implies single head), where H/U/D are the | |
| /// heads/key_value_tokens/value_channels, and * is the 0 or more dimensions treated as batch size. | |
| /// These batch size dimensions must be the same as query and key. | |
| /// @param scale_id - Value ID for the scale tensor. The scale tensor must be a 1D tensor defined in the @a subgraph | |
| /// with [C] dimensions. The query tensor is multiplied with this scale tensor before the dot product | |
| /// with the key tensor. | |
| /// @param mask_id - Value ID for the mask tensor. The mask tensor must be a 2D tensor defined in the @a subgraph with | |
| /// [T, U] dimensions. The mask tensor is added to the logits (query dot value). | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 3+-dimensional tensor defined in the | |
| /// @a subgraph with the dimensions as [*, H, T, D], where H/T/D are the heads/tokens/value_channels, | |
| /// and * is the 0 or more dimensions treated as batch size. These batch size dimensions must be the | |
| /// same as query, key, and value. | |
| /// @param flags - binary features of the Scaled Dot Product Attention Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_scaled_dot_product_attention( | |
| xnn_subgraph_t subgraph, | |
| enum xnn_attention_logits_cap_type cap_type, | |
| const void* cap_params, | |
| uint32_t query_id, | |
| uint32_t key_id, | |
| uint32_t value_id, | |
| uint32_t scale_id, | |
| uint32_t mask_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Subtract Node and add it to a Subgraph. | |
| /// | |
| /// The Subtract Node computes elementwise subtraction of two tensor inputs with numpy broadcasting rules. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the second | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an M-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the first | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a max(N,M)-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the maximum between the corresponding dimension | |
| /// of the two inputs. | |
| /// @param flags - binary features of the Subtract Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_subtract( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Divide Node and add it to a Subgraph. | |
| /// | |
| /// The Divide Node computes elementwise division of two tensor inputs with numpy broadcasting rules. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the second | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an M-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the first | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a max(N,M)-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the maximum between the corresponding dimension | |
| /// of the two inputs. | |
| /// @param flags - binary features of the Divide Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_divide( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2-Input Maximum Node and add it to a Subgraph. | |
| /// | |
| /// The 2-Input Maximum Node computes elementwise maximum of two tensor inputs with numpy broadcasting rules. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the second | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an M-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the first | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a max(N,M)-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the maximum between the corresponding dimension | |
| /// of the two inputs. | |
| /// @param flags - binary features of the Maximum Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_maximum2( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2-Input Minimum Node and add it to a Subgraph. | |
| /// | |
| /// The 2-Input Minimum Node computes elementwise minimum of two tensor inputs with numpy broadcasting rules. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the second | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an M-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the first | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a max(N,M)-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the maximum between the corresponding dimension | |
| /// of the two inputs. | |
| /// @param flags - binary features of the Minimum Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_minimum2( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Squared Difference Node and add it to a Subgraph. | |
| /// | |
| /// The Squared Difference Node computes elementwise squared difference of two tensor inputs with numpy broadcasting | |
| /// rules. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the second | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an M-dimensional tensor defined in | |
| /// the @a subgraph with each dimension either equal to the corresponding dimension of the first | |
| /// input, or equal to 1. In the latter case, the elements of the input tensor are broadcasted along | |
| /// that dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a max(N,M)-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the maximum between the corresponding dimension | |
| /// of the two inputs. | |
| /// @param flags - binary features of the Squared Difference Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_squared_difference( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Constant Pad Node with static padding specification and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param pre_paddings - number of padding elements to insert before input elements for every dimension. This array | |
| /// must have as many elements as the number of dimensions in the input tensor. | |
| /// @param post_paddings - number of padding elements to insert after input elements for every dimension. This array | |
| /// must have as many elements as the number of dimensions in the input tensor. | |
| /// @param padding_value - constant value used to initialize padding elements. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor with padding. | |
| /// @param flags - binary features of the Constant Pad Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_static_constant_pad( | |
| xnn_subgraph_t subgraph, | |
| const size_t* pre_paddings, | |
| const size_t* post_paddings, | |
| float padding_value, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Expand Dims Node with and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param num_new_axes - number of new axes of size 1 to be inserted. | |
| /// @param new_axes - The axis positions of the new axes in the expanded dimensions. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor with padding. | |
| /// @param flags - binary features of the Constant Pad Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_static_expand_dims( | |
| xnn_subgraph_t subgraph, | |
| size_t num_new_axes, | |
| const size_t* new_axes, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Mean Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param num_reduction_axes - number of axes along which mean is computed. | |
| /// @param reduction_axes - axes along which mean is computed. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a dense tensor with at least | |
| /// @a num_reduction_axes dimensions defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a dense tensor defined in the | |
| /// @a subgraph with @a num_reduction_axes fewer dimensions than the input tensor (if | |
| /// XNN_FLAG_KEEP_DIMS is not specified), or has same dimension rank but the dimension at | |
| /// @a reduction_axes reduced to 1 (if XNN_FLAG_KEEP_DIMS is specified). | |
| /// @param flags - binary features of the Mean Node. The only currently supported value is XNN_FLAG_KEEP_DIMS | |
| XNN_DEPRECATED enum xnn_status xnn_define_static_mean( | |
| xnn_subgraph_t subgraph, | |
| size_t num_reduction_axes, | |
| const size_t* reduction_axes, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| enum xnn_reduce_operator { | |
| xnn_reduce_invalid = -1, | |
| xnn_reduce_sum, | |
| xnn_reduce_mean, | |
| }; | |
| /// Define a Reduce Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param num_reduction_axes - number of axes along which reduce is computed. | |
| /// @param reduction_axes - axes along which reduce is computed. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a dense tensor with at least | |
| /// @a num_reduction_axes dimensions defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a dense tensor defined in the | |
| /// @a subgraph with @a num_reduction_axes fewer dimensions than the input tensor (if | |
| /// XNN_FLAG_KEEP_DIMS is not specified), or has same dimension rank but the dimension at | |
| /// @a reduction_axes reduced to 1 (if XNN_FLAG_KEEP_DIMS is specified). | |
| /// @param flags - binary features of the Reduce Node. The only currently supported value is XNN_FLAG_KEEP_DIMS | |
| enum xnn_status xnn_define_static_reduce( | |
| xnn_subgraph_t subgraph, | |
| enum xnn_reduce_operator reduce_operator_type, | |
| size_t num_reduction_axes, | |
| const size_t* reduction_axes, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Reduce Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param num_reduction_axes - number of axes along which reduce is computed. | |
| /// @param reduction_axes - axes along which reduce is computed. Negative values | |
| /// are interpreted as offsets from @a | |
| /// num_reduction_axes. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a | |
| /// dense tensor with at least @a num_reduction_axes | |
| /// dimensions defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be | |
| /// a dense tensor defined in the @a subgraph with @a | |
| /// num_reduction_axes fewer dimensions than the input tensor | |
| /// (if XNN_FLAG_KEEP_DIMS is not specified), or has same | |
| /// dimension rank but the dimension at | |
| /// @a reduction_axes reduced to 1 (if XNN_FLAG_KEEP_DIMS is | |
| /// specified). | |
| /// @param flags - binary features of the Reduce Node. The only currently | |
| /// supported value is XNN_FLAG_KEEP_DIMS | |
| enum xnn_status xnn_define_static_reduce_v2( // | |
| xnn_subgraph_t subgraph, // | |
| enum xnn_reduce_operator reduce_operator_type, // | |
| size_t num_reduction_axes, // | |
| const int64_t* reduction_axes, // | |
| uint32_t input_id, // | |
| uint32_t output_id, // | |
| uint32_t flags); | |
| /// Define a 2-Input Concatenate Node and add it to a Subgraph. | |
| /// | |
| /// The 2-Input Concatenate Node concatenates two tensors along a specified axis. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param axis - the axis to concatenate the two input tensors along. If this is less than zero, the number of | |
| /// dimensions is added to it. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// second input. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// first input. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a N-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the dimension of both inputs, except the axis | |
| /// dimension, where it is the sum of the corresponding dimensions of both inputs. | |
| /// @param flags - binary features of the Concatenate Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_concatenate2( | |
| xnn_subgraph_t subgraph, | |
| int32_t axis, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 3-Input Concatenate Node and add it to a Subgraph. | |
| /// | |
| /// The 3-Input Concatenate Node concatenates three tensors along a specified axis. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param axis - the axis to concatenate the two input tensors along. If this is less than zero, the number of | |
| /// dimensions is added to it. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param input3_id - Value ID for the third input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a N-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the dimension of all inputs, except the axis | |
| /// dimension, where it is the sum of the corresponding dimensions of all inputs. | |
| /// @param flags - binary features of the Concatenate Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_concatenate3( | |
| xnn_subgraph_t subgraph, | |
| int32_t axis, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t input3_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 4-Input Concatenate Node and add it to a Subgraph. | |
| /// | |
| /// The 4-Input Concatenate Node concatenates four tensors along a specified axis. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param axis - the axis to concatenate the two input tensors along. If this is less than zero, the number of | |
| /// dimensions is added to it. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param input3_id - Value ID for the third input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param input4_id - Value ID for the fourth input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a N-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the dimension of all inputs, except the axis | |
| /// dimension, where it is the sum of the corresponding dimensions of all inputs. | |
| /// @param flags - binary features of the Concatenate Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_concatenate4( | |
| xnn_subgraph_t subgraph, | |
| int32_t axis, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t input3_id, | |
| uint32_t input4_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 5-Input Concatenate Node and add it to a Subgraph. | |
| /// | |
| /// The 5-Input Concatenate Node concatenates four tensors along a specified axis. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param axis - the axis to concatenate the two input tensors along. If this is less than zero, the number of | |
| /// dimensions is added to it. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param input3_id - Value ID for the third input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param input4_id - Value ID for the fourth input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param input5_id - Value ID for the fourth input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph with each dimension, except the axis, equal to the corresponding dimension of the | |
| /// other inputs. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a N-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to the dimension of all inputs, except the axis | |
| /// dimension, where it is the sum of the corresponding dimensions of all inputs. | |
| enum xnn_status xnn_define_concatenate5( | |
| xnn_subgraph_t subgraph, | |
| int32_t axis, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t input3_id, | |
| uint32_t input4_id, | |
| uint32_t input5_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Copy Sign Node and add it to a Subgraph. | |
| /// | |
| /// The Copy Sign Node copies the sign of the second input to the first input. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. | |
| /// @param flags - binary features of the Copy Sign Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_copysign( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Copy Node and add it to a Subgraph. | |
| /// | |
| /// The Copy Node copies an input tensor to an output tensor. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the first input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Copy Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_copy( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2-Output Split Node and add it to a Subgraph. | |
| /// | |
| /// The 2-Output Split Node splits an input tensor into two output tensors along a specified axis evenly. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param split_dim - the dimension to split the input tensor along. If this is less than zero, the number of | |
| /// dimensions is added to it. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be an N-dimensional tensor defined in the @a | |
| /// subgraph. | |
| /// @param output1_id - Value ID for the first output tensor. The output tensor must be an N-dimensional tensor defined | |
| /// in the @a subgraph with each dimension, except the axis, equal to the corresponding dimension | |
| /// of the second output. The split_dim dimension is half of the input's split_dim. | |
| /// @param output2_id - Value ID for the second output tensor. The output tensor must be an N-dimensional tensor | |
| /// defined in the @a subgraph with each dimension, except the axis, equal to the corresponding | |
| /// dimension of the first output. The split_dim dimension is half of the input's split_dim. | |
| /// @param flags - binary features of the Split Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_even_split2( | |
| xnn_subgraph_t subgraph, | |
| int32_t split_dim, | |
| uint32_t input_id, | |
| uint32_t output1_id, | |
| uint32_t output2_id, | |
| uint32_t flags); | |
| /// Define a 3-Output Split Node and add it to a Subgraph. | |
| /// | |
| /// The 3-Output Split Node splits an input tensor into three output tensors along a specified axis evenly. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param split_dim - the dimension to split the input tensor along. If this is less than zero, the number of | |
| /// dimensions is added to it. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be an N-dimensional tensor defined in the @a | |
| /// subgraph. | |
| /// @param output1_id - Value ID for the first output tensor. The output tensor must be an N-dimensional tensor defined | |
| /// in the @a subgraph with each dimension, except the axis, equal to the corresponding dimension | |
| /// of the second and third output. The split_dim dimension is one third of the input's split_dim. | |
| /// @param output2_id - Value ID for the second output tensor. The output tensor must be an N-dimensional tensor | |
| /// defined in the @a subgraph with each dimension, except the axis, equal to the corresponding | |
| /// dimension of the first and third output. The split_dim dimension is one third of the input's | |
| /// split_dim. | |
| /// @param output3_id - Value ID for the third output tensor. The output tensor must be an N-dimensional tensor | |
| /// defined in the @a subgraph with each dimension, except the axis, equal to the corresponding | |
| /// dimension of the second and third output. The split_dim dimension is one third of the input's | |
| /// split_dim. | |
| /// @param flags - binary features of the Split Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_even_split3( | |
| xnn_subgraph_t subgraph, | |
| int32_t split_dim, | |
| uint32_t input_id, | |
| uint32_t output1_id, | |
| uint32_t output2_id, | |
| uint32_t output3_id, | |
| uint32_t flags); | |
| /// Define a 4-Output Split Node and add it to a Subgraph. | |
| /// | |
| /// The 4-Output Split Node splits an input tensor into four output tensors along a specified axis evenly. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param split_dim - the dimension to split the input tensor along. If this is less than zero, the number of | |
| /// dimensions is added to it. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be an N-dimensional tensor defined in the @a | |
| /// subgraph. | |
| /// @param output1_id - Value ID for the first output tensor. The output tensor must be an N-dimensional tensor defined | |
| /// in the @a subgraph with each dimension, except the axis, equal to the corresponding dimension | |
| /// of the other output tensors. The split_dim dimension is one fourth of the input's split_dim. | |
| /// @param output2_id - Value ID for the second output tensor. The output tensor must be an N-dimensional tensor | |
| /// defined in the @a subgraph with each dimension, except the axis, equal to the corresponding | |
| /// dimension of the other output tensors. The split_dim dimension is one fourth of the input's | |
| /// split_dim. | |
| /// @param output3_id - Value ID for the third output tensor. The output tensor must be an N-dimensional tensor | |
| /// defined in the @a subgraph with each dimension, except the axis, equal to the corresponding | |
| /// dimension of the other output tensors. The split_dim dimension is one fourth of the input's | |
| /// split_dim. | |
| /// @param output4_id - Value ID for the fourth output tensor. The output tensor must be an N-dimensional tensor | |
| /// defined in the @a subgraph with each dimension, except the axis, equal to the corresponding | |
| /// dimension of the other output tensors. The split_dim dimension is one fourth of the input's | |
| /// split_dim. | |
| /// @param flags - binary features of the Split Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_even_split4( | |
| xnn_subgraph_t subgraph, | |
| int32_t split_dim, | |
| uint32_t input_id, | |
| uint32_t output1_id, | |
| uint32_t output2_id, | |
| uint32_t output3_id, | |
| uint32_t output4_id, | |
| uint32_t flags); | |
| /// Define a Reshape Node with static shape specification and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param num_dims - number of shape dimensions in the output tensor. | |
| /// @param new_shape - shape dimensions of the output tensor. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor with padding. | |
| /// @param flags - binary features of the Reshape Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_static_reshape( | |
| xnn_subgraph_t subgraph, | |
| size_t num_dims, | |
| const size_t* new_shape, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a 2D Resize Bilinear Node with static output height & width specification and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param new_height - height dimension of the output tensor. | |
| /// @param new_width - width dimension of the output tensor. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, H, W, C] dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, new_height, new_width, C] dimensions. | |
| /// @param flags - binary features of the 2D Resize Bilinear Node. The only currently supported values are | |
| /// XNN_FLAG_TENSORFLOW_LEGACY_MODE and XNN_FLAG_ALIGN_CORNERS, which are mutually exclusive. | |
| enum xnn_status xnn_define_static_resize_bilinear_2d( | |
| xnn_subgraph_t subgraph, | |
| size_t new_height, | |
| size_t new_width, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a PReLU (Parametric ReLU) Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, H, W, channels] dimensions. | |
| /// @param slope_id - Value ID for the slope tensor. The slope tensor must be a 1D tensor defined in the @a subgraph with | |
| /// either [1] or [channels] dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, H, W, channels] dimensions. | |
| /// @param flags - binary features of the PReLU Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_prelu( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t slope_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a RoPE (Rotary Positional Embeddings) Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param max_tokens - deprecated. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [batch, tokens, heads, channels] dimensions. | |
| /// @param weights_id - Value ID for the weights tensor. The weights tensor must be a 2D tensor defined in the | |
| /// @a subgraph with [max_tokens, channels] dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [batch, tokens, heads, channels] dimensions. | |
| /// @param flags - binary features of the RoPE Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_rope( | |
| xnn_subgraph_t subgraph, | |
| size_t max_sequence_size, | |
| uint32_t input_id, | |
| uint32_t weights_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Abs Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Abs Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_abs( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Bankers' Rounding Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Bankers' Rounding Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_bankers_rounding( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Batch Matrix Multiply Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input1_id - Value ID for the first input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph. It must be at least 3D. The first N-2 dimensions must match the second input | |
| /// tensor. The last 2 dimensions are [M, K]. If XNN_FLAG_TRANSPOSE_B is not specified, the last | |
| /// dimension must match the second last dimension of the second input tensor. If | |
| /// XNN_FLAG_TRANSPOSE_B is specified, the last dimension must match the last dimension of the | |
| /// second input tensor. | |
| /// @param input2_id - Value ID for the second input tensor. The input tensor must be an N-dimensional tensor defined | |
| /// in the @a subgraph. It must be at least 3D. The first N-2 dimensions must match the first input | |
| /// tensor. If XNN_FLAG_TRANSPOSE_B is not specified, the last 2 dimensions are [K, N], and the | |
| /// second last dimension must match the last dimension of the first input tensor. If | |
| /// XNN_FLAG_TRANSPOSE_B is specified, the last 2 dimensions are [N, K], and the last dimension must | |
| /// match the last dimension of the first input tensor. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be an N-dimensional tensor defined in the | |
| /// @a subgraph. It must be at least 3D. The first N-2 dimensions must match the first and second | |
| /// input tensors . The last 2 dimensions must be [M, N]. | |
| /// @param flags - binary features of the Batch Matrix Multiply Node. The only currently supported value is | |
| /// XNN_FLAG_TRANSPOSE_B. | |
| enum xnn_status xnn_define_batch_matrix_multiply( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input1_id, | |
| uint32_t input2_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Ceiling Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Ceiling Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_ceiling( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Clamp Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param output_min - lower bound for clipping output values. | |
| /// @param output_max - upper bound for clipping output values. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Clamp Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_clamp( | |
| xnn_subgraph_t subgraph, | |
| float output_min, | |
| float output_max, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define an ELU (Exponential Linear Unit) Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param alpha - scale factor for negative output elements. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the ELU Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_elu( | |
| xnn_subgraph_t subgraph, | |
| float alpha, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Exp Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Exp Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_exp( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Floor Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Floor Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_floor( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define an GELU (Gaussian Error Linear Unit) Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the GELU Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_gelu( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a HardSwish Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the HardSwish Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_hardswish( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Leaky ReLU Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param negative_slope - scale factor for negative input elements. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Leaky ReLU Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_leaky_relu( | |
| xnn_subgraph_t subgraph, | |
| float negative_slope, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Log Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Log Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_log( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Negate Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Negate Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_negate( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Sigmoid Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Sigmoid Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_sigmoid( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a SoftMax Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph, and have at | |
| /// least one dimension. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the SoftMax Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_softmax( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Space To Depth 2D Node and add it to a Subgraph. | |
| /// | |
| /// The Space To Depth 2D Node rearranges blocks of spatial data into blocks (a reverse transform to Depth To Space 2D). | |
| /// For a given input pixel, an output square of pixels with side @a block_size is formed from values in the | |
| /// corresponding number of its channels. The output depth is therefore @a block_size x @a block_size times greater | |
| /// than that of the input. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param block_size - the size of the spatial block. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, IH * block_size, IW * block_size, OC] dimensions. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be a 4D tensor defined in the @a subgraph | |
| /// with [N, IH, IW, OC * block_size * block_size] dimensions. | |
| /// @param flags - binary features of the input_channels Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_space_to_depth_2d( | |
| xnn_subgraph_t subgraph, | |
| uint32_t block_size, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Square Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Square Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_square( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Square Root Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Square Root Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_square_root( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Reciprocal Square Root Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be | |
| /// defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be | |
| /// defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Square Root Node. No supported flags | |
| /// are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_reciprocal_square_root( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| enum xnn_status xnn_define_static_slice( | |
| xnn_subgraph_t subgraph, | |
| size_t num_dims, | |
| const size_t* offsets, | |
| const size_t* sizes, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Static Slice Node add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param num_dims - number of shape dimensions in the input and output tensor. | |
| /// @param offsets - offsets in each dimension of the input tensor. This array must have @a num_dims elements. Can be | |
| /// negative meaning that the offset is relative to the end of the dimension. | |
| /// @param sizes - size of each dimension in output tensor. This array must have @a num_dims elements. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// dimensions must match @a sizes. | |
| /// @param flags - binary features of the Static Slice Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_static_slice_v2( // | |
| xnn_subgraph_t subgraph, // | |
| size_t num_dims, // | |
| const int64_t* offsets, // | |
| const size_t* sizes, // | |
| uint32_t input_id, // | |
| uint32_t output_id, // | |
| uint32_t flags); | |
| /// Define a Static Transpose Node and add it to a Subgraph. | |
| /// | |
| /// The Static Transpose Node applies a generalized transpose to the input tensor using the permuation in perm. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be an N-dimensional tensor defined in | |
| /// the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be an N-dimensional tensor defined | |
| /// in the @a subgraph with each dimension equal to its corresponding permuted input dimension. | |
| /// @param num_dims - the number of permutation dimensions. This must be equal to the number of input dimensions. | |
| /// @param perm - The permutation of the axis of the input tensor. The perm array must must contain 0 to N-1 in the | |
| /// permuted order. | |
| /// @param flags - binary features of the Static Transpose Node. No supported flags are currently defined. | |
| enum xnn_status xnn_define_static_transpose( | |
| xnn_subgraph_t subgraph, | |
| size_t num_dims, | |
| const size_t* perm, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Define a Tanh Node and add it to a Subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object that will own the created Node. | |
| /// @param input_id - Value ID for the input tensor. The input tensor must be defined in the @a subgraph. | |
| /// @param output_id - Value ID for the output tensor. The output tensor must be defined in the @a subgraph, and its | |
| /// shape must match the shape of the input tensor. | |
| /// @param flags - binary features of the Tanh Node. No supported flags are currently defined. | |
| XNN_DEPRECATED enum xnn_status xnn_define_tanh( | |
| xnn_subgraph_t subgraph, | |
| uint32_t input_id, | |
| uint32_t output_id, | |
| uint32_t flags); | |
| /// Code cache is a cache for JIT generated code. | |
| typedef struct xnn_code_cache* xnn_code_cache_t; | |
| /// Weights cache can be finalized in these ways: | |
| enum xnn_weights_cache_finalization_kind { | |
| /// Weights cache is finalized, no insert operations into the weights cache is allowed, even if the "inserted" | |
| /// weights already exist in thee cache. Weights cache memory will also be trimmed to page boundary and set to | |
| /// read-only (to prevent writes). | |
| xnn_weights_cache_finalization_kind_hard, | |
| /// Weights cache will be finalized with some extra space at the end, this allows for "inserting" into the cache only | |
| /// if the weights are already in the cache, and errors on inserting uncached weights. There is memory overhead. | |
| xnn_weights_cache_finalization_kind_soft, | |
| }; | |
| /// A combination of multiple factors to uniquely locate the weights cache. | |
| struct xnn_weights_cache_look_up_key { | |
| /// The unique seed for each ukernel. It is guaranteed that each ukernel provides | |
| /// a consistent and identical seed. | |
| uint32_t seed; | |
| /// Pointer to the original kernel. | |
| const void* kernel; | |
| /// Pointer to the original bias, could be NULL. | |
| const void* bias; | |
| }; | |
| /// A group of function pointers to manage weights cache. All functions may be | |
| /// called on multi threads. | |
| struct xnn_weights_cache_provider { | |
| /// User-specified pointer that will be passed as-is to all functions in this | |
| /// structure. | |
| void* context; | |
| /// Looks up the tuple of {cache_key, kernel, bias} in the cache. If it is found, | |
| /// returns the offset to the found entry for reuse. Otherwise, returns SIZE_MAX. | |
| /// @param context - The user-specified pointer from xnn_weights_cache_provider structure. | |
| /// @param cache_key - The key used to locate the weights cache entry. | |
| size_t (*look_up)(void* context, const struct xnn_weights_cache_look_up_key* cache_key); | |
| /// Ensures that cache has enough space for `n` bytes. Returns the address to | |
| /// store weight cache. Returns NULL if fails to reserve space. | |
| /// @param context - The user-specified pointer from xnn_weights_cache_provider structure. | |
| /// @param n - size to be reserved. | |
| void* (*reserve_space)(void* context, size_t n); | |
| /// Looks up packed weights at `ptr` in the cache. If it is found, reuse it. | |
| /// Otherwise, it is added to the cache. Returns the offset to the cache. | |
| /// @param context - The user-specified pointer from xnn_weights_cache_provider structure. | |
| /// @param cache_key - The key used to locate the weights cache entry. | |
| /// @param ptr - pointer pointing to the packed weight. | |
| /// @param size - size of the packed weight. | |
| size_t (*look_up_or_insert)(void* context, const struct xnn_weights_cache_look_up_key* cache_key, void* ptr, size_t size); | |
| /// Returns whether the cache is finalized. | |
| /// @param context - The user-specified pointer from xnn_weights_cache_provider structure. | |
| bool (*is_finalized)(void* context); | |
| /// Returns the absolute pointer corresponding to `offset`, where the offset is returned from | |
| /// `look_up` or `get_or_insert`. This function must be called after finalize. | |
| /// @param context - The user-specified pointer from xnn_weights_cache_provider structure. | |
| /// @param offset - offset to the start of internal buffer | |
| void* (*offset_to_addr)(void* context, size_t offset); | |
| /// Destroy a weights cache object, as well as memory used for the cache. | |
| /// @param context - The user-specified pointer from xnn_weights_cache_provider structure. | |
| enum xnn_status (*delete_cache)(void* context); | |
| }; | |
| /// Weights cache is a cache for packed weights. It can be reused between runtimes. | |
| typedef struct xnn_weights_cache_provider* xnn_weights_cache_t; | |
| /// Create a weights cache object specifying the initial size of weights cache (in bytes). | |
| /// | |
| /// @param[in] size - initial capacity of the weights cache (in bytes), i.e. it can hold size bytes without growing. | |
| /// @param weights_cache_out - pointer to the variable that will be initialized to a handle to the weights cache provider | |
| /// upon successful return. Once created, the weights cache provider can be shared between | |
| /// different Runtime objects. | |
| enum xnn_status xnn_create_weights_cache_with_size(size_t size, xnn_weights_cache_t* weights_cache_out); | |
| enum xnn_status xnn_create_weights_cache(xnn_weights_cache_t* weights_cache_out); | |
| /// Finalizes the weights cache. The kind of finalization is specified by `finalization_kind`. | |
| /// @param weights_cache - the weights cache object to finalize. | |
| /// @param finalization_kind - the kind of finalization. | |
| enum xnn_status xnn_finalize_weights_cache( | |
| xnn_weights_cache_t weights_cache, | |
| enum xnn_weights_cache_finalization_kind finalization_kind); | |
| // Wrapper function of the function pointers in `xnn_weights_cache_t`. | |
| bool xnn_weights_cache_is_finalized(xnn_weights_cache_t cache); | |
| /// Destroy a weights cache object, as well as memory used for the cache. | |
| /// @param weights_cache - the weights cache object to destroy. | |
| enum xnn_status xnn_delete_weights_cache(xnn_weights_cache_t weights_cache); | |
| typedef struct xnn_workspace* xnn_workspace_t; | |
| /// Create a workspace object. | |
| /// @param workspace_out - pointer to the variable that will be initialized to a handle to the workspace object upon | |
| /// successful return. Once created, the workspace can be shared between different Runtime | |
| /// objects. | |
| enum xnn_status xnn_create_workspace(xnn_workspace_t* workspace_out); | |
| /// Destroy a workspace object, as well as memory used by the workspace. Object destruction can be deferred until all | |
| /// Runtime objects created with this workspace are destroyed. | |
| /// @param workspace - the workspace object to destroy. | |
| enum xnn_status xnn_release_workspace(xnn_workspace_t workspace); | |
| /// Runtime is a combination of an execution plan for subgraph Nodes and a memory manager for subgraph Values. | |
| typedef struct xnn_runtime* xnn_runtime_t; | |
| enum xnn_profile_info { | |
| /// Returns a size_t containing the number of operators. | |
| xnn_profile_info_num_operators, | |
| /// Returns a char[] containing the null character separated names of all operators. | |
| xnn_profile_info_operator_name, | |
| /// Returns a uint64_t[] with the runtimes of all operators in the same order as xnn_profile_info_operator_name. | |
| xnn_profile_info_operator_timing, | |
| }; | |
| /// Return profile information for all operators. | |
| /// | |
| /// @param runtime - a Runtime object created with @ref xnn_create_runtime, @ref xnn_create_runtime_v2 or | |
| /// @ref xnn_create_runtime_v3. | |
| /// @param param_name - type of profile information required. | |
| /// @param param_value_size - the size in bytes of memory pointed to by param_value. If this is not sufficient then | |
| /// param_value_size_ret will be set to the required size and xnn_status_out_of_memory will be | |
| /// returned. | |
| /// @param param_value - a pointer to memory location where appropriate values for a given param_value will be written. | |
| /// @param param_value_size_ret - returns number of bytes required to write the result if param_value_size is not | |
| /// sufficient. | |
| enum xnn_status xnn_get_runtime_profiling_info(xnn_runtime_t runtime, | |
| enum xnn_profile_info param_name, | |
| size_t param_value_size, | |
| void* param_value, | |
| size_t* param_value_size_ret); | |
| /// Create a Runtime object from a subgraph. | |
| /// | |
| /// @param subgraph - a Subgraph object with all Values and Nodes that would be handled by the runtime. No Values or | |
| /// Nodes can be added to the runtime once it is constructed. | |
| /// @param weights_cache - a cache for packed weights. The runtime will look up and reuse packed weights in this cache, | |
| /// this will reduce memory allocated for packed weights. | |
| /// @param workspace - a workspace to hold internal tensors. The runtime will allocate space used for internal tensors | |
| /// and track them using workspace. Workspace can be shared and reused across different runtimes. If | |
| /// workspace is NULL, there will be no sharing: each runtime has its own workspace. | |
| /// @param threadpool - the thread pool to be used for parallelisation of computations in the runtime. If the thread | |
| /// pool is NULL, the computation would run on the caller thread without parallelization. | |
| /// @param flags - binary features of the runtime. The only currently supported values are | |
| /// XNN_FLAG_HINT_SPARSE_INFERENCE, XNN_FLAG_HINT_FP16_INFERENCE, XNN_FLAG_FORCE_FP16_INFERENCE, | |
| /// XNN_FLAG_YIELD_WORKERS, and XNN_FLAG_TRANSIENT_INDIRECTION_BUFFER. If XNN_FLAG_YIELD_WORKERS is | |
| /// specified, worker threads would be yielded to the system scheduler after processing the last operator | |
| /// in the Runtime. If XNN_FLAG_TRANSIENT_INDIRECTION_BUFFER is specified, convolution operators will | |
| /// initialize indirection buffers on each inference run using temporary memory in the workspace, instead | |
| /// of initializing persistent indirection buffers once. | |
| /// @param runtime_out - pointer to the variable that will be initialized with a handle to the Runtime object upon | |
| /// successful return. Once constructed, the Runtime object is independent of the Subgraph object | |
| /// used to create it. | |
| enum xnn_status xnn_create_runtime_v4( | |
| xnn_subgraph_t subgraph, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_workspace_t workspace, | |
| pthreadpool_t threadpool, | |
| uint32_t flags, | |
| xnn_runtime_t* runtime_out); | |
| enum xnn_status xnn_create_runtime_v3( | |
| xnn_subgraph_t subgraph, | |
| xnn_weights_cache_t weights_cache, | |
| pthreadpool_t threadpool, | |
| uint32_t flags, | |
| xnn_runtime_t* runtime_out); | |
| enum xnn_status xnn_create_runtime_v2( | |
| xnn_subgraph_t subgraph, | |
| pthreadpool_t threadpool, | |
| uint32_t flags, | |
| xnn_runtime_t* runtime_out); | |
| enum xnn_status xnn_create_runtime( | |
| xnn_subgraph_t subgraph, | |
| xnn_runtime_t* runtime_out); | |
| struct xnn_external_value { | |
| uint32_t id; | |
| void* data; | |
| }; | |
| /// Reshape an external value. | |
| /// | |
| /// @param external_id - external ID for the Value. The ID must be within the range of reversed Value IDs specified on | |
| /// the Subgraph creation. If the external ID is XNN_INVALID_VALUE_ID, an internal ID will be | |
| /// created for the Value. | |
| /// @param num_dims - number of dimensions in the shape. | |
| /// @param dims - pointer to an array of @a num_dims shape dimensions. If num_dims is 0, this pointer can be NULL. | |
| /// XNNPACK does not keep any pointers to this array after the function returns. | |
| enum xnn_status xnn_reshape_external_value( | |
| xnn_runtime_t runtime, | |
| uint32_t external_id, | |
| size_t num_dims, | |
| const size_t* dims); | |
| /// Get the external value shape. | |
| /// | |
| /// @param external_id - external ID for the Value. The ID must be within the range of reversed Value IDs specified on | |
| /// the Subgraph creation. The external ID can not be XNN_INVALID_VALUE_ID. | |
| /// @param num_dims - A valid pointer into which the number of dimensions in the shape will be written. It can not be larger than XNN_MAX_TENSOR_DIMS. | |
| /// @param dims - pointer to an array of @a num_dims shape dimensions. This pointer can't be NULL. It must be large enough to hold | |
| /// at least @a num_dims elements. XNNPACK does not keep any pointers to this array after the function returns. | |
| enum xnn_status xnn_get_external_value_shape( | |
| xnn_runtime_t runtime, | |
| uint32_t external_id, | |
| size_t* num_dims, | |
| size_t* dims); | |
| /// Reshape the XNNPACK runtime. | |
| /// | |
| /// Propagates the shapes of input tensors through the graph to determine the shapes of intermediate and output tensors. | |
| /// Memory is allocated if required. Output tensor shapes are returned by xnn_get_external_value_shape. | |
| /// | |
| /// @param runtime - a Runtime object created with @ref xnn_create_runtime or @ref xnn_create_runtime_v2. | |
| enum xnn_status xnn_reshape_runtime( | |
| xnn_runtime_t runtime); | |
| /// Deprecated. Use xnn_reshape_runtime and xnn_setup_runtime_v2. | |
| /// | |
| /// Setup data pointers for external inputs and outputs in a Runtime object and | |
| /// allocate memory. | |
| /// | |
| /// @param runtime - a Runtime object created with @ref xnn_create_runtime or @ref xnn_create_runtime_v2. | |
| /// @param num_external_values - the number of external inputs and outputs specified in this call. This number must | |
| /// match the number of external inputs and outputs in the runtime, i.e. all external | |
| /// inputs and outputs in the runtime must be specified in one call. | |
| /// @param external_values - array with location information for all external inputs and outputs in the runtime. | |
| enum xnn_status xnn_setup_runtime( | |
| xnn_runtime_t runtime, | |
| size_t num_external_values, | |
| const struct xnn_external_value* external_values); | |
| /// Setup data pointers for external inputs and outputs in a Runtime object. | |
| /// Should be called after xnn_reshape_runtime. | |
| /// | |
| /// @param runtime - a Runtime object created with @ref xnn_create_runtime or @ref xnn_create_runtime_v2. | |
| /// @param num_external_values - the number of external inputs and outputs specified in this call. This number must | |
| /// match the number of external inputs and outputs in the runtime, i.e. all external | |
| /// inputs and outputs in the runtime must be specified in one call. | |
| /// @param external_values - array with location information for all external inputs and outputs in the runtime. | |
| enum xnn_status xnn_setup_runtime_v2( | |
| xnn_runtime_t runtime, | |
| size_t num_external_values, | |
| const struct xnn_external_value* external_values); | |
| /// Execute forward pass for all operators in the runtime. | |
| /// | |
| /// @param runtime - the Runtime object with the execution plan to invoke. | |
| enum xnn_status xnn_invoke_runtime( | |
| xnn_runtime_t runtime); | |
| /// Destroy a Runtime object, as well as operators and memory associated with it. | |
| /// | |
| /// @param runtime - the Runtime object to destroy. | |
| enum xnn_status xnn_delete_runtime( | |
| xnn_runtime_t runtime); | |
| typedef struct xnn_operator* xnn_operator_t; | |
| enum xnn_status xnn_run_operator( | |
| xnn_operator_t op, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_delete_operator( | |
| xnn_operator_t op); | |
| /// Operator API: | |
| /// - create operator will create and populate a xnn_operator_t | |
| /// - reshape operator will update fields in xnn_operator_t with shape/dimensions and parallelization information | |
| /// - setup operator will update pointers to input and outputs | |
| /// Each supported operator must have a create, reshape, and setup function. (Optionally a run function.) | |
| /// Operators listed below are in alphabetical order by operator name; within each operator, we sort alphabetically by | |
| /// data layout and type. We also group create, reshape, setup (and optionally run) functions of each operator together. | |
| enum xnn_status xnn_create_binary_elementwise_nd( | |
| enum xnn_binary_operator type, | |
| enum xnn_datatype datatype, | |
| const struct xnn_quantization_params* input1_quantization, | |
| const struct xnn_quantization_params* input2_quantization, | |
| const struct xnn_quantization_params* output_quantization, | |
| uint32_t flags, | |
| xnn_operator_t* binary_op_out); | |
| enum xnn_status xnn_reshape_binary_elementwise_nd( | |
| xnn_operator_t binary_op, | |
| size_t num_input1_dims, | |
| const size_t* input1_shape, | |
| size_t num_input2_dims, | |
| const size_t* input2_shape, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_binary_elementwise_nd( | |
| xnn_operator_t binary_op, | |
| const void* input1, | |
| const void* input2, | |
| void* output); | |
| enum xnn_status xnn_run_binary_elementwise_nd( | |
| enum xnn_binary_operator type, | |
| enum xnn_datatype datatype, | |
| const struct xnn_quantization_params* input1_quantization, | |
| const struct xnn_quantization_params* input2_quantization, | |
| const struct xnn_quantization_params* output_quantization, | |
| uint32_t flags, | |
| size_t num_input1_dims, | |
| const size_t* input1_shape, | |
| size_t num_input2_dims, | |
| const size_t* input2_shape, | |
| const void* input1, | |
| const void* input2, | |
| void* output, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_unary_elementwise_nc( | |
| enum xnn_unary_operator op_type, | |
| enum xnn_datatype input_datatype, | |
| enum xnn_datatype output_datatype, | |
| const union xnn_unary_params* params, | |
| const struct xnn_quantization_params* input_quantization, | |
| const struct xnn_quantization_params* output_quantization, | |
| uint32_t flags, | |
| xnn_operator_t* op_out); | |
| enum xnn_status xnn_reshape_unary_elementwise_nc( | |
| xnn_operator_t op, | |
| size_t batch_size, | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_unary_elementwise_nc( | |
| xnn_operator_t op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_run_unary_elementwise_nc( | |
| // create parameters | |
| enum xnn_unary_operator op_type, | |
| enum xnn_datatype input_datatype, | |
| enum xnn_datatype output_datatype, | |
| const union xnn_unary_params* params, | |
| const struct xnn_quantization_params* input_quantization, | |
| const struct xnn_quantization_params* output_quantization, | |
| uint32_t flags, | |
| // reshape parameters | |
| size_t batch_size, | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| pthreadpool_t threadpool, | |
| // setup parameters | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_argmax_pooling2d_nhwc_f32( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t flags, | |
| xnn_operator_t* argmax_pooling_op_out); | |
| enum xnn_status xnn_reshape_argmax_pooling2d_nhwc_f32( | |
| xnn_operator_t argmax_pooling_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_argmax_pooling2d_nhwc_f32( | |
| xnn_operator_t argmax_pooling_op, | |
| void* workspace, | |
| const float* input, | |
| float* output, | |
| uint32_t* index); | |
| enum xnn_status xnn_create_average_pooling2d_nhwc_f16( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_operator_t* average_pooling_op_out); | |
| enum xnn_status xnn_reshape_average_pooling2d_nhwc_f16( | |
| xnn_operator_t average_pooling_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_average_pooling2d_nhwc_f16( | |
| xnn_operator_t average_pooling_op, | |
| void* workspace, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_average_pooling2d_nhwc_f32( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_operator_t* average_pooling_op_out); | |
| enum xnn_status xnn_reshape_average_pooling2d_nhwc_f32( | |
| xnn_operator_t average_pooling_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_average_pooling2d_nhwc_f32( | |
| xnn_operator_t average_pooling_op, | |
| void* workspace, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_average_pooling2d_nhwc_qu8( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint8_t input_zero_point, | |
| float input_scale, | |
| uint8_t output_zero_point, | |
| float output_scale, | |
| uint8_t output_min, | |
| uint8_t output_max, | |
| uint32_t flags, | |
| xnn_operator_t* average_pooling_op_out); | |
| enum xnn_status xnn_reshape_average_pooling2d_nhwc_qu8( | |
| xnn_operator_t average_pooling_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_average_pooling2d_nhwc_qu8( | |
| xnn_operator_t average_pooling_op, | |
| void* workspace, | |
| const uint8_t* input, | |
| uint8_t* output); | |
| enum xnn_status xnn_create_batch_matrix_multiply_nc_f16( | |
| uint32_t flags, | |
| xnn_operator_t* batch_matrix_multiply_op); | |
| enum xnn_status xnn_reshape_batch_matrix_multiply_nc_f16( | |
| xnn_operator_t batch_matrix_multiply_op, size_t num_batch_dims, | |
| const size_t* batch_dims_a, const size_t* batch_dims_b, size_t m, size_t k, | |
| size_t n, size_t* workspace_size, size_t* workspace_alignment, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_batch_matrix_multiply_nc_f16( | |
| xnn_operator_t batch_matrix_multiply_op, void* workspace, | |
| const void* input_a, const void* input_b, void* output); | |
| enum xnn_status xnn_create_batch_matrix_multiply_nc_f32( | |
| uint32_t flags, xnn_operator_t* batch_matrix_multiply_op); | |
| enum xnn_status xnn_create_batch_matrix_multiply_nc_f32_const_weights( | |
| size_t batch_size_b, size_t k, size_t n, const float* data_b, | |
| uint32_t flags, xnn_operator_t* batch_matrix_multiply_op); | |
| enum xnn_status xnn_reshape_batch_matrix_multiply_nc_f32( | |
| xnn_operator_t batch_matrix_multiply_op, size_t num_batch_dims, | |
| const size_t* batch_dims_a, const size_t* batch_dims_b, size_t m, size_t k, | |
| size_t n, size_t* workspace_size, size_t* workspace_alignment, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_batch_matrix_multiply_nc_f32( | |
| xnn_operator_t batch_matrix_multiply_op, void* workspace, | |
| const float* input_a, const float* input_b, float* output); | |
| enum xnn_status xnn_create_batch_matrix_multiply_nc_qd8_f32_qc8w( | |
| size_t batch_size_b, size_t k, size_t n, const int8_t* data_b, | |
| const float* scale_b, uint32_t flags, | |
| xnn_operator_t* batch_matrix_multiply_op); | |
| enum xnn_status xnn_reshape_batch_matrix_multiply_nc_qd8_f32_qc8w( | |
| xnn_operator_t batch_matrix_multiply_op, size_t num_batch_dims, | |
| const size_t* batch_dims_a, const size_t* batch_dims_b, size_t m, size_t k, | |
| size_t n, pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_batch_matrix_multiply_nc_qd8_f32_qc8w( | |
| xnn_operator_t batch_matrix_multiply_op, const int8_t* input_a, | |
| const struct xnn_quantization_params* quantization_params, | |
| float* output); | |
| enum xnn_status xnn_create_channel_shuffle_nc_x8( | |
| size_t groups, | |
| size_t group_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| uint32_t flags, | |
| xnn_operator_t* channel_shuffle_op_out); | |
| enum xnn_status xnn_reshape_channel_shuffle_nc_x8( | |
| xnn_operator_t channel_shuffle_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_channel_shuffle_nc_x8( | |
| xnn_operator_t channel_shuffle_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_channel_shuffle_nc_x32( | |
| size_t groups, | |
| size_t group_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| uint32_t flags, | |
| xnn_operator_t* channel_shuffle_op_out); | |
| enum xnn_status xnn_reshape_channel_shuffle_nc_x32( | |
| xnn_operator_t channel_shuffle_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_channel_shuffle_nc_x32( | |
| xnn_operator_t channel_shuffle_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_constant_pad_nd_x8( | |
| const void* padding_value, | |
| uint32_t flags, | |
| xnn_operator_t* constant_pad_op_out); | |
| enum xnn_status xnn_reshape_constant_pad_nd_x8( | |
| xnn_operator_t constant_pad_op, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* pre_padding, | |
| const size_t* post_padding, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_constant_pad_nd_x8( | |
| xnn_operator_t constant_pad_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_run_constant_pad_nd_x8( | |
| uint32_t flags, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* pre_paddings, | |
| const size_t* post_paddings, | |
| const void* input, | |
| void* output, | |
| const void* padding_value, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_constant_pad_nd_x16( | |
| const void* padding_value, | |
| uint32_t flags, | |
| xnn_operator_t* constant_pad_op_out); | |
| enum xnn_status xnn_reshape_constant_pad_nd_x16( | |
| xnn_operator_t constant_pad_op, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* pre_padding, | |
| const size_t* post_padding, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_constant_pad_nd_x16( | |
| xnn_operator_t constant_pad_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_run_constant_pad_nd_x16( | |
| uint32_t flags, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* pre_paddings, | |
| const size_t* post_paddings, | |
| const void* input, | |
| void* output, | |
| const void* padding_value, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_constant_pad_nd_x32( | |
| const void* padding_value, | |
| uint32_t flags, | |
| xnn_operator_t* constant_pad_op_out); | |
| enum xnn_status xnn_reshape_constant_pad_nd_x32( | |
| xnn_operator_t constant_pad_op, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* pre_padding, | |
| const size_t* post_padding, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_constant_pad_nd_x32( | |
| xnn_operator_t constant_pad_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_run_constant_pad_nd_x32( | |
| uint32_t flags, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* pre_paddings, | |
| const size_t* post_paddings, | |
| const void* input, | |
| void* output, | |
| const void* padding_value, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_convert_nc_f16_qd8( | |
| uint32_t flags, | |
| xnn_operator_t* convert_op_out); | |
| enum xnn_status xnn_reshape_convert_nc_f16_qd8( | |
| xnn_operator_t convert_op, | |
| size_t batch_size, | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| pthreadpool_t threadpool); | |
| // quantization_params must be padded with at least XNN_EXTRA_QUANTIZATION_PARAMS entries. | |
| enum xnn_status xnn_setup_convert_nc_f16_qd8( | |
| xnn_operator_t convert_op, | |
| const void* input, | |
| int8_t* output, | |
| struct xnn_quantization_params* quantization_params); | |
| enum xnn_status xnn_create_convert_nc_f32_qd8( | |
| uint32_t flags, | |
| xnn_operator_t* convert_op_out); | |
| enum xnn_status xnn_reshape_convert_nc_f32_qd8( | |
| xnn_operator_t convert_op, | |
| size_t batch_size, | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| pthreadpool_t threadpool); | |
| // quantization_params must be padded with at least XNN_EXTRA_QUANTIZATION_PARAMS entries. | |
| enum xnn_status xnn_setup_convert_nc_f32_qd8( | |
| xnn_operator_t convert_op, | |
| const float* input, | |
| int8_t* output, | |
| struct xnn_quantization_params* quantization_params); | |
| XNN_DEPRECATED enum xnn_status xnn_run_convert_nc_f32_f16( | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| size_t batch_size, | |
| const float* input, | |
| void* output, | |
| uint32_t flags, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_convolution2d_nchw_f16( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_channel_stride, | |
| size_t output_channel_stride, | |
| const void* kernel, | |
| const void* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* convolution_op_out); | |
| enum xnn_status xnn_reshape_convolution2d_nchw_f16( | |
| xnn_operator_t convolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_convolution2d_nchw_f16( | |
| xnn_operator_t convolution_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_convolution2d_nchw_f32( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_channel_stride, | |
| size_t output_channel_stride, | |
| const float* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* convolution_op_out); | |
| enum xnn_status xnn_reshape_convolution2d_nchw_f32( | |
| xnn_operator_t convolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_convolution2d_nchw_f32( | |
| xnn_operator_t convolution_op, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_convolution2d_nhwc_f16( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_channel_stride, | |
| size_t output_channel_stride, | |
| const void* kernel, | |
| const void* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* convolution_op_out); | |
| enum xnn_status xnn_reshape_convolution2d_nhwc_f16( | |
| xnn_operator_t convolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_convolution2d_nhwc_f16( | |
| xnn_operator_t convolution_op, | |
| void* workspace, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_convolution2d_nhwc_f32( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_channel_stride, | |
| size_t output_channel_stride, | |
| const float* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* convolution_op_out); | |
| enum xnn_status xnn_create_convolution2d_nhwc_f32_f16( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_channel_stride, | |
| size_t output_channel_stride, | |
| const void* kernel, | |
| const void* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* convolution_op_out); | |
| // Forward declare. | |
| struct xnn_post_operation; | |
| /// Deprecated | |
| enum xnn_status xnn_create_fused_convolution2d_nhwc_f32( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_channel_stride, | |
| size_t output_channel_stride, | |
| const float* kernel, | |
| const float* bias, | |
| size_t num_post_operations, | |
| struct xnn_post_operation* post_operations, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* convolution_op_out); | |
| enum xnn_status xnn_reshape_convolution2d_nhwc_f32( | |
| xnn_operator_t convolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_convolution2d_nhwc_f32( | |
| xnn_operator_t convolution_op, | |
| void* workspace, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_convolution2d_nhwc_qd8_f16_qc8w( | |
| uint32_t input_padding_top, uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, uint32_t input_padding_left, | |
| uint32_t kernel_height, uint32_t kernel_width, uint32_t subsampling_height, | |
| uint32_t subsampling_width, uint32_t dilation_height, | |
| uint32_t dilation_width, uint32_t groups, size_t group_input_channels, | |
| size_t group_output_channels, size_t input_channel_stride, | |
| size_t output_channel_stride, const float* kernel_scale, | |
| const int8_t* kernel, const float* bias, float output_min, float output_max, | |
| uint32_t flags, xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, xnn_operator_t* convolution_op_out); | |
| enum xnn_status xnn_create_convolution2d_nhwc_qd8_f32_qc8w( | |
| uint32_t input_padding_top, uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, uint32_t input_padding_left, | |
| uint32_t kernel_height, uint32_t kernel_width, uint32_t subsampling_height, | |
| uint32_t subsampling_width, uint32_t dilation_height, | |
| uint32_t dilation_width, uint32_t groups, size_t group_input_channels, | |
| size_t group_output_channels, size_t input_channel_stride, | |
| size_t output_channel_stride, const float* kernel_scale, | |
| const int8_t* kernel, const float* bias, float output_min, float output_max, | |
| uint32_t flags, xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, xnn_operator_t* convolution_op_out); | |
| enum xnn_status xnn_create_convolution2d_nhwc_qs8( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_channel_stride, | |
| size_t output_channel_stride, | |
| int8_t input_zero_point, | |
| float input_scale, | |
| float kernel_scale, | |
| const int8_t* kernel, | |
| const int32_t* bias, | |
| int8_t output_zero_point, | |
| float output_scale, | |
| int8_t output_min, | |
| int8_t output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* convolution_op_out); | |
| enum xnn_status xnn_reshape_convolution2d_nhwc_qd8_f16_qc8w( | |
| xnn_operator_t convolution_op, size_t batch_size, size_t input_height, | |
| size_t input_width, size_t* workspace_size, size_t* workspace_alignment, | |
| size_t* output_height_out, size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_reshape_convolution2d_nhwc_qd8_f32_qc8w( | |
| xnn_operator_t convolution_op, size_t batch_size, size_t input_height, | |
| size_t input_width, size_t* workspace_size, size_t* workspace_alignment, | |
| size_t* output_height_out, size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_reshape_convolution2d_nhwc_qs8( | |
| xnn_operator_t convolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_convolution2d_nhwc_qd8_f16_qc8w( | |
| xnn_operator_t convolution_op, void* workspace, const int8_t* input, | |
| void* output, | |
| const struct xnn_quantization_params* quantization_params); | |
| enum xnn_status xnn_setup_convolution2d_nhwc_qd8_f32_qc8w( | |
| xnn_operator_t convolution_op, void* workspace, const int8_t* input, | |
| float* output, | |
| const struct xnn_quantization_params* quantization_params); | |
| enum xnn_status xnn_setup_convolution2d_nhwc_qs8( | |
| xnn_operator_t convolution_op, | |
| void* workspace, | |
| const int8_t* input, | |
| int8_t* output); | |
| enum xnn_status xnn_create_convolution2d_nhwc_qs8_qc8w( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_channel_stride, | |
| size_t output_channel_stride, | |
| int8_t input_zero_point, | |
| float input_scale, | |
| const float* kernel_scale, | |
| const int8_t* kernel, | |
| const int32_t* bias, | |
| int8_t output_zero_point, | |
| float output_scale, | |
| int8_t output_min, | |
| int8_t output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* convolution_op_out); | |
| enum xnn_status xnn_reshape_convolution2d_nhwc_qs8_qc8w( | |
| xnn_operator_t convolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_convolution2d_nhwc_qs8_qc8w( | |
| xnn_operator_t convolution_op, | |
| void* workspace, | |
| const int8_t* input, | |
| int8_t* output); | |
| enum xnn_status xnn_create_convolution2d_nhwc_qu8( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t subsampling_height, | |
| uint32_t subsampling_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_channel_stride, | |
| size_t output_channel_stride, | |
| uint8_t input_zero_point, | |
| float input_scale, | |
| uint8_t kernel_zero_point, | |
| float kernel_scale, | |
| const uint8_t* kernel, | |
| const int32_t* bias, | |
| uint8_t output_zero_point, | |
| float output_scale, | |
| uint8_t output_min, | |
| uint8_t output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* convolution_op_out); | |
| enum xnn_status xnn_reshape_convolution2d_nhwc_qu8( | |
| xnn_operator_t convolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_convolution2d_nhwc_qu8( | |
| xnn_operator_t convolution_op, | |
| void* workspace, | |
| const uint8_t* input, | |
| uint8_t* output); | |
| enum xnn_status xnn_create_copy_nc_x8( | |
| uint32_t flags, | |
| xnn_operator_t* copy_op_out); | |
| enum xnn_status xnn_reshape_copy_nc_x8( | |
| xnn_operator_t copy_op, | |
| size_t batch_size, | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_copy_nc_x8( | |
| xnn_operator_t copy_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_copy_nc_x16( | |
| uint32_t flags, | |
| xnn_operator_t* copy_op_out); | |
| enum xnn_status xnn_reshape_copy_nc_x16( | |
| xnn_operator_t copy_op, | |
| size_t batch_size, | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_copy_nc_x16( | |
| xnn_operator_t copy_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_copy_nc_x32( | |
| uint32_t flags, | |
| xnn_operator_t* copy_op_out); | |
| enum xnn_status xnn_reshape_copy_nc_x32( | |
| xnn_operator_t copy_op, | |
| size_t batch_size, | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_copy_nc_x32( | |
| xnn_operator_t copy_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_run_copy_nc_x32( | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| size_t batch_size, | |
| const uint32_t* input, | |
| uint32_t* output, | |
| uint32_t flags, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_deconvolution2d_nhwc_f16( | |
| uint32_t output_padding_top, | |
| uint32_t output_padding_right, | |
| uint32_t output_padding_bottom, | |
| uint32_t output_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| const void* kernel, | |
| const void* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* deconvolution_op_out); | |
| enum xnn_status xnn_reshape_deconvolution2d_nhwc_f16( | |
| xnn_operator_t deconvolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| uint32_t adjustment_height, | |
| uint32_t adjustment_width, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_deconvolution2d_nhwc_f16( | |
| xnn_operator_t deconvolution_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_deconvolution2d_nhwc_f32( | |
| uint32_t output_padding_top, | |
| uint32_t output_padding_right, | |
| uint32_t output_padding_bottom, | |
| uint32_t output_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| const float* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* deconvolution_op_out); | |
| enum xnn_status xnn_create_deconvolution2d_nhwc_f32_f16( | |
| uint32_t output_padding_top, | |
| uint32_t output_padding_right, | |
| uint32_t output_padding_bottom, | |
| uint32_t output_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| const void* kernel, | |
| const void* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* deconvolution_op_out); | |
| enum xnn_status xnn_reshape_deconvolution2d_nhwc_f32( | |
| xnn_operator_t deconvolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| uint32_t adjustment_height, | |
| uint32_t adjustment_width, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_deconvolution2d_nhwc_f32( | |
| xnn_operator_t deconvolution_op, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_deconvolution2d_nhwc_qd8_f32_qc8w( | |
| uint32_t output_padding_top, | |
| uint32_t output_padding_right, | |
| uint32_t output_padding_bottom, | |
| uint32_t output_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| const float* kernel_scale, | |
| const int8_t* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* deconvolution_op_out); | |
| enum xnn_status xnn_reshape_deconvolution2d_nhwc_qd8_f32_qc8w( | |
| xnn_operator_t deconvolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| uint32_t adjustment_height, | |
| uint32_t adjustment_width, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_deconvolution2d_nhwc_qd8_f32_qc8w( | |
| xnn_operator_t deconvolution_op, | |
| const int8_t* input, | |
| float* output, | |
| const struct xnn_quantization_params* quantization_params); | |
| enum xnn_status xnn_create_deconvolution2d_nhwc_qs8( | |
| uint32_t output_padding_top, | |
| uint32_t output_padding_right, | |
| uint32_t output_padding_bottom, | |
| uint32_t output_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| int8_t input_zero_point, | |
| float input_scale, | |
| float kernel_scale, | |
| const int8_t* kernel, | |
| const int32_t* bias, | |
| int8_t output_zero_point, | |
| float output_scale, | |
| int8_t output_min, | |
| int8_t output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* deconvolution_op_out); | |
| enum xnn_status xnn_reshape_deconvolution2d_nhwc_qs8( | |
| xnn_operator_t deconvolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| uint32_t adjustment_height, | |
| uint32_t adjustment_width, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_deconvolution2d_nhwc_qs8( | |
| xnn_operator_t deconvolution_op, | |
| const int8_t* input, | |
| int8_t* output); | |
| enum xnn_status xnn_create_deconvolution2d_nhwc_qs8_qc8w( | |
| uint32_t output_padding_top, | |
| uint32_t output_padding_right, | |
| uint32_t output_padding_bottom, | |
| uint32_t output_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| int8_t input_zero_point, | |
| float input_scale, | |
| const float* kernel_scale, | |
| const int8_t* kernel, | |
| const int32_t* bias, | |
| int8_t output_zero_point, | |
| float output_scale, | |
| int8_t output_min, | |
| int8_t output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* deconvolution_op_out); | |
| enum xnn_status xnn_reshape_deconvolution2d_nhwc_qs8_qc8w( | |
| xnn_operator_t deconvolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| uint32_t adjustment_height, | |
| uint32_t adjustment_width, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_deconvolution2d_nhwc_qs8_qc8w( | |
| xnn_operator_t deconvolution_op, | |
| const int8_t* input, | |
| int8_t* output); | |
| enum xnn_status xnn_create_deconvolution2d_nhwc_qu8( | |
| uint32_t output_padding_top, | |
| uint32_t output_padding_right, | |
| uint32_t output_padding_bottom, | |
| uint32_t output_padding_left, | |
| uint32_t kernel_height, | |
| uint32_t kernel_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint32_t groups, | |
| size_t group_input_channels, | |
| size_t group_output_channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| uint8_t input_zero_point, | |
| float input_scale, | |
| uint8_t kernel_zero_point, | |
| float kernel_scale, | |
| const uint8_t* kernel, | |
| const int32_t* bias, | |
| uint8_t output_zero_point, | |
| float output_scale, | |
| uint8_t output_min, | |
| uint8_t output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* deconvolution_op_out); | |
| enum xnn_status xnn_reshape_deconvolution2d_nhwc_qu8( | |
| xnn_operator_t deconvolution_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| uint32_t adjustment_height, | |
| uint32_t adjustment_width, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_deconvolution2d_nhwc_qu8( | |
| xnn_operator_t deconvolution_op, | |
| const uint8_t* input, | |
| uint8_t* output); | |
| enum xnn_status xnn_create_depth_to_space_nchw2nhwc_x16( | |
| uint32_t block_size, | |
| uint32_t flags, | |
| xnn_operator_t* depth_to_space_op_out); | |
| enum xnn_status xnn_reshape_depth_to_space_nchw2nhwc_x16( | |
| xnn_operator_t depth_to_space_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t input_channels, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| size_t* output_channels_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_depth_to_space_nchw2nhwc_x16( | |
| xnn_operator_t depth_to_space_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_depth_to_space_nchw2nhwc_x32( | |
| uint32_t block_size, | |
| uint32_t flags, | |
| xnn_operator_t* depth_to_space_op_out); | |
| enum xnn_status xnn_reshape_depth_to_space_nchw2nhwc_x32( | |
| xnn_operator_t depth_to_space_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t input_channels, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| size_t* output_channels_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_depth_to_space_nchw2nhwc_x32( | |
| xnn_operator_t depth_to_space_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_depth_to_space_nhwc_x8( | |
| uint32_t block_size, | |
| uint32_t flags, | |
| xnn_operator_t* depth_to_space_op_out); | |
| enum xnn_status xnn_reshape_depth_to_space_nhwc_x8( | |
| xnn_operator_t depth_to_space_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t input_channels, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| size_t* output_channels_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_depth_to_space_nhwc_x8( | |
| xnn_operator_t depth_to_space_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_depth_to_space_nhwc_x16( | |
| uint32_t block_size, | |
| uint32_t flags, | |
| xnn_operator_t* depth_to_space_op_out); | |
| enum xnn_status xnn_reshape_depth_to_space_nhwc_x16( | |
| xnn_operator_t depth_to_space_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t input_channels, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| size_t* output_channels_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_depth_to_space_nhwc_x16( | |
| xnn_operator_t depth_to_space_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_depth_to_space_nhwc_x32( | |
| uint32_t block_size, | |
| uint32_t flags, | |
| xnn_operator_t* depth_to_space_op_out); | |
| enum xnn_status xnn_reshape_depth_to_space_nhwc_x32( | |
| xnn_operator_t depth_to_space_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t input_channels, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| size_t* output_channels_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_depth_to_space_nhwc_x32( | |
| xnn_operator_t depth_to_space_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_dynamic_fully_connected_nc_f16( | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_operator_t* dynamic_fully_connected_op_out); | |
| enum xnn_status xnn_reshape_dynamic_fully_connected_nc_f16( | |
| xnn_operator_t dynamic_fully_connected_op, | |
| size_t batch_size, | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_dynamic_fully_connected_nc_f16( | |
| xnn_operator_t dynamic_fully_connected_op, | |
| void* workspace, | |
| const void* input, | |
| const void* kernel, | |
| const void* bias, | |
| void* output); | |
| enum xnn_status xnn_create_dynamic_fully_connected_nc_f32( | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_operator_t* dynamic_fully_connected_op_out); | |
| enum xnn_status xnn_reshape_dynamic_fully_connected_nc_f32( | |
| xnn_operator_t dynamic_fully_connected_op, | |
| size_t batch_size, | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_dynamic_fully_connected_nc_f32( | |
| xnn_operator_t dynamic_fully_connected_op, | |
| void* workspace, | |
| const float* input, | |
| const float* kernel, | |
| const float* bias, | |
| float* output); | |
| enum xnn_status xnn_create_fully_connected_nc_f16( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| const void* kernel, | |
| const void* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_reshape_fully_connected_nc_f16( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_fully_connected_nc_f16( | |
| xnn_operator_t fully_connected_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_fully_connected_nc_f32_f16( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| const void* kernel, | |
| const void* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_create_fully_connected_nc_f32( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| const float* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_reshape_fully_connected_nc_f32_f16( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_reshape_fully_connected_nc_f32( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_fully_connected_nc_f32_f16( | |
| xnn_operator_t fully_connected_op, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_setup_fully_connected_nc_f32( | |
| xnn_operator_t fully_connected_op, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_fully_connected_nc_f32_qc4w( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| uint8_t kernel_zero_point, | |
| const float* kernel_scale, | |
| const uint8_t* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_reshape_fully_connected_nc_f32_qc4w( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_fully_connected_nc_f32_qc4w( | |
| xnn_operator_t fully_connected_op, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_fully_connected_nc_f32_qc8w( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| const float* kernel_scale, | |
| const int8_t* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_reshape_fully_connected_nc_f32_qc8w( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_fully_connected_nc_f32_qc8w( | |
| xnn_operator_t fully_connected_op, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_fully_connected_nc_qd8_f16_qc4w( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| uint8_t kernel_zero_point, | |
| const float* kernel_scale, | |
| const void* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_setup_fully_connected_nc_qd8_f16_qc4w( | |
| xnn_operator_t fully_connected_op, | |
| const int8_t* input, | |
| void* output, | |
| const struct xnn_quantization_params* quantization_params); | |
| enum xnn_status xnn_reshape_fully_connected_nc_qd8_f16_qc4w( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_fully_connected_nc_qd8_f16_qb4w( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| size_t block_size, | |
| uint8_t kernel_zero_point, | |
| const uint16_t* kernel_scale, | |
| const void* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_reshape_fully_connected_nc_qd8_f16_qb4w( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_fully_connected_nc_qd8_f16_qb4w( | |
| xnn_operator_t fully_connected_op, | |
| const int8_t* input, | |
| void* output, | |
| const struct xnn_quantization_params* quantization_params); | |
| enum xnn_status xnn_create_fully_connected_nc_qd8_f32_qc4w( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| uint8_t kernel_zero_point, | |
| const float* kernel_scale, | |
| const void* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_setup_fully_connected_nc_qd8_f32_qc4w( | |
| xnn_operator_t fully_connected_op, | |
| const int8_t* input, | |
| float* output, | |
| const struct xnn_quantization_params* quantization_params); | |
| enum xnn_status xnn_reshape_fully_connected_nc_qd8_f32_qc4w( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_fully_connected_nc_qd8_f32_qb4w( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| size_t block_size, | |
| uint8_t kernel_zero_point, | |
| const uint16_t* kernel_scale, | |
| const void* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_reshape_fully_connected_nc_qd8_f32_qb4w( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_fully_connected_nc_qd8_f32_qb4w( | |
| xnn_operator_t fully_connected_op, | |
| const int8_t* input, | |
| float* output, | |
| const struct xnn_quantization_params* quantization_params); | |
| enum xnn_status xnn_create_fully_connected_nc_qd8_f16_qc8w( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| const float* kernel_scale, | |
| const int8_t* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_setup_fully_connected_nc_qd8_f16_qc8w( | |
| xnn_operator_t fully_connected_op, | |
| const int8_t* input, | |
| void* output, | |
| const struct xnn_quantization_params* quantization_params); | |
| enum xnn_status xnn_reshape_fully_connected_nc_qd8_f16_qc8w( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_fully_connected_nc_qd8_f32_qc8w( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| const float* kernel_scale, | |
| const int8_t* kernel, | |
| const float* bias, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_setup_fully_connected_nc_qd8_f32_qc8w( | |
| xnn_operator_t fully_connected_op, | |
| const int8_t* input, | |
| float* output, | |
| const struct xnn_quantization_params* quantization_params); | |
| enum xnn_status xnn_reshape_fully_connected_nc_qd8_f32_qc8w( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_fully_connected_nc_qs8( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| int8_t input_zero_point, | |
| float input_scale, | |
| float kernel_scale, | |
| const int8_t* kernel, | |
| const int32_t* bias, | |
| int8_t output_zero_point, | |
| float output_scale, | |
| int8_t output_min, | |
| int8_t output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_reshape_fully_connected_nc_qs8( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_fully_connected_nc_qs8( | |
| xnn_operator_t fully_connected_op, | |
| const int8_t* input, | |
| int8_t* output); | |
| enum xnn_status xnn_create_fully_connected_nc_qs8_qc8w( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| int8_t input_zero_point, | |
| float input_scale, | |
| const float* kernel_scale, | |
| const int8_t* kernel, | |
| const int32_t* bias, | |
| int8_t output_zero_point, | |
| float output_scale, | |
| int8_t output_min, | |
| int8_t output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_reshape_fully_connected_nc_qs8_qc8w( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_fully_connected_nc_qs8_qc8w( | |
| xnn_operator_t fully_connected_op, | |
| const int8_t* input, | |
| int8_t* output); | |
| enum xnn_status xnn_create_fully_connected_nc_qu8( | |
| size_t input_channels, | |
| size_t output_channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| uint8_t input_zero_point, | |
| float input_scale, | |
| uint8_t kernel_zero_point, | |
| float kernel_scale, | |
| const uint8_t* kernel, | |
| const int32_t* bias, | |
| uint8_t output_zero_point, | |
| float output_scale, | |
| uint8_t output_min, | |
| uint8_t output_max, | |
| uint32_t flags, | |
| xnn_code_cache_t code_cache, | |
| xnn_weights_cache_t weights_cache, | |
| xnn_operator_t* fully_connected_op_out); | |
| enum xnn_status xnn_reshape_fully_connected_nc_qu8( | |
| xnn_operator_t fully_connected_op, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_fully_connected_nc_qu8( | |
| xnn_operator_t fully_connected_op, | |
| const uint8_t* input, | |
| uint8_t* output); | |
| enum xnn_status xnn_create_max_pooling2d_nhwc_f16( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_operator_t* max_pooling_op_out); | |
| enum xnn_status xnn_reshape_max_pooling2d_nhwc_f16( | |
| xnn_operator_t max_pooling_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_max_pooling2d_nhwc_f16( | |
| xnn_operator_t max_pooling_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_max_pooling2d_nhwc_f32( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| float output_min, | |
| float output_max, | |
| uint32_t flags, | |
| xnn_operator_t* max_pooling_op_out); | |
| enum xnn_status xnn_reshape_max_pooling2d_nhwc_f32( | |
| xnn_operator_t max_pooling_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_max_pooling2d_nhwc_f32( | |
| xnn_operator_t max_pooling_op, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_max_pooling2d_nhwc_s8( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| int8_t output_min, | |
| int8_t output_max, | |
| uint32_t flags, | |
| xnn_operator_t* max_pooling_op_out); | |
| enum xnn_status xnn_reshape_max_pooling2d_nhwc_s8( | |
| xnn_operator_t max_pooling_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_max_pooling2d_nhwc_s8( | |
| xnn_operator_t max_pooling_op, | |
| const int8_t* input, | |
| int8_t* output); | |
| enum xnn_status xnn_create_max_pooling2d_nhwc_u8( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| uint32_t stride_height, | |
| uint32_t stride_width, | |
| uint32_t dilation_height, | |
| uint32_t dilation_width, | |
| uint8_t output_min, | |
| uint8_t output_max, | |
| uint32_t flags, | |
| xnn_operator_t* max_pooling_op_out); | |
| enum xnn_status xnn_reshape_max_pooling2d_nhwc_u8( | |
| xnn_operator_t max_pooling_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_max_pooling2d_nhwc_u8( | |
| xnn_operator_t max_pooling_op, | |
| const uint8_t* input, | |
| uint8_t* output); | |
| enum xnn_status xnn_create_reduce_nd( | |
| enum xnn_reduce_operator reduce_operator_type, | |
| enum xnn_datatype datatype, | |
| const struct xnn_quantization_params* input_quantization, | |
| const struct xnn_quantization_params* output_quantization, | |
| uint32_t flags, | |
| xnn_operator_t* reduce_op_out); | |
| enum xnn_status xnn_reshape_reduce_nd( // | |
| xnn_operator_t reduce_op, // | |
| size_t num_reduction_axes, // | |
| const int64_t* reduction_axes, // | |
| size_t num_input_dims, // | |
| const size_t* input_shape, // | |
| size_t* workspace_size, // | |
| size_t* workspace_alignment, // | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_reduce_nd( | |
| xnn_operator_t reduce_op, | |
| void* workspace, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_resize_bilinear2d_nchw_f32( | |
| size_t output_height, | |
| size_t output_width, | |
| uint32_t flags, | |
| xnn_operator_t* resize_op_out); | |
| enum xnn_status xnn_reshape_resize_bilinear2d_nchw_f32( | |
| xnn_operator_t resize_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_resize_bilinear2d_nchw_f32( | |
| xnn_operator_t resize_op, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_resize_bilinear2d_nchw_f16( | |
| size_t output_height, | |
| size_t output_width, | |
| uint32_t flags, | |
| xnn_operator_t* resize_op_out); | |
| enum xnn_status xnn_reshape_resize_bilinear2d_nchw_f16( | |
| xnn_operator_t resize_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_resize_bilinear2d_nchw_f16( | |
| xnn_operator_t resize_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_resize_bilinear2d_nhwc_f16( | |
| size_t output_height, | |
| size_t output_width, | |
| uint32_t flags, | |
| xnn_operator_t* resize_op_out); | |
| enum xnn_status xnn_reshape_resize_bilinear2d_nhwc_f16( | |
| xnn_operator_t resize_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_resize_bilinear2d_nhwc_f16( | |
| xnn_operator_t resize_op, | |
| void* workspace, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_resize_bilinear2d_nhwc_f32( | |
| size_t output_height, | |
| size_t output_width, | |
| uint32_t flags, | |
| xnn_operator_t* resize_op_out); | |
| enum xnn_status xnn_reshape_resize_bilinear2d_nhwc_f32( | |
| xnn_operator_t resize_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_resize_bilinear2d_nhwc_f32( | |
| xnn_operator_t resize_op, | |
| void* workspace, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_resize_bilinear2d_nhwc_s8( | |
| size_t output_height, | |
| size_t output_width, | |
| uint32_t flags, | |
| xnn_operator_t* resize_op_out); | |
| enum xnn_status xnn_reshape_resize_bilinear2d_nhwc_s8( | |
| xnn_operator_t resize_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* workspace_size, | |
| size_t* workspace, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_resize_bilinear2d_nhwc_s8( | |
| xnn_operator_t resize_op, | |
| void* workspace, | |
| const int8_t* input, | |
| int8_t* output); | |
| enum xnn_status xnn_create_resize_bilinear2d_nhwc_u8( | |
| size_t output_height, | |
| size_t output_width, | |
| uint32_t flags, | |
| xnn_operator_t* resize_op_out); | |
| enum xnn_status xnn_reshape_resize_bilinear2d_nhwc_u8( | |
| xnn_operator_t resize_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_resize_bilinear2d_nhwc_u8( | |
| xnn_operator_t resize_op, | |
| void* workspace, | |
| const uint8_t* input, | |
| uint8_t* output); | |
| enum xnn_status xnn_create_rope_nthc_f16( | |
| uint32_t flags, | |
| xnn_operator_t* rope_op_out); | |
| enum xnn_status xnn_reshape_rope_nthc_f16( | |
| xnn_operator_t rope_op, | |
| size_t batch_size, | |
| size_t tokens, | |
| size_t heads, | |
| size_t channels, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_rope_nthc_f16( | |
| xnn_operator_t rope_op, | |
| const void* input, | |
| const void* weights, | |
| void* output); | |
| enum xnn_status xnn_create_rope_nthc_f32( | |
| uint32_t flags, | |
| xnn_operator_t* rope_op_out); | |
| enum xnn_status xnn_reshape_rope_nthc_f32( | |
| xnn_operator_t rope_op, | |
| size_t batch_size, | |
| size_t tokens, | |
| size_t heads, | |
| size_t channels, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_rope_nthc_f32( | |
| xnn_operator_t rope_op, | |
| const float* input, | |
| const float* weights, | |
| float* output); | |
| // N: batch size | |
| // H: number of heads | |
| // T: tokens (sequence length) | |
| // C: channels (head dimension) | |
| enum xnn_status xnn_create_scaled_dot_product_attention_nhtc_f16( | |
| enum xnn_attention_logits_cap_type cap_type, | |
| const void* cap_params, | |
| uint32_t flags, | |
| xnn_operator_t* attention_op_out); | |
| enum xnn_status xnn_reshape_scaled_dot_product_attention_nhtc_f16( | |
| xnn_operator_t attention_op, | |
| size_t batch_size, | |
| size_t query_heads, | |
| // Number of tokens in query. | |
| size_t query_tokens, | |
| size_t key_value_heads, | |
| // Number of tokens in key/value. For self-attention, this is same as tokens. | |
| size_t key_value_tokens, | |
| size_t query_key_channels, | |
| size_t value_channels, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| pthreadpool_t threadpool); | |
| // Query is of dimension [batch_size, query_heads, query_tokens, channels]. | |
| // Key and value are of dimension [batch_size, key_value_heads, key_value_tokens, channels]. | |
| // Scale is of dimension [channels]. | |
| // Mask is of dimension [query_tokens, key_value_tokens]. | |
| enum xnn_status xnn_setup_scaled_dot_product_attention_nhtc_f16( | |
| xnn_operator_t attention_op, | |
| void* workspace, | |
| const void* query, | |
| const void* key, | |
| const void* value, | |
| const void* scale, | |
| const void* mask, | |
| void* output); | |
| // N: batch size | |
| // H: number of heads | |
| // T: tokens (sequence length) | |
| // C: channels (head dimension) | |
| enum xnn_status xnn_create_scaled_dot_product_attention_nhtc_f32( | |
| enum xnn_attention_logits_cap_type cap_type, | |
| const void* cap_params, | |
| uint32_t flags, | |
| xnn_operator_t* attention_op_out); | |
| enum xnn_status xnn_reshape_scaled_dot_product_attention_nhtc_f32( | |
| xnn_operator_t attention_op, | |
| size_t batch_size, | |
| size_t query_heads, | |
| // Number of tokens in query. | |
| size_t query_tokens, | |
| size_t key_value_heads, | |
| // Number of tokens in key/value. For self-attention, this is same as tokens. | |
| size_t key_value_tokens, | |
| size_t query_key_channels, | |
| size_t value_channels, | |
| size_t* workspace_size, | |
| size_t* workspace_alignment, | |
| pthreadpool_t threadpool); | |
| // Query is of dimension [batch_size, query_heads, query_tokens, query_key_channels]. | |
| // Key and value are of dimension [batch_size, key_value_heads, key_value_tokens, query_key_channels]. | |
| // Scale is of dimension [query_key_channels]. | |
| // Mask is of dimension [query_tokens, key_value_tokens]. | |
| // Output is of dimension [batch_size, query_heads, query_tokens, value_channels]. | |
| enum xnn_status xnn_setup_scaled_dot_product_attention_nhtc_f32( | |
| xnn_operator_t attention_op, | |
| void* workspace, | |
| const float* query, | |
| const float* key, | |
| const float* value, | |
| const float* scale, | |
| const float* mask, | |
| float* output); | |
| enum xnn_status xnn_create_slice_nd_x16( | |
| uint32_t flags, | |
| xnn_operator_t* slice_op_out); | |
| enum xnn_status xnn_reshape_slice_nd_x16( | |
| xnn_operator_t slice_op, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* offsets, | |
| const size_t* sizes, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_slice_nd_x16( | |
| xnn_operator_t slice_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_slice_nd_x32( | |
| uint32_t flags, | |
| xnn_operator_t* slice_op_out); | |
| enum xnn_status xnn_reshape_slice_nd_x32( | |
| xnn_operator_t slice_op, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* offsets, | |
| const size_t* sizes, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_slice_nd_x32( | |
| xnn_operator_t slice_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_run_slice_nd_x32( | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* offsets, | |
| const size_t* sizes, | |
| const void* input, | |
| void* output, | |
| uint32_t flags, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_softmax_nc_f16( | |
| uint32_t flags, | |
| xnn_operator_t* softmax_op_out); | |
| enum xnn_status xnn_reshape_softmax_nc_f16( | |
| xnn_operator_t softmax_op, | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_softmax_nc_f16( | |
| xnn_operator_t softmax_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_softmax_nc_f32( | |
| uint32_t flags, | |
| xnn_operator_t* softmax_op_out); | |
| enum xnn_status xnn_reshape_softmax_nc_f32( | |
| xnn_operator_t softmax_op, | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_softmax_nc_f32( | |
| xnn_operator_t softmax_op, | |
| const float* input, | |
| float* output); | |
| enum xnn_status xnn_create_softmax_nc_qu8( | |
| float input_scale, | |
| uint8_t output_zero_point, | |
| float output_scale, | |
| uint32_t flags, | |
| xnn_operator_t* softmax_op_out); | |
| enum xnn_status xnn_reshape_softmax_nc_qu8( | |
| xnn_operator_t softmax_op, | |
| size_t channels, | |
| size_t input_stride, | |
| size_t output_stride, | |
| size_t batch_size, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_softmax_nc_qu8( | |
| xnn_operator_t softmax_op, | |
| const uint8_t* input, | |
| uint8_t* output); | |
| enum xnn_status xnn_create_space_to_depth_nhwc_x16( | |
| uint32_t block_size, | |
| uint32_t flags, | |
| xnn_operator_t* space_to_depth_op_out); | |
| enum xnn_status xnn_reshape_space_to_depth_nhwc_x16( | |
| xnn_operator_t space_to_depth_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t input_channels, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| size_t* output_channels_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_space_to_depth_nhwc_x16( | |
| xnn_operator_t space_to_depth_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_space_to_depth_nhwc_x32( | |
| uint32_t block_size, | |
| uint32_t flags, | |
| xnn_operator_t* space_to_depth_op_out); | |
| enum xnn_status xnn_reshape_space_to_depth_nhwc_x32( | |
| xnn_operator_t space_to_depth_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t input_channels, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| size_t* output_channels_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_space_to_depth_nhwc_x32( | |
| xnn_operator_t space_to_depth_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_transpose_nd_x8( | |
| uint32_t flags, | |
| xnn_operator_t* transpose_op_out); | |
| enum xnn_status xnn_reshape_transpose_nd_x8( | |
| xnn_operator_t transpose_op, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* output_perm, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_transpose_nd_x8( | |
| xnn_operator_t transpose_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_run_transpose_nd_x8( | |
| const void* input, | |
| void* output, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* output_perm, | |
| uint32_t flags, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_transpose_nd_x16( | |
| uint32_t flags, | |
| xnn_operator_t* transpose_op_out); | |
| enum xnn_status xnn_reshape_transpose_nd_x16( | |
| xnn_operator_t transpose_op, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* output_perm, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_transpose_nd_x16( | |
| xnn_operator_t transpose_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_run_transpose_nd_x16( | |
| const void* input, | |
| void* output, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* output_perm, | |
| uint32_t flags, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_transpose_nd_x32( | |
| uint32_t flags, | |
| xnn_operator_t* transpose_op_out); | |
| enum xnn_status xnn_reshape_transpose_nd_x32( | |
| xnn_operator_t transpose_op, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* output_perm, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_transpose_nd_x32( | |
| xnn_operator_t transpose_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_run_transpose_nd_x32( | |
| const void* input, | |
| void* output, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* output_perm, | |
| uint32_t flags, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_transpose_nd_x64( | |
| uint32_t flags, | |
| xnn_operator_t* transpose_op_out); | |
| enum xnn_status xnn_reshape_transpose_nd_x64( | |
| xnn_operator_t transpose_op, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* output_perm, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_transpose_nd_x64( | |
| xnn_operator_t transpose_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_run_transpose_nd_x64( | |
| const void* input, | |
| void* output, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* output_perm, | |
| uint32_t flags, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_create_unpooling2d_nhwc_x32( | |
| uint32_t input_padding_top, | |
| uint32_t input_padding_right, | |
| uint32_t input_padding_bottom, | |
| uint32_t input_padding_left, | |
| uint32_t pooling_height, | |
| uint32_t pooling_width, | |
| size_t channels, | |
| size_t input_pixel_stride, | |
| size_t output_pixel_stride, | |
| uint32_t flags, | |
| xnn_operator_t* unpooling_op_out); | |
| enum xnn_status xnn_reshape_unpooling2d_nhwc_x32( | |
| xnn_operator_t unpooling_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_unpooling2d_nhwc_x32( | |
| xnn_operator_t unpooling_op, | |
| const void* input, | |
| const uint32_t* index, | |
| void* output); | |
| enum xnn_status xnn_create_slice_nd_x8( | |
| uint32_t flags, | |
| xnn_operator_t* slice_op_out); | |
| enum xnn_status xnn_reshape_slice_nd_x8( | |
| xnn_operator_t slice_op, | |
| size_t num_dims, | |
| const size_t* input_shape, | |
| const size_t* offsets, | |
| const size_t* sizes, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_slice_nd_x8( | |
| xnn_operator_t slice_op, | |
| const void* input, | |
| void* output); | |
| enum xnn_status xnn_create_space_to_depth_nhwc_x8( | |
| uint32_t block_size, | |
| uint32_t flags, | |
| xnn_operator_t* space_to_depth_op_out); | |
| enum xnn_status xnn_reshape_space_to_depth_nhwc_x8( | |
| xnn_operator_t space_to_depth_op, | |
| size_t batch_size, | |
| size_t input_height, | |
| size_t input_width, | |
| size_t input_channels, | |
| size_t* output_height_out, | |
| size_t* output_width_out, | |
| size_t* output_channels_out, | |
| pthreadpool_t threadpool); | |
| enum xnn_status xnn_setup_space_to_depth_nhwc_x8( | |
| xnn_operator_t space_to_depth_op, | |
| const void* input, | |
| void* output); | |
| } // extern "C" | |