|
|
|
|
|
|
|
|
|
|
|
|
|
|
#ifndef OPENCV_HAL_INTRIN_WASM_HPP |
|
|
#define OPENCV_HAL_INTRIN_WASM_HPP |
|
|
|
|
|
#include <limits> |
|
|
#include <cstring> |
|
|
#include <algorithm> |
|
|
#include "opencv2/core/saturate.hpp" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#ifndef __EMSCRIPTEN_major__ |
|
|
#include <emscripten/version.h> |
|
|
#endif |
|
|
|
|
|
#define CV_SIMD128 1 |
|
|
#define CV_SIMD128_64F 0 |
|
|
#define CV_SIMD128_FP16 0 |
|
|
|
|
|
namespace cv |
|
|
{ |
|
|
|
|
|
|
|
|
|
|
|
CV_CPU_OPTIMIZATION_HAL_NAMESPACE_BEGIN |
|
|
|
|
|
#if (__EMSCRIPTEN_major__ * 1000000 + __EMSCRIPTEN_minor__ * 1000 + __EMSCRIPTEN_tiny__) < (1038046) |
|
|
|
|
|
#define wasm_i32x4_trunc_saturate_f32x4 wasm_trunc_saturate_i32x4_f32x4 |
|
|
#define wasm_u32x4_trunc_saturate_f32x4 wasm_trunc_saturate_u32x4_f32x4 |
|
|
#define wasm_i64x2_trunc_saturate_f64x2 wasm_trunc_saturate_i64x2_f64x2 |
|
|
#define wasm_u64x2_trunc_saturate_f64x2 wasm_trunc_saturate_u64x2_f64x2 |
|
|
#define wasm_f32x4_convert_i32x4 wasm_convert_f32x4_i32x4 |
|
|
#define wasm_f32x4_convert_u32x4 wasm_convert_f32x4_u32x4 |
|
|
#define wasm_f64x2_convert_i64x2 wasm_convert_f64x2_i64x2 |
|
|
#define wasm_f64x2_convert_u64x2 wasm_convert_f64x2_u64x2 |
|
|
#endif |
|
|
|
|
|
|
|
|
|
|
|
struct v_uint8x16 |
|
|
{ |
|
|
typedef uchar lane_type; |
|
|
typedef v128_t vector_type; |
|
|
enum { nlanes = 16 }; |
|
|
|
|
|
v_uint8x16() {} |
|
|
explicit v_uint8x16(v128_t v) : val(v) {} |
|
|
v_uint8x16(uchar v0, uchar v1, uchar v2, uchar v3, uchar v4, uchar v5, uchar v6, uchar v7, |
|
|
uchar v8, uchar v9, uchar v10, uchar v11, uchar v12, uchar v13, uchar v14, uchar v15) |
|
|
{ |
|
|
uchar v[] = {v0, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15}; |
|
|
val = wasm_v128_load(v); |
|
|
} |
|
|
|
|
|
uchar get0() const |
|
|
{ |
|
|
return (uchar)wasm_i8x16_extract_lane(val, 0); |
|
|
} |
|
|
|
|
|
v128_t val; |
|
|
}; |
|
|
|
|
|
struct v_int8x16 |
|
|
{ |
|
|
typedef schar lane_type; |
|
|
typedef v128_t vector_type; |
|
|
enum { nlanes = 16 }; |
|
|
|
|
|
v_int8x16() {} |
|
|
explicit v_int8x16(v128_t v) : val(v) {} |
|
|
v_int8x16(schar v0, schar v1, schar v2, schar v3, schar v4, schar v5, schar v6, schar v7, |
|
|
schar v8, schar v9, schar v10, schar v11, schar v12, schar v13, schar v14, schar v15) |
|
|
{ |
|
|
schar v[] = {v0, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15}; |
|
|
val = wasm_v128_load(v); |
|
|
} |
|
|
|
|
|
schar get0() const |
|
|
{ |
|
|
return wasm_i8x16_extract_lane(val, 0); |
|
|
} |
|
|
|
|
|
v128_t val; |
|
|
}; |
|
|
|
|
|
struct v_uint16x8 |
|
|
{ |
|
|
typedef ushort lane_type; |
|
|
typedef v128_t vector_type; |
|
|
enum { nlanes = 8 }; |
|
|
|
|
|
v_uint16x8() {} |
|
|
explicit v_uint16x8(v128_t v) : val(v) {} |
|
|
v_uint16x8(ushort v0, ushort v1, ushort v2, ushort v3, ushort v4, ushort v5, ushort v6, ushort v7) |
|
|
{ |
|
|
ushort v[] = {v0, v1, v2, v3, v4, v5, v6, v7}; |
|
|
val = wasm_v128_load(v); |
|
|
} |
|
|
|
|
|
ushort get0() const |
|
|
{ |
|
|
return (ushort)wasm_i16x8_extract_lane(val, 0); |
|
|
} |
|
|
|
|
|
v128_t val; |
|
|
}; |
|
|
|
|
|
struct v_int16x8 |
|
|
{ |
|
|
typedef short lane_type; |
|
|
typedef v128_t vector_type; |
|
|
enum { nlanes = 8 }; |
|
|
|
|
|
v_int16x8() {} |
|
|
explicit v_int16x8(v128_t v) : val(v) {} |
|
|
v_int16x8(short v0, short v1, short v2, short v3, short v4, short v5, short v6, short v7) |
|
|
{ |
|
|
short v[] = {v0, v1, v2, v3, v4, v5, v6, v7}; |
|
|
val = wasm_v128_load(v); |
|
|
} |
|
|
|
|
|
short get0() const |
|
|
{ |
|
|
return wasm_i16x8_extract_lane(val, 0); |
|
|
} |
|
|
|
|
|
v128_t val; |
|
|
}; |
|
|
|
|
|
struct v_uint32x4 |
|
|
{ |
|
|
typedef unsigned lane_type; |
|
|
typedef v128_t vector_type; |
|
|
enum { nlanes = 4 }; |
|
|
|
|
|
v_uint32x4() {} |
|
|
explicit v_uint32x4(v128_t v) : val(v) {} |
|
|
v_uint32x4(unsigned v0, unsigned v1, unsigned v2, unsigned v3) |
|
|
{ |
|
|
unsigned v[] = {v0, v1, v2, v3}; |
|
|
val = wasm_v128_load(v); |
|
|
} |
|
|
|
|
|
unsigned get0() const |
|
|
{ |
|
|
return (unsigned)wasm_i32x4_extract_lane(val, 0); |
|
|
} |
|
|
|
|
|
v128_t val; |
|
|
}; |
|
|
|
|
|
struct v_int32x4 |
|
|
{ |
|
|
typedef int lane_type; |
|
|
typedef v128_t vector_type; |
|
|
enum { nlanes = 4 }; |
|
|
|
|
|
v_int32x4() {} |
|
|
explicit v_int32x4(v128_t v) : val(v) {} |
|
|
v_int32x4(int v0, int v1, int v2, int v3) |
|
|
{ |
|
|
int v[] = {v0, v1, v2, v3}; |
|
|
val = wasm_v128_load(v); |
|
|
} |
|
|
|
|
|
int get0() const |
|
|
{ |
|
|
return wasm_i32x4_extract_lane(val, 0); |
|
|
} |
|
|
|
|
|
v128_t val; |
|
|
}; |
|
|
|
|
|
struct v_float32x4 |
|
|
{ |
|
|
typedef float lane_type; |
|
|
typedef v128_t vector_type; |
|
|
enum { nlanes = 4 }; |
|
|
|
|
|
v_float32x4() {} |
|
|
explicit v_float32x4(v128_t v) : val(v) {} |
|
|
v_float32x4(float v0, float v1, float v2, float v3) |
|
|
{ |
|
|
float v[] = {v0, v1, v2, v3}; |
|
|
val = wasm_v128_load(v); |
|
|
} |
|
|
|
|
|
float get0() const |
|
|
{ |
|
|
return wasm_f32x4_extract_lane(val, 0); |
|
|
} |
|
|
|
|
|
v128_t val; |
|
|
}; |
|
|
|
|
|
struct v_uint64x2 |
|
|
{ |
|
|
typedef uint64 lane_type; |
|
|
typedef v128_t vector_type; |
|
|
enum { nlanes = 2 }; |
|
|
|
|
|
v_uint64x2() {} |
|
|
explicit v_uint64x2(v128_t v) : val(v) {} |
|
|
v_uint64x2(uint64 v0, uint64 v1) |
|
|
{ |
|
|
uint64 v[] = {v0, v1}; |
|
|
val = wasm_v128_load(v); |
|
|
} |
|
|
|
|
|
uint64 get0() const |
|
|
{ |
|
|
return (uint64)wasm_i64x2_extract_lane(val, 0); |
|
|
} |
|
|
|
|
|
v128_t val; |
|
|
}; |
|
|
|
|
|
struct v_int64x2 |
|
|
{ |
|
|
typedef int64 lane_type; |
|
|
typedef v128_t vector_type; |
|
|
enum { nlanes = 2 }; |
|
|
|
|
|
v_int64x2() {} |
|
|
explicit v_int64x2(v128_t v) : val(v) {} |
|
|
v_int64x2(int64 v0, int64 v1) |
|
|
{ |
|
|
int64 v[] = {v0, v1}; |
|
|
val = wasm_v128_load(v); |
|
|
} |
|
|
|
|
|
int64 get0() const |
|
|
{ |
|
|
return wasm_i64x2_extract_lane(val, 0); |
|
|
} |
|
|
|
|
|
v128_t val; |
|
|
}; |
|
|
|
|
|
struct v_float64x2 |
|
|
{ |
|
|
typedef double lane_type; |
|
|
typedef v128_t vector_type; |
|
|
enum { nlanes = 2 }; |
|
|
|
|
|
v_float64x2() {} |
|
|
explicit v_float64x2(v128_t v) : val(v) {} |
|
|
v_float64x2(double v0, double v1) |
|
|
{ |
|
|
double v[] = {v0, v1}; |
|
|
val = wasm_v128_load(v); |
|
|
} |
|
|
|
|
|
double get0() const |
|
|
{ |
|
|
return wasm_f64x2_extract_lane(val, 0); |
|
|
} |
|
|
|
|
|
v128_t val; |
|
|
}; |
|
|
|
|
|
namespace |
|
|
{ |
|
|
#define OPENCV_HAL_IMPL_REINTERPRET_INT(ft, tt) \ |
|
|
inline tt reinterpret_int(ft x) { union { ft l; tt i; } v; v.l = x; return v.i; } |
|
|
OPENCV_HAL_IMPL_REINTERPRET_INT(uchar, schar) |
|
|
OPENCV_HAL_IMPL_REINTERPRET_INT(schar, schar) |
|
|
OPENCV_HAL_IMPL_REINTERPRET_INT(ushort, short) |
|
|
OPENCV_HAL_IMPL_REINTERPRET_INT(short, short) |
|
|
OPENCV_HAL_IMPL_REINTERPRET_INT(unsigned, int) |
|
|
OPENCV_HAL_IMPL_REINTERPRET_INT(int, int) |
|
|
OPENCV_HAL_IMPL_REINTERPRET_INT(float, int) |
|
|
OPENCV_HAL_IMPL_REINTERPRET_INT(uint64, int64) |
|
|
OPENCV_HAL_IMPL_REINTERPRET_INT(int64, int64) |
|
|
OPENCV_HAL_IMPL_REINTERPRET_INT(double, int64) |
|
|
|
|
|
static const unsigned char popCountTable[] = |
|
|
{ |
|
|
0, 1, 1, 2, 1, 2, 2, 3, 1, 2, 2, 3, 2, 3, 3, 4, |
|
|
1, 2, 2, 3, 2, 3, 3, 4, 2, 3, 3, 4, 3, 4, 4, 5, |
|
|
1, 2, 2, 3, 2, 3, 3, 4, 2, 3, 3, 4, 3, 4, 4, 5, |
|
|
2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, |
|
|
1, 2, 2, 3, 2, 3, 3, 4, 2, 3, 3, 4, 3, 4, 4, 5, |
|
|
2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, |
|
|
2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, |
|
|
3, 4, 4, 5, 4, 5, 5, 6, 4, 5, 5, 6, 5, 6, 6, 7, |
|
|
1, 2, 2, 3, 2, 3, 3, 4, 2, 3, 3, 4, 3, 4, 4, 5, |
|
|
2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, |
|
|
2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, |
|
|
3, 4, 4, 5, 4, 5, 5, 6, 4, 5, 5, 6, 5, 6, 6, 7, |
|
|
2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, |
|
|
3, 4, 4, 5, 4, 5, 5, 6, 4, 5, 5, 6, 5, 6, 6, 7, |
|
|
3, 4, 4, 5, 4, 5, 5, 6, 4, 5, 5, 6, 5, 6, 6, 7, |
|
|
4, 5, 5, 6, 5, 6, 6, 7, 5, 6, 6, 7, 6, 7, 7, 8, |
|
|
}; |
|
|
} |
|
|
|
|
|
static v128_t wasm_unpacklo_i8x16(v128_t a, v128_t b) { |
|
|
return wasm_i8x16_shuffle(a, b, 0,16,1,17,2,18,3,19,4,20,5,21,6,22,7,23); |
|
|
} |
|
|
|
|
|
static v128_t wasm_unpacklo_i16x8(v128_t a, v128_t b) { |
|
|
return wasm_i8x16_shuffle(a, b, 0,1,16,17,2,3,18,19,4,5,20,21,6,7,22,23); |
|
|
} |
|
|
|
|
|
static v128_t wasm_unpacklo_i32x4(v128_t a, v128_t b) { |
|
|
return wasm_i8x16_shuffle(a, b, 0,1,2,3,16,17,18,19,4,5,6,7,20,21,22,23); |
|
|
} |
|
|
|
|
|
static v128_t wasm_unpacklo_i64x2(v128_t a, v128_t b) { |
|
|
return wasm_i8x16_shuffle(a, b, 0,1,2,3,4,5,6,7,16,17,18,19,20,21,22,23); |
|
|
} |
|
|
|
|
|
static v128_t wasm_unpackhi_i8x16(v128_t a, v128_t b) { |
|
|
return wasm_i8x16_shuffle(a, b, 8,24,9,25,10,26,11,27,12,28,13,29,14,30,15,31); |
|
|
} |
|
|
|
|
|
static v128_t wasm_unpackhi_i16x8(v128_t a, v128_t b) { |
|
|
return wasm_i8x16_shuffle(a, b, 8,9,24,25,10,11,26,27,12,13,28,29,14,15,30,31); |
|
|
} |
|
|
|
|
|
static v128_t wasm_unpackhi_i32x4(v128_t a, v128_t b) { |
|
|
return wasm_i8x16_shuffle(a, b, 8,9,10,11,24,25,26,27,12,13,14,15,28,29,30,31); |
|
|
} |
|
|
|
|
|
static v128_t wasm_unpackhi_i64x2(v128_t a, v128_t b) { |
|
|
return wasm_i8x16_shuffle(a, b, 8,9,10,11,12,13,14,15,24,25,26,27,28,29,30,31); |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
inline v128_t v128_cvtu8x16_i16x8(const v128_t& a) |
|
|
{ |
|
|
const v128_t z = wasm_i8x16_splat(0); |
|
|
return wasm_unpacklo_i8x16(a, z); |
|
|
} |
|
|
inline v128_t v128_cvti8x16_i16x8(const v128_t& a) |
|
|
{ return wasm_i16x8_shr(wasm_unpacklo_i8x16(a, a), 8); } |
|
|
|
|
|
inline v128_t v128_cvtu8x16_i32x4(const v128_t& a) |
|
|
{ |
|
|
const v128_t z = wasm_i8x16_splat(0); |
|
|
return wasm_unpacklo_i16x8(wasm_unpacklo_i8x16(a, z), z); |
|
|
} |
|
|
inline v128_t v128_cvti8x16_i32x4(const v128_t& a) |
|
|
{ |
|
|
v128_t r = wasm_unpacklo_i8x16(a, a); |
|
|
r = wasm_unpacklo_i8x16(r, r); |
|
|
return wasm_i32x4_shr(r, 24); |
|
|
} |
|
|
|
|
|
inline v128_t v128_cvtu16x8_i32x4(const v128_t& a) |
|
|
{ |
|
|
const v128_t z = wasm_i8x16_splat(0); |
|
|
return wasm_unpacklo_i16x8(a, z); |
|
|
} |
|
|
inline v128_t v128_cvti16x8_i32x4(const v128_t& a) |
|
|
{ return wasm_i32x4_shr(wasm_unpacklo_i16x8(a, a), 16); } |
|
|
|
|
|
inline v128_t v128_cvtu32x4_i64x2(const v128_t& a) |
|
|
{ |
|
|
const v128_t z = wasm_i8x16_splat(0); |
|
|
return wasm_unpacklo_i32x4(a, z); |
|
|
} |
|
|
inline v128_t v128_cvti32x4_i64x2(const v128_t& a) |
|
|
{ return wasm_unpacklo_i32x4(a, wasm_i32x4_shr(a, 31)); } |
|
|
|
|
|
|
|
|
inline v128_t v128_cvtu8x16_i16x8_high(const v128_t& a) |
|
|
{ |
|
|
const v128_t z = wasm_i8x16_splat(0); |
|
|
return wasm_unpackhi_i8x16(a, z); |
|
|
} |
|
|
inline v128_t v128_cvti8x16_i16x8_high(const v128_t& a) |
|
|
{ return wasm_i16x8_shr(wasm_unpackhi_i8x16(a, a), 8); } |
|
|
|
|
|
inline v128_t v128_cvtu16x8_i32x4_high(const v128_t& a) |
|
|
{ |
|
|
const v128_t z = wasm_i8x16_splat(0); |
|
|
return wasm_unpackhi_i16x8(a, z); |
|
|
} |
|
|
inline v128_t v128_cvti16x8_i32x4_high(const v128_t& a) |
|
|
{ return wasm_i32x4_shr(wasm_unpackhi_i16x8(a, a), 16); } |
|
|
|
|
|
inline v128_t v128_cvtu32x4_i64x2_high(const v128_t& a) |
|
|
{ |
|
|
const v128_t z = wasm_i8x16_splat(0); |
|
|
return wasm_unpackhi_i32x4(a, z); |
|
|
} |
|
|
inline v128_t v128_cvti32x4_i64x2_high(const v128_t& a) |
|
|
{ return wasm_unpackhi_i32x4(a, wasm_i32x4_shr(a, 31)); } |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_INITVEC(_Tpvec, _Tp, suffix, zsuffix, _Tps) \ |
|
|
inline _Tpvec v_setzero_##suffix() { return _Tpvec(wasm_##zsuffix##_splat((_Tps)0)); } \ |
|
|
inline _Tpvec v_setall_##suffix(_Tp v) { return _Tpvec(wasm_##zsuffix##_splat((_Tps)v)); } \ |
|
|
template <> inline _Tpvec v_setzero_() { return v_setzero_##suffix(); } \ |
|
|
template <> inline _Tpvec v_setall_(_Tp v) { return v_setall_##suffix(v); } \ |
|
|
template<typename _Tpvec0> inline _Tpvec v_reinterpret_as_##suffix(const _Tpvec0& a) \ |
|
|
{ return _Tpvec(a.val); } |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_INITVEC(v_uint8x16, uchar, u8, i8x16, schar) |
|
|
OPENCV_HAL_IMPL_WASM_INITVEC(v_int8x16, schar, s8, i8x16, schar) |
|
|
OPENCV_HAL_IMPL_WASM_INITVEC(v_uint16x8, ushort, u16, i16x8, short) |
|
|
OPENCV_HAL_IMPL_WASM_INITVEC(v_int16x8, short, s16, i16x8, short) |
|
|
OPENCV_HAL_IMPL_WASM_INITVEC(v_uint32x4, unsigned, u32, i32x4, int) |
|
|
OPENCV_HAL_IMPL_WASM_INITVEC(v_int32x4, int, s32, i32x4, int) |
|
|
OPENCV_HAL_IMPL_WASM_INITVEC(v_float32x4, float, f32, f32x4, float) |
|
|
OPENCV_HAL_IMPL_WASM_INITVEC(v_uint64x2, uint64, u64, i64x2, int64) |
|
|
OPENCV_HAL_IMPL_WASM_INITVEC(v_int64x2, int64, s64, i64x2, int64) |
|
|
OPENCV_HAL_IMPL_WASM_INITVEC(v_float64x2, double, f64, f64x2, double) |
|
|
|
|
|
|
|
|
inline v_uint8x16 v_pack(const v_uint16x8& a, const v_uint16x8& b) |
|
|
{ |
|
|
v128_t maxval = wasm_i16x8_splat(255); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_u16x8_gt(a.val, maxval)); |
|
|
v128_t b1 = wasm_v128_bitselect(maxval, b.val, wasm_u16x8_gt(b.val, maxval)); |
|
|
return v_uint8x16(wasm_i8x16_shuffle(a1, b1, 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30)); |
|
|
} |
|
|
inline v_int8x16 v_pack(const v_int16x8& a, const v_int16x8& b) |
|
|
{ |
|
|
v128_t maxval = wasm_i16x8_splat(127); |
|
|
v128_t minval = wasm_i16x8_splat(-128); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_i16x8_gt(a.val, maxval)); |
|
|
v128_t b1 = wasm_v128_bitselect(maxval, b.val, wasm_i16x8_gt(b.val, maxval)); |
|
|
v128_t a2 = wasm_v128_bitselect(minval, a1, wasm_i16x8_lt(a1, minval)); |
|
|
v128_t b2 = wasm_v128_bitselect(minval, b1, wasm_i16x8_lt(b1, minval)); |
|
|
return v_int8x16(wasm_i8x16_shuffle(a2, b2, 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30)); |
|
|
} |
|
|
inline v_uint16x8 v_pack(const v_uint32x4& a, const v_uint32x4& b) |
|
|
{ |
|
|
v128_t maxval = wasm_i32x4_splat(65535); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_u32x4_gt(a.val, maxval)); |
|
|
v128_t b1 = wasm_v128_bitselect(maxval, b.val, wasm_u32x4_gt(b.val, maxval)); |
|
|
return v_uint16x8(wasm_i8x16_shuffle(a1, b1, 0,1,4,5,8,9,12,13,16,17,20,21,24,25,28,29)); |
|
|
} |
|
|
inline v_int16x8 v_pack(const v_int32x4& a, const v_int32x4& b) |
|
|
{ |
|
|
v128_t maxval = wasm_i32x4_splat(32767); |
|
|
v128_t minval = wasm_i32x4_splat(-32768); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_i32x4_gt(a.val, maxval)); |
|
|
v128_t b1 = wasm_v128_bitselect(maxval, b.val, wasm_i32x4_gt(b.val, maxval)); |
|
|
v128_t a2 = wasm_v128_bitselect(minval, a1, wasm_i32x4_lt(a1, minval)); |
|
|
v128_t b2 = wasm_v128_bitselect(minval, b1, wasm_i32x4_lt(b1, minval)); |
|
|
return v_int16x8(wasm_i8x16_shuffle(a2, b2, 0,1,4,5,8,9,12,13,16,17,20,21,24,25,28,29)); |
|
|
} |
|
|
inline v_uint32x4 v_pack(const v_uint64x2& a, const v_uint64x2& b) |
|
|
{ |
|
|
return v_uint32x4(wasm_i8x16_shuffle(a.val, b.val, 0,1,2,3,8,9,10,11,16,17,18,19,24,25,26,27)); |
|
|
} |
|
|
inline v_int32x4 v_pack(const v_int64x2& a, const v_int64x2& b) |
|
|
{ |
|
|
return v_int32x4(wasm_i8x16_shuffle(a.val, b.val, 0,1,2,3,8,9,10,11,16,17,18,19,24,25,26,27)); |
|
|
} |
|
|
inline v_uint8x16 v_pack_u(const v_int16x8& a, const v_int16x8& b) |
|
|
{ |
|
|
v128_t maxval = wasm_i16x8_splat(255); |
|
|
v128_t minval = wasm_i16x8_splat(0); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_i16x8_gt(a.val, maxval)); |
|
|
v128_t b1 = wasm_v128_bitselect(maxval, b.val, wasm_i16x8_gt(b.val, maxval)); |
|
|
v128_t a2 = wasm_v128_bitselect(minval, a1, wasm_i16x8_lt(a1, minval)); |
|
|
v128_t b2 = wasm_v128_bitselect(minval, b1, wasm_i16x8_lt(b1, minval)); |
|
|
return v_uint8x16(wasm_i8x16_shuffle(a2, b2, 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30)); |
|
|
} |
|
|
inline v_uint16x8 v_pack_u(const v_int32x4& a, const v_int32x4& b) |
|
|
{ |
|
|
v128_t maxval = wasm_i32x4_splat(65535); |
|
|
v128_t minval = wasm_i32x4_splat(0); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_i32x4_gt(a.val, maxval)); |
|
|
v128_t b1 = wasm_v128_bitselect(maxval, b.val, wasm_i32x4_gt(b.val, maxval)); |
|
|
v128_t a2 = wasm_v128_bitselect(minval, a1, wasm_i32x4_lt(a1, minval)); |
|
|
v128_t b2 = wasm_v128_bitselect(minval, b1, wasm_i32x4_lt(b1, minval)); |
|
|
return v_uint16x8(wasm_i8x16_shuffle(a2, b2, 0,1,4,5,8,9,12,13,16,17,20,21,24,25,28,29)); |
|
|
} |
|
|
|
|
|
template<int n> |
|
|
inline v_uint8x16 v_rshr_pack(const v_uint16x8& a, const v_uint16x8& b) |
|
|
{ |
|
|
v128_t delta = wasm_i16x8_splat(((short)1 << (n-1))); |
|
|
v128_t a1 = wasm_u16x8_shr(wasm_i16x8_add(a.val, delta), n); |
|
|
v128_t b1 = wasm_u16x8_shr(wasm_i16x8_add(b.val, delta), n); |
|
|
v128_t maxval = wasm_i16x8_splat(255); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_u16x8_gt(a1, maxval)); |
|
|
v128_t b2 = wasm_v128_bitselect(maxval, b1, wasm_u16x8_gt(b1, maxval)); |
|
|
return v_uint8x16(wasm_i8x16_shuffle(a2, b2, 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30)); |
|
|
} |
|
|
template<int n> |
|
|
inline v_int8x16 v_rshr_pack(const v_int16x8& a, const v_int16x8& b) |
|
|
{ |
|
|
v128_t delta = wasm_i16x8_splat(((short)1 << (n-1))); |
|
|
v128_t a1 = wasm_i16x8_shr(wasm_i16x8_add(a.val, delta), n); |
|
|
v128_t b1 = wasm_i16x8_shr(wasm_i16x8_add(b.val, delta), n); |
|
|
v128_t maxval = wasm_i16x8_splat(127); |
|
|
v128_t minval = wasm_i16x8_splat(-128); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_i16x8_gt(a1, maxval)); |
|
|
v128_t b2 = wasm_v128_bitselect(maxval, b1, wasm_i16x8_gt(b1, maxval)); |
|
|
v128_t a3 = wasm_v128_bitselect(minval, a2, wasm_i16x8_lt(a1, minval)); |
|
|
v128_t b3 = wasm_v128_bitselect(minval, b2, wasm_i16x8_lt(b1, minval)); |
|
|
return v_int8x16(wasm_i8x16_shuffle(a3, b3, 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30)); |
|
|
} |
|
|
template<int n> |
|
|
inline v_uint16x8 v_rshr_pack(const v_uint32x4& a, const v_uint32x4& b) |
|
|
{ |
|
|
v128_t delta = wasm_i32x4_splat(((int)1 << (n-1))); |
|
|
v128_t a1 = wasm_u32x4_shr(wasm_i32x4_add(a.val, delta), n); |
|
|
v128_t b1 = wasm_u32x4_shr(wasm_i32x4_add(b.val, delta), n); |
|
|
v128_t maxval = wasm_i32x4_splat(65535); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_u32x4_gt(a1, maxval)); |
|
|
v128_t b2 = wasm_v128_bitselect(maxval, b1, wasm_u32x4_gt(b1, maxval)); |
|
|
return v_uint16x8(wasm_i8x16_shuffle(a2, b2, 0,1,4,5,8,9,12,13,16,17,20,21,24,25,28,29)); |
|
|
} |
|
|
template<int n> |
|
|
inline v_int16x8 v_rshr_pack(const v_int32x4& a, const v_int32x4& b) |
|
|
{ |
|
|
v128_t delta = wasm_i32x4_splat(((int)1 << (n-1))); |
|
|
v128_t a1 = wasm_i32x4_shr(wasm_i32x4_add(a.val, delta), n); |
|
|
v128_t b1 = wasm_i32x4_shr(wasm_i32x4_add(b.val, delta), n); |
|
|
v128_t maxval = wasm_i32x4_splat(32767); |
|
|
v128_t minval = wasm_i16x8_splat(-32768); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_i32x4_gt(a1, maxval)); |
|
|
v128_t b2 = wasm_v128_bitselect(maxval, b1, wasm_i32x4_gt(b1, maxval)); |
|
|
v128_t a3 = wasm_v128_bitselect(minval, a2, wasm_i32x4_lt(a1, minval)); |
|
|
v128_t b3 = wasm_v128_bitselect(minval, b2, wasm_i32x4_lt(b1, minval)); |
|
|
return v_int16x8(wasm_i8x16_shuffle(a3, b3, 0,1,4,5,8,9,12,13,16,17,20,21,24,25,28,29)); |
|
|
} |
|
|
template<int n> |
|
|
inline v_uint32x4 v_rshr_pack(const v_uint64x2& a, const v_uint64x2& b) |
|
|
{ |
|
|
v128_t delta = wasm_i64x2_splat(((int64)1 << (n-1))); |
|
|
v128_t a1 = wasm_u64x2_shr(wasm_i64x2_add(a.val, delta), n); |
|
|
v128_t b1 = wasm_u64x2_shr(wasm_i64x2_add(b.val, delta), n); |
|
|
return v_uint32x4(wasm_i8x16_shuffle(a1, b1, 0,1,2,3,8,9,10,11,16,17,18,19,24,25,26,27)); |
|
|
} |
|
|
template<int n> |
|
|
inline v_int32x4 v_rshr_pack(const v_int64x2& a, const v_int64x2& b) |
|
|
{ |
|
|
v128_t delta = wasm_i64x2_splat(((int64)1 << (n-1))); |
|
|
v128_t a1 = wasm_i64x2_shr(wasm_i64x2_add(a.val, delta), n); |
|
|
v128_t b1 = wasm_i64x2_shr(wasm_i64x2_add(b.val, delta), n); |
|
|
return v_int32x4(wasm_i8x16_shuffle(a1, b1, 0,1,2,3,8,9,10,11,16,17,18,19,24,25,26,27)); |
|
|
} |
|
|
template<int n> |
|
|
inline v_uint8x16 v_rshr_pack_u(const v_int16x8& a, const v_int16x8& b) |
|
|
{ |
|
|
v128_t delta = wasm_i16x8_splat(((short)1 << (n-1))); |
|
|
v128_t a1 = wasm_i16x8_shr(wasm_i16x8_add(a.val, delta), n); |
|
|
v128_t b1 = wasm_i16x8_shr(wasm_i16x8_add(b.val, delta), n); |
|
|
v128_t maxval = wasm_i16x8_splat(255); |
|
|
v128_t minval = wasm_i16x8_splat(0); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_i16x8_gt(a1, maxval)); |
|
|
v128_t b2 = wasm_v128_bitselect(maxval, b1, wasm_i16x8_gt(b1, maxval)); |
|
|
v128_t a3 = wasm_v128_bitselect(minval, a2, wasm_i16x8_lt(a1, minval)); |
|
|
v128_t b3 = wasm_v128_bitselect(minval, b2, wasm_i16x8_lt(b1, minval)); |
|
|
return v_uint8x16(wasm_i8x16_shuffle(a3, b3, 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30)); |
|
|
} |
|
|
template<int n> |
|
|
inline v_uint16x8 v_rshr_pack_u(const v_int32x4& a, const v_int32x4& b) |
|
|
{ |
|
|
v128_t delta = wasm_i32x4_splat(((int)1 << (n-1))); |
|
|
v128_t a1 = wasm_i32x4_shr(wasm_i32x4_add(a.val, delta), n); |
|
|
v128_t b1 = wasm_i32x4_shr(wasm_i32x4_add(b.val, delta), n); |
|
|
v128_t maxval = wasm_i32x4_splat(65535); |
|
|
v128_t minval = wasm_i16x8_splat(0); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_i32x4_gt(a1, maxval)); |
|
|
v128_t b2 = wasm_v128_bitselect(maxval, b1, wasm_i32x4_gt(b1, maxval)); |
|
|
v128_t a3 = wasm_v128_bitselect(minval, a2, wasm_i32x4_lt(a1, minval)); |
|
|
v128_t b3 = wasm_v128_bitselect(minval, b2, wasm_i32x4_lt(b1, minval)); |
|
|
return v_uint16x8(wasm_i8x16_shuffle(a3, b3, 0,1,4,5,8,9,12,13,16,17,20,21,24,25,28,29)); |
|
|
} |
|
|
|
|
|
inline void v_pack_store(uchar* ptr, const v_uint16x8& a) |
|
|
{ |
|
|
v128_t maxval = wasm_i16x8_splat(255); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_u16x8_gt(a.val, maxval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a1, a1, 0,2,4,6,8,10,12,14,0,2,4,6,8,10,12,14); |
|
|
uchar t_ptr[16]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<8; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
inline void v_pack_store(schar* ptr, const v_int16x8& a) |
|
|
{ |
|
|
v128_t maxval = wasm_i16x8_splat(127); |
|
|
v128_t minval = wasm_i16x8_splat(-128); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_i16x8_gt(a.val, maxval)); |
|
|
v128_t a2 = wasm_v128_bitselect(minval, a1, wasm_i16x8_lt(a1, minval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a2, a2, 0,2,4,6,8,10,12,14,0,2,4,6,8,10,12,14); |
|
|
schar t_ptr[16]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<8; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
inline void v_pack_store(ushort* ptr, const v_uint32x4& a) |
|
|
{ |
|
|
v128_t maxval = wasm_i32x4_splat(65535); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_u32x4_gt(a.val, maxval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a1, a1, 0,1,4,5,8,9,12,13,0,1,4,5,8,9,12,13); |
|
|
ushort t_ptr[8]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<4; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
inline void v_pack_store(short* ptr, const v_int32x4& a) |
|
|
{ |
|
|
v128_t maxval = wasm_i32x4_splat(32767); |
|
|
v128_t minval = wasm_i32x4_splat(-32768); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_i32x4_gt(a.val, maxval)); |
|
|
v128_t a2 = wasm_v128_bitselect(minval, a1, wasm_i32x4_lt(a1, minval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a2, a2, 0,1,4,5,8,9,12,13,0,1,4,5,8,9,12,13); |
|
|
short t_ptr[8]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<4; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
inline void v_pack_store(unsigned* ptr, const v_uint64x2& a) |
|
|
{ |
|
|
v128_t r = wasm_i8x16_shuffle(a.val, a.val, 0,1,2,3,8,9,10,11,0,1,2,3,8,9,10,11); |
|
|
unsigned t_ptr[4]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<2; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
inline void v_pack_store(int* ptr, const v_int64x2& a) |
|
|
{ |
|
|
v128_t r = wasm_i8x16_shuffle(a.val, a.val, 0,1,2,3,8,9,10,11,0,1,2,3,8,9,10,11); |
|
|
int t_ptr[4]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<2; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
inline void v_pack_u_store(uchar* ptr, const v_int16x8& a) |
|
|
{ |
|
|
v128_t maxval = wasm_i16x8_splat(255); |
|
|
v128_t minval = wasm_i16x8_splat(0); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_i16x8_gt(a.val, maxval)); |
|
|
v128_t a2 = wasm_v128_bitselect(minval, a1, wasm_i16x8_lt(a1, minval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a2, a2, 0,2,4,6,8,10,12,14,0,2,4,6,8,10,12,14); |
|
|
uchar t_ptr[16]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<8; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
inline void v_pack_u_store(ushort* ptr, const v_int32x4& a) |
|
|
{ |
|
|
v128_t maxval = wasm_i32x4_splat(65535); |
|
|
v128_t minval = wasm_i32x4_splat(0); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_i32x4_gt(a.val, maxval)); |
|
|
v128_t a2 = wasm_v128_bitselect(minval, a1, wasm_i32x4_lt(a1, minval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a2, a2, 0,1,4,5,8,9,12,13,0,1,4,5,8,9,12,13); |
|
|
ushort t_ptr[8]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<4; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
|
|
|
template<int n> |
|
|
inline void v_rshr_pack_store(uchar* ptr, const v_uint16x8& a) |
|
|
{ |
|
|
v128_t delta = wasm_i16x8_splat((short)(1 << (n-1))); |
|
|
v128_t a1 = wasm_u16x8_shr(wasm_i16x8_add(a.val, delta), n); |
|
|
v128_t maxval = wasm_i16x8_splat(255); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_u16x8_gt(a1, maxval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a2, a2, 0,2,4,6,8,10,12,14,0,2,4,6,8,10,12,14); |
|
|
uchar t_ptr[16]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<8; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
template<int n> |
|
|
inline void v_rshr_pack_store(schar* ptr, const v_int16x8& a) |
|
|
{ |
|
|
v128_t delta = wasm_i16x8_splat(((short)1 << (n-1))); |
|
|
v128_t a1 = wasm_i16x8_shr(wasm_i16x8_add(a.val, delta), n); |
|
|
v128_t maxval = wasm_i16x8_splat(127); |
|
|
v128_t minval = wasm_i16x8_splat(-128); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_i16x8_gt(a1, maxval)); |
|
|
v128_t a3 = wasm_v128_bitselect(minval, a2, wasm_i16x8_lt(a1, minval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a3, a3, 0,2,4,6,8,10,12,14,0,2,4,6,8,10,12,14); |
|
|
schar t_ptr[16]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<8; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
template<int n> |
|
|
inline void v_rshr_pack_store(ushort* ptr, const v_uint32x4& a) |
|
|
{ |
|
|
v128_t delta = wasm_i32x4_splat(((int)1 << (n-1))); |
|
|
v128_t a1 = wasm_u32x4_shr(wasm_i32x4_add(a.val, delta), n); |
|
|
v128_t maxval = wasm_i32x4_splat(65535); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_u32x4_gt(a1, maxval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a2, a2, 0,1,4,5,8,9,12,13,0,1,4,5,8,9,12,13); |
|
|
ushort t_ptr[8]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<4; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
template<int n> |
|
|
inline void v_rshr_pack_store(short* ptr, const v_int32x4& a) |
|
|
{ |
|
|
v128_t delta = wasm_i32x4_splat(((int)1 << (n-1))); |
|
|
v128_t a1 = wasm_i32x4_shr(wasm_i32x4_add(a.val, delta), n); |
|
|
v128_t maxval = wasm_i32x4_splat(32767); |
|
|
v128_t minval = wasm_i32x4_splat(-32768); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_i32x4_gt(a1, maxval)); |
|
|
v128_t a3 = wasm_v128_bitselect(minval, a2, wasm_i32x4_lt(a1, minval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a3, a3, 0,1,4,5,8,9,12,13,0,1,4,5,8,9,12,13); |
|
|
short t_ptr[8]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<4; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
template<int n> |
|
|
inline void v_rshr_pack_store(unsigned* ptr, const v_uint64x2& a) |
|
|
{ |
|
|
v128_t delta = wasm_i64x2_splat(((int64)1 << (n-1))); |
|
|
v128_t a1 = wasm_u64x2_shr(wasm_i64x2_add(a.val, delta), n); |
|
|
v128_t r = wasm_i8x16_shuffle(a1, a1, 0,1,2,3,8,9,10,11,0,1,2,3,8,9,10,11); |
|
|
unsigned t_ptr[4]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<2; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
template<int n> |
|
|
inline void v_rshr_pack_store(int* ptr, const v_int64x2& a) |
|
|
{ |
|
|
v128_t delta = wasm_i64x2_splat(((int64)1 << (n-1))); |
|
|
v128_t a1 = wasm_i64x2_shr(wasm_i64x2_add(a.val, delta), n); |
|
|
v128_t r = wasm_i8x16_shuffle(a1, a1, 0,1,2,3,8,9,10,11,0,1,2,3,8,9,10,11); |
|
|
int t_ptr[4]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<2; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
template<int n> |
|
|
inline void v_rshr_pack_u_store(uchar* ptr, const v_int16x8& a) |
|
|
{ |
|
|
v128_t delta = wasm_i16x8_splat(((short)1 << (n-1))); |
|
|
v128_t a1 = wasm_i16x8_shr(wasm_i16x8_add(a.val, delta), n); |
|
|
v128_t maxval = wasm_i16x8_splat(255); |
|
|
v128_t minval = wasm_i16x8_splat(0); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_i16x8_gt(a1, maxval)); |
|
|
v128_t a3 = wasm_v128_bitselect(minval, a2, wasm_i16x8_lt(a1, minval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a3, a3, 0,2,4,6,8,10,12,14,0,2,4,6,8,10,12,14); |
|
|
uchar t_ptr[16]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<8; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
template<int n> |
|
|
inline void v_rshr_pack_u_store(ushort* ptr, const v_int32x4& a) |
|
|
{ |
|
|
v128_t delta = wasm_i32x4_splat(((int)1 << (n-1))); |
|
|
v128_t a1 = wasm_i32x4_shr(wasm_i32x4_add(a.val, delta), n); |
|
|
v128_t maxval = wasm_i32x4_splat(65535); |
|
|
v128_t minval = wasm_i32x4_splat(0); |
|
|
v128_t a2 = wasm_v128_bitselect(maxval, a1, wasm_i32x4_gt(a1, maxval)); |
|
|
v128_t a3 = wasm_v128_bitselect(minval, a2, wasm_i32x4_lt(a1, minval)); |
|
|
v128_t r = wasm_i8x16_shuffle(a3, a3, 0,1,4,5,8,9,12,13,0,1,4,5,8,9,12,13); |
|
|
ushort t_ptr[8]; |
|
|
wasm_v128_store(t_ptr, r); |
|
|
for (int i=0; i<4; ++i) { |
|
|
ptr[i] = t_ptr[i]; |
|
|
} |
|
|
} |
|
|
|
|
|
inline v_uint8x16 v_pack_b(const v_uint16x8& a, const v_uint16x8& b) |
|
|
{ |
|
|
v128_t maxval = wasm_i16x8_splat(255); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_u16x8_gt(a.val, maxval)); |
|
|
v128_t b1 = wasm_v128_bitselect(maxval, b.val, wasm_u16x8_gt(b.val, maxval)); |
|
|
return v_uint8x16(wasm_i8x16_shuffle(a1, b1, 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30)); |
|
|
} |
|
|
|
|
|
inline v_uint8x16 v_pack_b(const v_uint32x4& a, const v_uint32x4& b, |
|
|
const v_uint32x4& c, const v_uint32x4& d) |
|
|
{ |
|
|
v128_t maxval = wasm_i32x4_splat(255); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, wasm_u32x4_gt(a.val, maxval)); |
|
|
v128_t b1 = wasm_v128_bitselect(maxval, b.val, wasm_u32x4_gt(b.val, maxval)); |
|
|
v128_t c1 = wasm_v128_bitselect(maxval, c.val, wasm_u32x4_gt(c.val, maxval)); |
|
|
v128_t d1 = wasm_v128_bitselect(maxval, d.val, wasm_u32x4_gt(d.val, maxval)); |
|
|
v128_t ab = wasm_i8x16_shuffle(a1, b1, 0,4,8,12,16,20,24,28,0,4,8,12,16,20,24,28); |
|
|
v128_t cd = wasm_i8x16_shuffle(c1, d1, 0,4,8,12,16,20,24,28,0,4,8,12,16,20,24,28); |
|
|
return v_uint8x16(wasm_i8x16_shuffle(ab, cd, 0,1,2,3,4,5,6,7,16,17,18,19,20,21,22,23)); |
|
|
} |
|
|
|
|
|
inline v_uint8x16 v_pack_b(const v_uint64x2& a, const v_uint64x2& b, const v_uint64x2& c, |
|
|
const v_uint64x2& d, const v_uint64x2& e, const v_uint64x2& f, |
|
|
const v_uint64x2& g, const v_uint64x2& h) |
|
|
{ |
|
|
v128_t maxval = wasm_i32x4_splat(255); |
|
|
v128_t a1 = wasm_v128_bitselect(maxval, a.val, ((__u64x2)(a.val) > (__u64x2)maxval)); |
|
|
v128_t b1 = wasm_v128_bitselect(maxval, b.val, ((__u64x2)(b.val) > (__u64x2)maxval)); |
|
|
v128_t c1 = wasm_v128_bitselect(maxval, c.val, ((__u64x2)(c.val) > (__u64x2)maxval)); |
|
|
v128_t d1 = wasm_v128_bitselect(maxval, d.val, ((__u64x2)(d.val) > (__u64x2)maxval)); |
|
|
v128_t e1 = wasm_v128_bitselect(maxval, e.val, ((__u64x2)(e.val) > (__u64x2)maxval)); |
|
|
v128_t f1 = wasm_v128_bitselect(maxval, f.val, ((__u64x2)(f.val) > (__u64x2)maxval)); |
|
|
v128_t g1 = wasm_v128_bitselect(maxval, g.val, ((__u64x2)(g.val) > (__u64x2)maxval)); |
|
|
v128_t h1 = wasm_v128_bitselect(maxval, h.val, ((__u64x2)(h.val) > (__u64x2)maxval)); |
|
|
v128_t ab = wasm_i8x16_shuffle(a1, b1, 0,8,16,24,0,8,16,24,0,8,16,24,0,8,16,24); |
|
|
v128_t cd = wasm_i8x16_shuffle(c1, d1, 0,8,16,24,0,8,16,24,0,8,16,24,0,8,16,24); |
|
|
v128_t ef = wasm_i8x16_shuffle(e1, f1, 0,8,16,24,0,8,16,24,0,8,16,24,0,8,16,24); |
|
|
v128_t gh = wasm_i8x16_shuffle(g1, h1, 0,8,16,24,0,8,16,24,0,8,16,24,0,8,16,24); |
|
|
v128_t abcd = wasm_i8x16_shuffle(ab, cd, 0,1,2,3,16,17,18,19,0,1,2,3,16,17,18,19); |
|
|
v128_t efgh = wasm_i8x16_shuffle(ef, gh, 0,1,2,3,16,17,18,19,0,1,2,3,16,17,18,19); |
|
|
return v_uint8x16(wasm_i8x16_shuffle(abcd, efgh, 0,1,2,3,4,5,6,7,16,17,18,19,20,21,22,23)); |
|
|
} |
|
|
|
|
|
inline v_float32x4 v_matmul(const v_float32x4& v, const v_float32x4& m0, |
|
|
const v_float32x4& m1, const v_float32x4& m2, |
|
|
const v_float32x4& m3) |
|
|
{ |
|
|
v128_t v0 = wasm_f32x4_splat(wasm_f32x4_extract_lane(v.val, 0)); |
|
|
v128_t v1 = wasm_f32x4_splat(wasm_f32x4_extract_lane(v.val, 1)); |
|
|
v128_t v2 = wasm_f32x4_splat(wasm_f32x4_extract_lane(v.val, 2)); |
|
|
v128_t v3 = wasm_f32x4_splat(wasm_f32x4_extract_lane(v.val, 3)); |
|
|
v0 = wasm_f32x4_mul(v0, m0.val); |
|
|
v1 = wasm_f32x4_mul(v1, m1.val); |
|
|
v2 = wasm_f32x4_mul(v2, m2.val); |
|
|
v3 = wasm_f32x4_mul(v3, m3.val); |
|
|
|
|
|
return v_float32x4(wasm_f32x4_add(wasm_f32x4_add(v0, v1), wasm_f32x4_add(v2, v3))); |
|
|
} |
|
|
|
|
|
inline v_float32x4 v_matmuladd(const v_float32x4& v, const v_float32x4& m0, |
|
|
const v_float32x4& m1, const v_float32x4& m2, |
|
|
const v_float32x4& a) |
|
|
{ |
|
|
v128_t v0 = wasm_f32x4_splat(wasm_f32x4_extract_lane(v.val, 0)); |
|
|
v128_t v1 = wasm_f32x4_splat(wasm_f32x4_extract_lane(v.val, 1)); |
|
|
v128_t v2 = wasm_f32x4_splat(wasm_f32x4_extract_lane(v.val, 2)); |
|
|
v0 = wasm_f32x4_mul(v0, m0.val); |
|
|
v1 = wasm_f32x4_mul(v1, m1.val); |
|
|
v2 = wasm_f32x4_mul(v2, m2.val); |
|
|
|
|
|
return v_float32x4(wasm_f32x4_add(wasm_f32x4_add(v0, v1), wasm_f32x4_add(v2, a.val))); |
|
|
} |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_BIN_OP(bin_op, _Tpvec, intrin) \ |
|
|
inline _Tpvec bin_op(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
return _Tpvec(intrin(a.val, b.val)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_add, v_uint8x16, wasm_u8x16_add_saturate) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_sub, v_uint8x16, wasm_u8x16_sub_saturate) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_add, v_int8x16, wasm_i8x16_add_saturate) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_sub, v_int8x16, wasm_i8x16_sub_saturate) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_add, v_uint16x8, wasm_u16x8_add_saturate) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_sub, v_uint16x8, wasm_u16x8_sub_saturate) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_add, v_int16x8, wasm_i16x8_add_saturate) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_sub, v_int16x8, wasm_i16x8_sub_saturate) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_add, v_uint32x4, wasm_i32x4_add) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_sub, v_uint32x4, wasm_i32x4_sub) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_mul, v_uint32x4, wasm_i32x4_mul) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_add, v_int32x4, wasm_i32x4_add) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_sub, v_int32x4, wasm_i32x4_sub) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_mul, v_int32x4, wasm_i32x4_mul) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_add, v_float32x4, wasm_f32x4_add) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_sub, v_float32x4, wasm_f32x4_sub) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_mul, v_float32x4, wasm_f32x4_mul) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_div, v_float32x4, wasm_f32x4_div) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_add, v_uint64x2, wasm_i64x2_add) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_sub, v_uint64x2, wasm_i64x2_sub) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_add, v_int64x2, wasm_i64x2_add) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_sub, v_int64x2, wasm_i64x2_sub) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_add, v_float64x2, wasm_f64x2_add) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_sub, v_float64x2, wasm_f64x2_sub) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_mul, v_float64x2, wasm_f64x2_mul) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_div, v_float64x2, wasm_f64x2_div) |
|
|
|
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_MUL_SAT(_Tpvec, _Tpwvec) \ |
|
|
inline _Tpvec v_mul(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
_Tpwvec c, d; \ |
|
|
v_mul_expand(a, b, c, d); \ |
|
|
return v_pack(c, d); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_MUL_SAT(v_uint8x16, v_uint16x8) |
|
|
OPENCV_HAL_IMPL_WASM_MUL_SAT(v_int8x16, v_int16x8) |
|
|
OPENCV_HAL_IMPL_WASM_MUL_SAT(v_uint16x8, v_uint32x4) |
|
|
OPENCV_HAL_IMPL_WASM_MUL_SAT(v_int16x8, v_int32x4) |
|
|
|
|
|
|
|
|
inline void v_mul_expand(const v_uint8x16& a, const v_uint8x16& b, |
|
|
v_uint16x8& c, v_uint16x8& d) |
|
|
{ |
|
|
v_uint16x8 a0, a1, b0, b1; |
|
|
v_expand(a, a0, a1); |
|
|
v_expand(b, b0, b1); |
|
|
c = v_mul_wrap(a0, b0); |
|
|
d = v_mul_wrap(a1, b1); |
|
|
} |
|
|
|
|
|
inline void v_mul_expand(const v_int8x16& a, const v_int8x16& b, |
|
|
v_int16x8& c, v_int16x8& d) |
|
|
{ |
|
|
v_int16x8 a0, a1, b0, b1; |
|
|
v_expand(a, a0, a1); |
|
|
v_expand(b, b0, b1); |
|
|
c = v_mul_wrap(a0, b0); |
|
|
d = v_mul_wrap(a1, b1); |
|
|
} |
|
|
|
|
|
inline void v_mul_expand(const v_int16x8& a, const v_int16x8& b, |
|
|
v_int32x4& c, v_int32x4& d) |
|
|
{ |
|
|
v_int32x4 a0, a1, b0, b1; |
|
|
v_expand(a, a0, a1); |
|
|
v_expand(b, b0, b1); |
|
|
c.val = wasm_i32x4_mul(a0.val, b0.val); |
|
|
d.val = wasm_i32x4_mul(a1.val, b1.val); |
|
|
} |
|
|
|
|
|
inline void v_mul_expand(const v_uint16x8& a, const v_uint16x8& b, |
|
|
v_uint32x4& c, v_uint32x4& d) |
|
|
{ |
|
|
v_uint32x4 a0, a1, b0, b1; |
|
|
v_expand(a, a0, a1); |
|
|
v_expand(b, b0, b1); |
|
|
c.val = wasm_i32x4_mul(a0.val, b0.val); |
|
|
d.val = wasm_i32x4_mul(a1.val, b1.val); |
|
|
} |
|
|
|
|
|
inline void v_mul_expand(const v_uint32x4& a, const v_uint32x4& b, |
|
|
v_uint64x2& c, v_uint64x2& d) |
|
|
{ |
|
|
v_uint64x2 a0, a1, b0, b1; |
|
|
v_expand(a, a0, a1); |
|
|
v_expand(b, b0, b1); |
|
|
c.val = ((__u64x2)(a0.val) * (__u64x2)(b0.val)); |
|
|
d.val = ((__u64x2)(a1.val) * (__u64x2)(b1.val)); |
|
|
} |
|
|
|
|
|
inline v_int16x8 v_mul_hi(const v_int16x8& a, const v_int16x8& b) |
|
|
{ |
|
|
v_int32x4 a0, a1, b0, b1; |
|
|
v_expand(a, a0, a1); |
|
|
v_expand(b, b0, b1); |
|
|
v128_t c = wasm_i32x4_mul(a0.val, b0.val); |
|
|
v128_t d = wasm_i32x4_mul(a1.val, b1.val); |
|
|
return v_int16x8(wasm_i8x16_shuffle(c, d, 2,3,6,7,10,11,14,15,18,19,22,23,26,27,30,31)); |
|
|
} |
|
|
inline v_uint16x8 v_mul_hi(const v_uint16x8& a, const v_uint16x8& b) |
|
|
{ |
|
|
v_uint32x4 a0, a1, b0, b1; |
|
|
v_expand(a, a0, a1); |
|
|
v_expand(b, b0, b1); |
|
|
v128_t c = wasm_i32x4_mul(a0.val, b0.val); |
|
|
v128_t d = wasm_i32x4_mul(a1.val, b1.val); |
|
|
return v_uint16x8(wasm_i8x16_shuffle(c, d, 2,3,6,7,10,11,14,15,18,19,22,23,26,27,30,31)); |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
inline v_int32x4 v_dotprod(const v_int16x8& a, const v_int16x8& b) |
|
|
{ |
|
|
v128_t a0 = wasm_i32x4_shr(wasm_i32x4_shl(a.val, 16), 16); |
|
|
v128_t a1 = wasm_i32x4_shr(a.val, 16); |
|
|
v128_t b0 = wasm_i32x4_shr(wasm_i32x4_shl(b.val, 16), 16); |
|
|
v128_t b1 = wasm_i32x4_shr(b.val, 16); |
|
|
v128_t c = wasm_i32x4_mul(a0, b0); |
|
|
v128_t d = wasm_i32x4_mul(a1, b1); |
|
|
return v_int32x4(wasm_i32x4_add(c, d)); |
|
|
} |
|
|
|
|
|
inline v_int32x4 v_dotprod(const v_int16x8& a, const v_int16x8& b, const v_int32x4& c) |
|
|
{ return v_add(v_dotprod(a, b), c); } |
|
|
|
|
|
inline v_int64x2 v_dotprod(const v_int32x4& a, const v_int32x4& b) |
|
|
{ |
|
|
v128_t a0 = wasm_i64x2_shr(wasm_i64x2_shl(a.val, 32), 32); |
|
|
v128_t a1 = wasm_i64x2_shr(a.val, 32); |
|
|
v128_t b0 = wasm_i64x2_shr(wasm_i64x2_shl(b.val, 32), 32); |
|
|
v128_t b1 = wasm_i64x2_shr(b.val, 32); |
|
|
v128_t c = (v128_t)((__i64x2)a0 * (__i64x2)b0); |
|
|
v128_t d = (v128_t)((__i64x2)a1 * (__i64x2)b1); |
|
|
return v_int64x2(wasm_i64x2_add(c, d)); |
|
|
} |
|
|
inline v_int64x2 v_dotprod(const v_int32x4& a, const v_int32x4& b, const v_int64x2& c) |
|
|
{ |
|
|
return v_add(v_dotprod(a, b), c); |
|
|
} |
|
|
|
|
|
|
|
|
inline v_uint32x4 v_dotprod_expand(const v_uint8x16& a, const v_uint8x16& b) |
|
|
{ |
|
|
v128_t a0 = wasm_u16x8_shr(wasm_i16x8_shl(a.val, 8), 8); |
|
|
v128_t a1 = wasm_u16x8_shr(a.val, 8); |
|
|
v128_t b0 = wasm_u16x8_shr(wasm_i16x8_shl(b.val, 8), 8); |
|
|
v128_t b1 = wasm_u16x8_shr(b.val, 8); |
|
|
return v_uint32x4((v_add( |
|
|
v_dotprod(v_int16x8(a0), v_int16x8(b0)), |
|
|
v_dotprod(v_int16x8(a1), v_int16x8(b1)))).val |
|
|
); |
|
|
} |
|
|
inline v_uint32x4 v_dotprod_expand(const v_uint8x16& a, const v_uint8x16& b, const v_uint32x4& c) |
|
|
{ return v_add(v_dotprod_expand(a, b), c); } |
|
|
|
|
|
inline v_int32x4 v_dotprod_expand(const v_int8x16& a, const v_int8x16& b) |
|
|
{ |
|
|
v128_t a0 = wasm_i16x8_shr(wasm_i16x8_shl(a.val, 8), 8); |
|
|
v128_t a1 = wasm_i16x8_shr(a.val, 8); |
|
|
v128_t b0 = wasm_i16x8_shr(wasm_i16x8_shl(b.val, 8), 8); |
|
|
v128_t b1 = wasm_i16x8_shr(b.val, 8); |
|
|
return v_int32x4(v_add( |
|
|
v_dotprod(v_int16x8(a0), v_int16x8(b0)), |
|
|
v_dotprod(v_int16x8(a1), v_int16x8(b1)) |
|
|
)); |
|
|
} |
|
|
inline v_int32x4 v_dotprod_expand(const v_int8x16& a, const v_int8x16& b, const v_int32x4& c) |
|
|
{ return v_add(v_dotprod_expand(a, b), c); } |
|
|
|
|
|
|
|
|
inline v_uint64x2 v_dotprod_expand(const v_uint16x8& a, const v_uint16x8& b) |
|
|
{ |
|
|
v128_t a0 = wasm_u32x4_shr(wasm_i32x4_shl(a.val, 16), 16); |
|
|
v128_t a1 = wasm_u32x4_shr(a.val, 16); |
|
|
v128_t b0 = wasm_u32x4_shr(wasm_i32x4_shl(b.val, 16), 16); |
|
|
v128_t b1 = wasm_u32x4_shr(b.val, 16); |
|
|
return v_uint64x2((v_add( |
|
|
v_dotprod(v_int32x4(a0), v_int32x4(b0)), |
|
|
v_dotprod(v_int32x4(a1), v_int32x4(b1))).val |
|
|
)); |
|
|
} |
|
|
inline v_uint64x2 v_dotprod_expand(const v_uint16x8& a, const v_uint16x8& b, const v_uint64x2& c) |
|
|
{ return v_add(v_dotprod_expand(a, b), c); } |
|
|
|
|
|
inline v_int64x2 v_dotprod_expand(const v_int16x8& a, const v_int16x8& b) |
|
|
{ |
|
|
v128_t a0 = wasm_i32x4_shr(wasm_i32x4_shl(a.val, 16), 16); |
|
|
v128_t a1 = wasm_i32x4_shr(a.val, 16); |
|
|
v128_t b0 = wasm_i32x4_shr(wasm_i32x4_shl(b.val, 16), 16); |
|
|
v128_t b1 = wasm_i32x4_shr(b.val, 16); |
|
|
return v_int64x2((v_add( |
|
|
v_dotprod(v_int32x4(a0), v_int32x4(b0)), |
|
|
v_dotprod(v_int32x4(a1), v_int32x4(b1))) |
|
|
)); |
|
|
} |
|
|
|
|
|
inline v_int64x2 v_dotprod_expand(const v_int16x8& a, const v_int16x8& b, const v_int64x2& c) |
|
|
{ return v_add(v_dotprod_expand(a, b), c); } |
|
|
|
|
|
|
|
|
inline v_float64x2 v_dotprod_expand(const v_int32x4& a, const v_int32x4& b) |
|
|
{ return v_cvt_f64(v_dotprod(a, b)); } |
|
|
inline v_float64x2 v_dotprod_expand(const v_int32x4& a, const v_int32x4& b, const v_float64x2& c) |
|
|
{ return v_add(v_dotprod_expand(a, b), c); } |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
inline v_int32x4 v_dotprod_fast(const v_int16x8& a, const v_int16x8& b) |
|
|
{ return v_dotprod(a, b); } |
|
|
inline v_int32x4 v_dotprod_fast(const v_int16x8& a, const v_int16x8& b, const v_int32x4& c) |
|
|
{ return v_dotprod(a, b, c); } |
|
|
|
|
|
|
|
|
inline v_int64x2 v_dotprod_fast(const v_int32x4& a, const v_int32x4& b) |
|
|
{ return v_dotprod(a, b); } |
|
|
inline v_int64x2 v_dotprod_fast(const v_int32x4& a, const v_int32x4& b, const v_int64x2& c) |
|
|
{ return v_dotprod(a, b, c); } |
|
|
|
|
|
|
|
|
inline v_uint32x4 v_dotprod_expand_fast(const v_uint8x16& a, const v_uint8x16& b) |
|
|
{ return v_dotprod_expand(a, b); } |
|
|
inline v_uint32x4 v_dotprod_expand_fast(const v_uint8x16& a, const v_uint8x16& b, const v_uint32x4& c) |
|
|
{ return v_dotprod_expand(a, b, c); } |
|
|
inline v_int32x4 v_dotprod_expand_fast(const v_int8x16& a, const v_int8x16& b) |
|
|
{ return v_dotprod_expand(a, b); } |
|
|
inline v_int32x4 v_dotprod_expand_fast(const v_int8x16& a, const v_int8x16& b, const v_int32x4& c) |
|
|
{ return v_dotprod_expand(a, b, c); } |
|
|
|
|
|
|
|
|
inline v_uint64x2 v_dotprod_expand_fast(const v_uint16x8& a, const v_uint16x8& b) |
|
|
{ return v_dotprod_expand(a, b); } |
|
|
inline v_uint64x2 v_dotprod_expand_fast(const v_uint16x8& a, const v_uint16x8& b, const v_uint64x2& c) |
|
|
{ return v_dotprod_expand(a, b, c); } |
|
|
inline v_int64x2 v_dotprod_expand_fast(const v_int16x8& a, const v_int16x8& b) |
|
|
{ return v_dotprod_expand(a, b); } |
|
|
inline v_int64x2 v_dotprod_expand_fast(const v_int16x8& a, const v_int16x8& b, const v_int64x2& c) |
|
|
{ return v_dotprod_expand(a, b, c); } |
|
|
|
|
|
|
|
|
inline v_float64x2 v_dotprod_expand_fast(const v_int32x4& a, const v_int32x4& b) |
|
|
{ return v_dotprod_expand(a, b); } |
|
|
inline v_float64x2 v_dotprod_expand_fast(const v_int32x4& a, const v_int32x4& b, const v_float64x2& c) |
|
|
{ return v_dotprod_expand(a, b, c); } |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_LOGIC_OP(_Tpvec) \ |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_and, _Tpvec, wasm_v128_and) \ |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_or, _Tpvec, wasm_v128_or) \ |
|
|
OPENCV_HAL_IMPL_WASM_BIN_OP(v_xor, _Tpvec, wasm_v128_xor) \ |
|
|
inline _Tpvec v_not(const _Tpvec& a) \ |
|
|
{ \ |
|
|
return _Tpvec(wasm_v128_not(a.val)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_LOGIC_OP(v_uint8x16) |
|
|
OPENCV_HAL_IMPL_WASM_LOGIC_OP(v_int8x16) |
|
|
OPENCV_HAL_IMPL_WASM_LOGIC_OP(v_uint16x8) |
|
|
OPENCV_HAL_IMPL_WASM_LOGIC_OP(v_int16x8) |
|
|
OPENCV_HAL_IMPL_WASM_LOGIC_OP(v_uint32x4) |
|
|
OPENCV_HAL_IMPL_WASM_LOGIC_OP(v_int32x4) |
|
|
OPENCV_HAL_IMPL_WASM_LOGIC_OP(v_uint64x2) |
|
|
OPENCV_HAL_IMPL_WASM_LOGIC_OP(v_int64x2) |
|
|
OPENCV_HAL_IMPL_WASM_LOGIC_OP(v_float32x4) |
|
|
OPENCV_HAL_IMPL_WASM_LOGIC_OP(v_float64x2) |
|
|
|
|
|
inline v_float32x4 v_sqrt(const v_float32x4& x) |
|
|
{ |
|
|
return v_float32x4(wasm_f32x4_sqrt(x.val)); |
|
|
} |
|
|
|
|
|
inline v_float32x4 v_invsqrt(const v_float32x4& x) |
|
|
{ |
|
|
const v128_t _1_0 = wasm_f32x4_splat(1.0); |
|
|
return v_float32x4(wasm_f32x4_div(_1_0, wasm_f32x4_sqrt(x.val))); |
|
|
} |
|
|
|
|
|
inline v_float64x2 v_sqrt(const v_float64x2& x) |
|
|
{ |
|
|
return v_float64x2(wasm_f64x2_sqrt(x.val)); |
|
|
} |
|
|
|
|
|
inline v_float64x2 v_invsqrt(const v_float64x2& x) |
|
|
{ |
|
|
const v128_t _1_0 = wasm_f64x2_splat(1.0); |
|
|
return v_float64x2(wasm_f64x2_div(_1_0, wasm_f64x2_sqrt(x.val))); |
|
|
} |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_ABS_INT_FUNC(_Tpuvec, _Tpsvec, suffix, zsuffix, shiftWidth) \ |
|
|
inline _Tpuvec v_abs(const _Tpsvec& x) \ |
|
|
{ \ |
|
|
v128_t s = wasm_##suffix##_shr(x.val, shiftWidth); \ |
|
|
v128_t f = wasm_##zsuffix##_shr(x.val, shiftWidth); \ |
|
|
return _Tpuvec(wasm_##zsuffix##_add(wasm_v128_xor(x.val, f), s)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_ABS_INT_FUNC(v_uint8x16, v_int8x16, u8x16, i8x16, 7) |
|
|
OPENCV_HAL_IMPL_WASM_ABS_INT_FUNC(v_uint16x8, v_int16x8, u16x8, i16x8, 15) |
|
|
OPENCV_HAL_IMPL_WASM_ABS_INT_FUNC(v_uint32x4, v_int32x4, u32x4, i32x4, 31) |
|
|
|
|
|
inline v_float32x4 v_abs(const v_float32x4& x) |
|
|
{ return v_float32x4(wasm_f32x4_abs(x.val)); } |
|
|
inline v_float64x2 v_abs(const v_float64x2& x) |
|
|
{ |
|
|
return v_float64x2(wasm_f64x2_abs(x.val)); |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_BIN_FUNC(_Tpvec, func, intrin) \ |
|
|
inline _Tpvec func(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
return _Tpvec(intrin(a.val, b.val)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_float32x4, v_min, wasm_f32x4_min) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_float32x4, v_max, wasm_f32x4_max) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_float64x2, v_min, wasm_f64x2_min) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_float64x2, v_max, wasm_f64x2_max) |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_MINMAX_S_INIT_FUNC(_Tpvec, suffix) \ |
|
|
inline _Tpvec v_min(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
return _Tpvec(wasm_v128_bitselect(b.val, a.val, wasm_##suffix##_gt(a.val, b.val))); \ |
|
|
} \ |
|
|
inline _Tpvec v_max(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
return _Tpvec(wasm_v128_bitselect(a.val, b.val, wasm_##suffix##_gt(a.val, b.val))); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_MINMAX_S_INIT_FUNC(v_int8x16, i8x16) |
|
|
OPENCV_HAL_IMPL_WASM_MINMAX_S_INIT_FUNC(v_int16x8, i16x8) |
|
|
OPENCV_HAL_IMPL_WASM_MINMAX_S_INIT_FUNC(v_int32x4, i32x4) |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_MINMAX_U_INIT_FUNC(_Tpvec, suffix, deltaNum) \ |
|
|
inline _Tpvec v_min(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
v128_t delta = wasm_##suffix##_splat(deltaNum); \ |
|
|
v128_t mask = wasm_##suffix##_gt(wasm_v128_xor(a.val, delta), wasm_v128_xor(b.val, delta)); \ |
|
|
return _Tpvec(wasm_v128_bitselect(b.val, a.val, mask)); \ |
|
|
} \ |
|
|
inline _Tpvec v_max(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
v128_t delta = wasm_##suffix##_splat(deltaNum); \ |
|
|
v128_t mask = wasm_##suffix##_gt(wasm_v128_xor(a.val, delta), wasm_v128_xor(b.val, delta)); \ |
|
|
return _Tpvec(wasm_v128_bitselect(a.val, b.val, mask)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_MINMAX_U_INIT_FUNC(v_uint8x16, i8x16, (schar)0x80) |
|
|
OPENCV_HAL_IMPL_WASM_MINMAX_U_INIT_FUNC(v_uint16x8, i16x8, (short)0x8000) |
|
|
OPENCV_HAL_IMPL_WASM_MINMAX_U_INIT_FUNC(v_uint32x4, i32x4, (int)0x80000000) |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_INIT_CMP_OP(_Tpvec, suffix, esuffix) \ |
|
|
inline _Tpvec v_eq(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ return _Tpvec(wasm_##esuffix##_eq(a.val, b.val)); } \ |
|
|
inline _Tpvec v_ne(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ return _Tpvec(wasm_##esuffix##_ne(a.val, b.val)); } \ |
|
|
inline _Tpvec v_lt(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ return _Tpvec(wasm_##suffix##_lt(a.val, b.val)); } \ |
|
|
inline _Tpvec v_gt(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ return _Tpvec(wasm_##suffix##_gt(a.val, b.val)); } \ |
|
|
inline _Tpvec v_le(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ return _Tpvec(wasm_##suffix##_le(a.val, b.val)); } \ |
|
|
inline _Tpvec v_ge(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ return _Tpvec(wasm_##suffix##_ge(a.val, b.val)); } |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_INIT_CMP_OP(v_uint8x16, u8x16, i8x16) |
|
|
OPENCV_HAL_IMPL_WASM_INIT_CMP_OP(v_int8x16, i8x16, i8x16) |
|
|
OPENCV_HAL_IMPL_WASM_INIT_CMP_OP(v_uint16x8, u16x8, i16x8) |
|
|
OPENCV_HAL_IMPL_WASM_INIT_CMP_OP(v_int16x8, i16x8, i16x8) |
|
|
OPENCV_HAL_IMPL_WASM_INIT_CMP_OP(v_uint32x4, u32x4, i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_INIT_CMP_OP(v_int32x4, i32x4, i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_INIT_CMP_OP(v_float32x4, f32x4, f32x4) |
|
|
OPENCV_HAL_IMPL_WASM_INIT_CMP_OP(v_float64x2, f64x2, f64x2) |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_64BIT_CMP_OP(_Tpvec, cast) \ |
|
|
inline _Tpvec v_eq(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ return cast(v_eq(v_reinterpret_as_f64(a), v_reinterpret_as_f64(b))); } \ |
|
|
inline _Tpvec v_ne(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ return cast(v_ne(v_reinterpret_as_f64(a), v_reinterpret_as_f64(b))); } |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_64BIT_CMP_OP(v_uint64x2, v_reinterpret_as_u64) |
|
|
OPENCV_HAL_IMPL_WASM_64BIT_CMP_OP(v_int64x2, v_reinterpret_as_s64) |
|
|
|
|
|
inline v_float32x4 v_not_nan(const v_float32x4& a) |
|
|
{ |
|
|
v128_t z = wasm_i32x4_splat(0x7fffffff); |
|
|
v128_t t = wasm_i32x4_splat(0x7f800000); |
|
|
return v_float32x4(wasm_u32x4_lt(wasm_v128_and(a.val, z), t)); |
|
|
} |
|
|
inline v_float64x2 v_not_nan(const v_float64x2& a) |
|
|
{ |
|
|
v128_t z = wasm_i64x2_splat(0x7fffffffffffffff); |
|
|
v128_t t = wasm_i64x2_splat(0x7ff0000000000000); |
|
|
return v_float64x2((__u64x2)(wasm_v128_and(a.val, z)) < (__u64x2)t); |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_uint8x16, v_add_wrap, wasm_i8x16_add) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_int8x16, v_add_wrap, wasm_i8x16_add) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_uint16x8, v_add_wrap, wasm_i16x8_add) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_int16x8, v_add_wrap, wasm_i16x8_add) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_uint8x16, v_sub_wrap, wasm_i8x16_sub) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_int8x16, v_sub_wrap, wasm_i8x16_sub) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_uint16x8, v_sub_wrap, wasm_i16x8_sub) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_int16x8, v_sub_wrap, wasm_i16x8_sub) |
|
|
#if (__EMSCRIPTEN_major__ * 1000000 + __EMSCRIPTEN_minor__ * 1000 + __EMSCRIPTEN_tiny__) >= (1039012) |
|
|
|
|
|
|
|
|
inline v_uint8x16 v_mul_wrap(const v_uint8x16& a, const v_uint8x16& b) |
|
|
{ |
|
|
uchar a_[16], b_[16]; |
|
|
wasm_v128_store(a_, a.val); |
|
|
wasm_v128_store(b_, b.val); |
|
|
for (int i = 0; i < 16; i++) |
|
|
a_[i] = (uchar)(a_[i] * b_[i]); |
|
|
return v_uint8x16(wasm_v128_load(a_)); |
|
|
} |
|
|
inline v_int8x16 v_mul_wrap(const v_int8x16& a, const v_int8x16& b) |
|
|
{ |
|
|
schar a_[16], b_[16]; |
|
|
wasm_v128_store(a_, a.val); |
|
|
wasm_v128_store(b_, b.val); |
|
|
for (int i = 0; i < 16; i++) |
|
|
a_[i] = (schar)(a_[i] * b_[i]); |
|
|
return v_int8x16(wasm_v128_load(a_)); |
|
|
} |
|
|
#else |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_uint8x16, v_mul_wrap, wasm_i8x16_mul) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_int8x16, v_mul_wrap, wasm_i8x16_mul) |
|
|
#endif |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_uint16x8, v_mul_wrap, wasm_i16x8_mul) |
|
|
OPENCV_HAL_IMPL_WASM_BIN_FUNC(v_int16x8, v_mul_wrap, wasm_i16x8_mul) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
inline v_uint8x16 v_absdiff(const v_uint8x16& a, const v_uint8x16& b) |
|
|
{ return v_add_wrap(v_sub(a, b), v_sub(b, a)); } |
|
|
inline v_uint16x8 v_absdiff(const v_uint16x8& a, const v_uint16x8& b) |
|
|
{ return v_add_wrap(v_sub(a, b), v_sub(b, a)); } |
|
|
inline v_uint32x4 v_absdiff(const v_uint32x4& a, const v_uint32x4& b) |
|
|
{ return v_sub(v_max(a, b), v_min(a, b)); } |
|
|
|
|
|
inline v_uint8x16 v_absdiff(const v_int8x16& a, const v_int8x16& b) |
|
|
{ |
|
|
v_int8x16 d = v_sub_wrap(a, b); |
|
|
v_int8x16 m = v_lt(a, b); |
|
|
return v_reinterpret_as_u8(v_sub_wrap(v_xor(d, m), m)); |
|
|
} |
|
|
inline v_uint16x8 v_absdiff(const v_int16x8& a, const v_int16x8& b) |
|
|
{ |
|
|
return v_reinterpret_as_u16(v_sub_wrap(v_max(a, b), v_min(a, b))); |
|
|
} |
|
|
inline v_uint32x4 v_absdiff(const v_int32x4& a, const v_int32x4& b) |
|
|
{ |
|
|
v_int32x4 d = v_sub(a, b); |
|
|
v_int32x4 m = v_lt(a, b); |
|
|
return v_reinterpret_as_u32(v_sub(v_xor(d, m), m)); |
|
|
} |
|
|
|
|
|
|
|
|
inline v_int8x16 v_absdiffs(const v_int8x16& a, const v_int8x16& b) |
|
|
{ |
|
|
v_int8x16 d = v_sub(a, b); |
|
|
v_int8x16 m = v_lt(a, b); |
|
|
return v_sub(v_xor(d, m), m); |
|
|
} |
|
|
inline v_int16x8 v_absdiffs(const v_int16x8& a, const v_int16x8& b) |
|
|
{ return v_sub(v_max(a, b), v_min(a, b)); } |
|
|
|
|
|
|
|
|
inline v_int32x4 v_fma(const v_int32x4& a, const v_int32x4& b, const v_int32x4& c) |
|
|
{ |
|
|
return v_add(v_mul(a, b), c); |
|
|
} |
|
|
|
|
|
inline v_int32x4 v_muladd(const v_int32x4& a, const v_int32x4& b, const v_int32x4& c) |
|
|
{ |
|
|
return v_fma(a, b, c); |
|
|
} |
|
|
|
|
|
inline v_float32x4 v_fma(const v_float32x4& a, const v_float32x4& b, const v_float32x4& c) |
|
|
{ |
|
|
return v_add(v_mul(a, b), c); |
|
|
} |
|
|
|
|
|
inline v_float64x2 v_fma(const v_float64x2& a, const v_float64x2& b, const v_float64x2& c) |
|
|
{ |
|
|
return v_add(v_mul(a, b), c); |
|
|
} |
|
|
|
|
|
inline v_float32x4 v_absdiff(const v_float32x4& a, const v_float32x4& b) |
|
|
{ |
|
|
v128_t absmask_vec = wasm_i32x4_splat(0x7fffffff); |
|
|
return v_float32x4(wasm_v128_and(wasm_f32x4_sub(a.val, b.val), absmask_vec)); |
|
|
} |
|
|
inline v_float64x2 v_absdiff(const v_float64x2& a, const v_float64x2& b) |
|
|
{ |
|
|
v128_t absmask_vec = wasm_u64x2_shr(wasm_i32x4_splat(-1), 1); |
|
|
return v_float64x2(wasm_v128_and(wasm_f64x2_sub(a.val, b.val), absmask_vec)); |
|
|
} |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_MISC_FLT_OP(_Tpvec, suffix) \ |
|
|
inline _Tpvec v_magnitude(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
v128_t a_Square = wasm_##suffix##_mul(a.val, a.val); \ |
|
|
v128_t b_Square = wasm_##suffix##_mul(b.val, b.val); \ |
|
|
return _Tpvec(wasm_##suffix##_sqrt(wasm_##suffix##_add(a_Square, b_Square))); \ |
|
|
} \ |
|
|
inline _Tpvec v_sqr_magnitude(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
v128_t a_Square = wasm_##suffix##_mul(a.val, a.val); \ |
|
|
v128_t b_Square = wasm_##suffix##_mul(b.val, b.val); \ |
|
|
return _Tpvec(wasm_##suffix##_add(a_Square, b_Square)); \ |
|
|
} \ |
|
|
inline _Tpvec v_muladd(const _Tpvec& a, const _Tpvec& b, const _Tpvec& c) \ |
|
|
{ \ |
|
|
return _Tpvec(wasm_##suffix##_add(wasm_##suffix##_mul(a.val, b.val), c.val)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_MISC_FLT_OP(v_float32x4, f32x4) |
|
|
OPENCV_HAL_IMPL_WASM_MISC_FLT_OP(v_float64x2, f64x2) |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_SHIFT_OP(_Tpuvec, _Tpsvec, suffix, ssuffix) \ |
|
|
inline _Tpuvec v_shl(const _Tpuvec& a, int imm) \ |
|
|
{ \ |
|
|
return _Tpuvec(wasm_##suffix##_shl(a.val, imm)); \ |
|
|
} \ |
|
|
inline _Tpsvec v_shl(const _Tpsvec& a, int imm) \ |
|
|
{ \ |
|
|
return _Tpsvec(wasm_##suffix##_shl(a.val, imm)); \ |
|
|
} \ |
|
|
inline _Tpuvec v_shr(const _Tpuvec& a, int imm) \ |
|
|
{ \ |
|
|
return _Tpuvec(wasm_##ssuffix##_shr(a.val, imm)); \ |
|
|
} \ |
|
|
inline _Tpsvec v_shr(const _Tpsvec& a, int imm) \ |
|
|
{ \ |
|
|
return _Tpsvec(wasm_##suffix##_shr(a.val, imm)); \ |
|
|
} \ |
|
|
template<int imm> \ |
|
|
inline _Tpuvec v_shl(const _Tpuvec& a) \ |
|
|
{ \ |
|
|
return _Tpuvec(wasm_##suffix##_shl(a.val, imm)); \ |
|
|
} \ |
|
|
template<int imm> \ |
|
|
inline _Tpsvec v_shl(const _Tpsvec& a) \ |
|
|
{ \ |
|
|
return _Tpsvec(wasm_##suffix##_shl(a.val, imm)); \ |
|
|
} \ |
|
|
template<int imm> \ |
|
|
inline _Tpuvec v_shr(const _Tpuvec& a) \ |
|
|
{ \ |
|
|
return _Tpuvec(wasm_##ssuffix##_shr(a.val, imm)); \ |
|
|
} \ |
|
|
template<int imm> \ |
|
|
inline _Tpsvec v_shr(const _Tpsvec& a) \ |
|
|
{ \ |
|
|
return _Tpsvec(wasm_##suffix##_shr(a.val, imm)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_SHIFT_OP(v_uint8x16, v_int8x16, i8x16, u8x16) |
|
|
OPENCV_HAL_IMPL_WASM_SHIFT_OP(v_uint16x8, v_int16x8, i16x8, u16x8) |
|
|
OPENCV_HAL_IMPL_WASM_SHIFT_OP(v_uint32x4, v_int32x4, i32x4, u32x4) |
|
|
OPENCV_HAL_IMPL_WASM_SHIFT_OP(v_uint64x2, v_int64x2, i64x2, u64x2) |
|
|
|
|
|
namespace hal_wasm_internal |
|
|
{ |
|
|
template <int imm, |
|
|
bool is_invalid = ((imm < 0) || (imm > 16)), |
|
|
bool is_first = (imm == 0), |
|
|
bool is_second = (imm == 16), |
|
|
bool is_other = (((imm > 0) && (imm < 16)))> |
|
|
class v_wasm_palignr_u8_class; |
|
|
|
|
|
template <int imm> |
|
|
class v_wasm_palignr_u8_class<imm, true, false, false, false>; |
|
|
|
|
|
template <int imm> |
|
|
class v_wasm_palignr_u8_class<imm, false, true, false, false> |
|
|
{ |
|
|
public: |
|
|
inline v128_t operator()(const v128_t& a, const v128_t&) const |
|
|
{ |
|
|
return a; |
|
|
} |
|
|
}; |
|
|
|
|
|
template <int imm> |
|
|
class v_wasm_palignr_u8_class<imm, false, false, true, false> |
|
|
{ |
|
|
public: |
|
|
inline v128_t operator()(const v128_t&, const v128_t& b) const |
|
|
{ |
|
|
return b; |
|
|
} |
|
|
}; |
|
|
|
|
|
template <int imm> |
|
|
class v_wasm_palignr_u8_class<imm, false, false, false, true> |
|
|
{ |
|
|
public: |
|
|
inline v128_t operator()(const v128_t& a, const v128_t& b) const |
|
|
{ |
|
|
enum { imm2 = (sizeof(v128_t) - imm) }; |
|
|
return wasm_i8x16_shuffle(a, b, |
|
|
imm, imm+1, imm+2, imm+3, |
|
|
imm+4, imm+5, imm+6, imm+7, |
|
|
imm+8, imm+9, imm+10, imm+11, |
|
|
imm+12, imm+13, imm+14, imm+15); |
|
|
} |
|
|
}; |
|
|
|
|
|
template <int imm> |
|
|
inline v128_t v_wasm_palignr_u8(const v128_t& a, const v128_t& b) |
|
|
{ |
|
|
CV_StaticAssert((imm >= 0) && (imm <= 16), "Invalid imm for v_wasm_palignr_u8."); |
|
|
return v_wasm_palignr_u8_class<imm>()(a, b); |
|
|
} |
|
|
} |
|
|
|
|
|
template<int imm, typename _Tpvec> |
|
|
inline _Tpvec v_rotate_right(const _Tpvec &a) |
|
|
{ |
|
|
using namespace hal_wasm_internal; |
|
|
enum { imm2 = (imm * sizeof(typename _Tpvec::lane_type)) }; |
|
|
v128_t z = wasm_i8x16_splat(0); |
|
|
return _Tpvec(v_wasm_palignr_u8<imm2>(a.val, z)); |
|
|
} |
|
|
|
|
|
template<int imm, typename _Tpvec> |
|
|
inline _Tpvec v_rotate_left(const _Tpvec &a) |
|
|
{ |
|
|
using namespace hal_wasm_internal; |
|
|
enum { imm2 = ((_Tpvec::nlanes - imm) * sizeof(typename _Tpvec::lane_type)) }; |
|
|
v128_t z = wasm_i8x16_splat(0); |
|
|
return _Tpvec(v_wasm_palignr_u8<imm2>(z, a.val)); |
|
|
} |
|
|
|
|
|
template<int imm, typename _Tpvec> |
|
|
inline _Tpvec v_rotate_right(const _Tpvec &a, const _Tpvec &b) |
|
|
{ |
|
|
using namespace hal_wasm_internal; |
|
|
enum { imm2 = (imm * sizeof(typename _Tpvec::lane_type)) }; |
|
|
return _Tpvec(v_wasm_palignr_u8<imm2>(a.val, b.val)); |
|
|
} |
|
|
|
|
|
template<int imm, typename _Tpvec> |
|
|
inline _Tpvec v_rotate_left(const _Tpvec &a, const _Tpvec &b) |
|
|
{ |
|
|
using namespace hal_wasm_internal; |
|
|
enum { imm2 = ((_Tpvec::nlanes - imm) * sizeof(typename _Tpvec::lane_type)) }; |
|
|
return _Tpvec(v_wasm_palignr_u8<imm2>(b.val, a.val)); |
|
|
} |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(_Tpvec, _Tp) \ |
|
|
inline _Tpvec v_load(const _Tp* ptr) \ |
|
|
{ return _Tpvec(wasm_v128_load(ptr)); } \ |
|
|
inline _Tpvec v_load_aligned(const _Tp* ptr) \ |
|
|
{ return _Tpvec(wasm_v128_load(ptr)); } \ |
|
|
inline _Tpvec v_load_low(const _Tp* ptr) \ |
|
|
{ \ |
|
|
_Tp tmp[_Tpvec::nlanes] = {0}; \ |
|
|
for (int i=0; i<_Tpvec::nlanes/2; ++i) { \ |
|
|
tmp[i] = ptr[i]; \ |
|
|
} \ |
|
|
return _Tpvec(wasm_v128_load(tmp)); \ |
|
|
} \ |
|
|
inline _Tpvec v_load_halves(const _Tp* ptr0, const _Tp* ptr1) \ |
|
|
{ \ |
|
|
_Tp tmp[_Tpvec::nlanes]; \ |
|
|
for (int i=0; i<_Tpvec::nlanes/2; ++i) { \ |
|
|
tmp[i] = ptr0[i]; \ |
|
|
tmp[i+_Tpvec::nlanes/2] = ptr1[i]; \ |
|
|
} \ |
|
|
return _Tpvec(wasm_v128_load(tmp)); \ |
|
|
} \ |
|
|
inline void v_store(_Tp* ptr, const _Tpvec& a) \ |
|
|
{ wasm_v128_store(ptr, a.val); } \ |
|
|
inline void v_store_aligned(_Tp* ptr, const _Tpvec& a) \ |
|
|
{ wasm_v128_store(ptr, a.val); } \ |
|
|
inline void v_store_aligned_nocache(_Tp* ptr, const _Tpvec& a) \ |
|
|
{ wasm_v128_store(ptr, a.val); } \ |
|
|
inline void v_store(_Tp* ptr, const _Tpvec& a, hal::StoreMode ) \ |
|
|
{ \ |
|
|
wasm_v128_store(ptr, a.val); \ |
|
|
} \ |
|
|
inline void v_store_low(_Tp* ptr, const _Tpvec& a) \ |
|
|
{ \ |
|
|
_Tpvec::lane_type a_[_Tpvec::nlanes]; \ |
|
|
wasm_v128_store(a_, a.val); \ |
|
|
for (int i = 0; i < (_Tpvec::nlanes / 2); i++) \ |
|
|
ptr[i] = a_[i]; \ |
|
|
} \ |
|
|
inline void v_store_high(_Tp* ptr, const _Tpvec& a) \ |
|
|
{ \ |
|
|
_Tpvec::lane_type a_[_Tpvec::nlanes]; \ |
|
|
wasm_v128_store(a_, a.val); \ |
|
|
for (int i = 0; i < (_Tpvec::nlanes / 2); i++) \ |
|
|
ptr[i] = a_[i + (_Tpvec::nlanes / 2)]; \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(v_uint8x16, uchar) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(v_int8x16, schar) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(v_uint16x8, ushort) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(v_int16x8, short) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(v_uint32x4, unsigned) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(v_int32x4, int) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(v_uint64x2, uint64) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(v_int64x2, int64) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(v_float32x4, float) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INT_OP(v_float64x2, double) |
|
|
|
|
|
|
|
|
|
|
|
inline v_uint8x16 v_reverse(const v_uint8x16 &a) |
|
|
{ return v_uint8x16(wasm_i8x16_shuffle(a.val, a.val, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0)); } |
|
|
|
|
|
inline v_int8x16 v_reverse(const v_int8x16 &a) |
|
|
{ return v_reinterpret_as_s8(v_reverse(v_reinterpret_as_u8(a))); } |
|
|
|
|
|
inline v_uint16x8 v_reverse(const v_uint16x8 &a) |
|
|
{ return v_uint16x8(wasm_i8x16_shuffle(a.val, a.val, 14, 15, 12, 13, 10, 11, 8, 9, 6, 7, 4, 5, 2, 3, 0, 1)); } |
|
|
|
|
|
inline v_int16x8 v_reverse(const v_int16x8 &a) |
|
|
{ return v_reinterpret_as_s16(v_reverse(v_reinterpret_as_u16(a))); } |
|
|
|
|
|
inline v_uint32x4 v_reverse(const v_uint32x4 &a) |
|
|
{ return v_uint32x4(wasm_i8x16_shuffle(a.val, a.val, 12, 13, 14, 15, 8, 9, 10, 11, 4, 5, 6, 7, 0, 1, 2, 3)); } |
|
|
|
|
|
inline v_int32x4 v_reverse(const v_int32x4 &a) |
|
|
{ return v_reinterpret_as_s32(v_reverse(v_reinterpret_as_u32(a))); } |
|
|
|
|
|
inline v_float32x4 v_reverse(const v_float32x4 &a) |
|
|
{ return v_reinterpret_as_f32(v_reverse(v_reinterpret_as_u32(a))); } |
|
|
|
|
|
inline v_uint64x2 v_reverse(const v_uint64x2 &a) |
|
|
{ return v_uint64x2(wasm_i8x16_shuffle(a.val, a.val, 8, 9, 10, 11, 12, 13, 14, 15, 0, 1, 2, 3, 4, 5, 6, 7)); } |
|
|
|
|
|
inline v_int64x2 v_reverse(const v_int64x2 &a) |
|
|
{ return v_reinterpret_as_s64(v_reverse(v_reinterpret_as_u64(a))); } |
|
|
|
|
|
inline v_float64x2 v_reverse(const v_float64x2 &a) |
|
|
{ return v_reinterpret_as_f64(v_reverse(v_reinterpret_as_u64(a))); } |
|
|
|
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_REDUCE_OP_4_SUM(_Tpvec, scalartype, regtype, suffix, esuffix) \ |
|
|
inline scalartype v_reduce_sum(const _Tpvec& a) \ |
|
|
{ \ |
|
|
regtype val = a.val; \ |
|
|
val = wasm_##suffix##_add(val, wasm_i8x16_shuffle(val, val, 8,9,10,11,12,13,14,15,0,1,2,3,4,5,6,7)); \ |
|
|
val = wasm_##suffix##_add(val, wasm_i8x16_shuffle(val, val, 4,5,6,7,8,9,10,11,12,13,14,15,0,1,2,3)); \ |
|
|
return (scalartype)wasm_##esuffix##_extract_lane(val, 0); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP_4_SUM(v_uint32x4, unsigned, v128_t, i32x4, i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP_4_SUM(v_int32x4, int, v128_t, i32x4, i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP_4_SUM(v_float32x4, float, v128_t, f32x4, f32x4) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#define OPENCV_HAL_IMPL_FALLBACK_REDUCE_OP_SUM(_Tpvec, scalartype) \ |
|
|
inline scalartype v_reduce_sum(const _Tpvec& a) \ |
|
|
{ \ |
|
|
_Tpvec::lane_type a_[_Tpvec::nlanes]; \ |
|
|
wasm_v128_store(a_, a.val); \ |
|
|
scalartype c = a_[0]; \ |
|
|
for (int i = 1; i < _Tpvec::nlanes; i++) \ |
|
|
c += a_[i]; \ |
|
|
return c; \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_FALLBACK_REDUCE_OP_SUM(v_uint8x16, unsigned) |
|
|
OPENCV_HAL_IMPL_FALLBACK_REDUCE_OP_SUM(v_int8x16, int) |
|
|
OPENCV_HAL_IMPL_FALLBACK_REDUCE_OP_SUM(v_uint16x8, unsigned) |
|
|
OPENCV_HAL_IMPL_FALLBACK_REDUCE_OP_SUM(v_int16x8, int) |
|
|
|
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_REDUCE_OP_2_SUM(_Tpvec, scalartype, regtype, suffix, esuffix) \ |
|
|
inline scalartype v_reduce_sum(const _Tpvec& a) \ |
|
|
{ \ |
|
|
regtype val = a.val; \ |
|
|
val = wasm_##suffix##_add(val, wasm_i8x16_shuffle(val, val, 8,9,10,11,12,13,14,15,0,1,2,3,4,5,6,7)); \ |
|
|
return (scalartype)wasm_##esuffix##_extract_lane(val, 0); \ |
|
|
} |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP_2_SUM(v_uint64x2, uint64, v128_t, i64x2, i64x2) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP_2_SUM(v_int64x2, int64, v128_t, i64x2, i64x2) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP_2_SUM(v_float64x2, double, v128_t, f64x2,f64x2) |
|
|
|
|
|
inline v_float32x4 v_reduce_sum4(const v_float32x4& a, const v_float32x4& b, |
|
|
const v_float32x4& c, const v_float32x4& d) |
|
|
{ |
|
|
v128_t ac = wasm_f32x4_add(wasm_unpacklo_i32x4(a.val, c.val), wasm_unpackhi_i32x4(a.val, c.val)); |
|
|
v128_t bd = wasm_f32x4_add(wasm_unpacklo_i32x4(b.val, d.val), wasm_unpackhi_i32x4(b.val, d.val)); |
|
|
return v_float32x4(wasm_f32x4_add(wasm_unpacklo_i32x4(ac, bd), wasm_unpackhi_i32x4(ac, bd))); |
|
|
} |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_REDUCE_OP(_Tpvec, scalartype, func, scalar_func) \ |
|
|
inline scalartype v_reduce_##func(const _Tpvec& a) \ |
|
|
{ \ |
|
|
scalartype buf[_Tpvec::nlanes]; \ |
|
|
v_store(buf, a); \ |
|
|
scalartype tmp = buf[0]; \ |
|
|
for (int i=1; i<_Tpvec::nlanes; ++i) { \ |
|
|
tmp = scalar_func(tmp, buf[i]); \ |
|
|
} \ |
|
|
return tmp; \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_uint8x16, uchar, max, std::max) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_uint8x16, uchar, min, std::min) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_int8x16, schar, max, std::max) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_int8x16, schar, min, std::min) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_uint16x8, ushort, max, std::max) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_uint16x8, ushort, min, std::min) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_int16x8, short, max, std::max) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_int16x8, short, min, std::min) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_uint32x4, unsigned, max, std::max) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_uint32x4, unsigned, min, std::min) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_int32x4, int, max, std::max) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_int32x4, int, min, std::min) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_float32x4, float, max, std::max) |
|
|
OPENCV_HAL_IMPL_WASM_REDUCE_OP(v_float32x4, float, min, std::min) |
|
|
|
|
|
inline unsigned v_reduce_sad(const v_uint8x16& a, const v_uint8x16& b) |
|
|
{ |
|
|
v_uint16x8 l16, h16; |
|
|
v_uint32x4 l16_l32, l16_h32, h16_l32, h16_h32; |
|
|
v_expand(v_absdiff(a, b), l16, h16); |
|
|
v_expand(l16, l16_l32, l16_h32); |
|
|
v_expand(h16, h16_l32, h16_h32); |
|
|
return v_reduce_sum(v_add(v_add(l16_l32, l16_h32), v_add(h16_l32, h16_h32))); |
|
|
} |
|
|
inline unsigned v_reduce_sad(const v_int8x16& a, const v_int8x16& b) |
|
|
{ |
|
|
v_uint16x8 l16, h16; |
|
|
v_uint32x4 l16_l32, l16_h32, h16_l32, h16_h32; |
|
|
v_expand(v_absdiff(a, b), l16, h16); |
|
|
v_expand(l16, l16_l32, l16_h32); |
|
|
v_expand(h16, h16_l32, h16_h32); |
|
|
return v_reduce_sum(v_add(v_add(l16_l32, l16_h32), v_add(h16_l32, h16_h32))); |
|
|
} |
|
|
inline unsigned v_reduce_sad(const v_uint16x8& a, const v_uint16x8& b) |
|
|
{ |
|
|
v_uint32x4 l, h; |
|
|
v_expand(v_absdiff(a, b), l, h); |
|
|
return v_reduce_sum(v_add(l, h)); |
|
|
} |
|
|
inline unsigned v_reduce_sad(const v_int16x8& a, const v_int16x8& b) |
|
|
{ |
|
|
v_uint32x4 l, h; |
|
|
v_expand(v_absdiff(a, b), l, h); |
|
|
return v_reduce_sum(v_add(l, h)); |
|
|
} |
|
|
inline unsigned v_reduce_sad(const v_uint32x4& a, const v_uint32x4& b) |
|
|
{ |
|
|
return v_reduce_sum(v_absdiff(a, b)); |
|
|
} |
|
|
inline unsigned v_reduce_sad(const v_int32x4& a, const v_int32x4& b) |
|
|
{ |
|
|
return v_reduce_sum(v_absdiff(a, b)); |
|
|
} |
|
|
inline float v_reduce_sad(const v_float32x4& a, const v_float32x4& b) |
|
|
{ |
|
|
return v_reduce_sum(v_absdiff(a, b)); |
|
|
} |
|
|
|
|
|
inline v_uint8x16 v_popcount(const v_uint8x16& a) |
|
|
{ |
|
|
v128_t m1 = wasm_i32x4_splat(0x55555555); |
|
|
v128_t m2 = wasm_i32x4_splat(0x33333333); |
|
|
v128_t m4 = wasm_i32x4_splat(0x0f0f0f0f); |
|
|
v128_t p = a.val; |
|
|
p = wasm_i32x4_add(wasm_v128_and(wasm_u32x4_shr(p, 1), m1), wasm_v128_and(p, m1)); |
|
|
p = wasm_i32x4_add(wasm_v128_and(wasm_u32x4_shr(p, 2), m2), wasm_v128_and(p, m2)); |
|
|
p = wasm_i32x4_add(wasm_v128_and(wasm_u32x4_shr(p, 4), m4), wasm_v128_and(p, m4)); |
|
|
return v_uint8x16(p); |
|
|
} |
|
|
inline v_uint16x8 v_popcount(const v_uint16x8& a) |
|
|
{ |
|
|
v_uint8x16 p = v_popcount(v_reinterpret_as_u8(a)); |
|
|
p = v_add(p, v_rotate_right<1>(p)); |
|
|
return v_and(v_reinterpret_as_u16(p), v_setall_u16(0x00ff)); |
|
|
} |
|
|
inline v_uint32x4 v_popcount(const v_uint32x4& a) |
|
|
{ |
|
|
v_uint8x16 p = v_popcount(v_reinterpret_as_u8(a)); |
|
|
p = v_add(p, v_rotate_right<1>(p)); |
|
|
p = v_add(p, v_rotate_right<2>(p)); |
|
|
return v_and(v_reinterpret_as_u32(p), v_setall_u32(0x000000ff)); |
|
|
} |
|
|
inline v_uint64x2 v_popcount(const v_uint64x2& a) |
|
|
{ |
|
|
uint64 a_[2], b_[2] = { 0 }; |
|
|
wasm_v128_store(a_, a.val); |
|
|
for (int i = 0; i < 16; i++) |
|
|
b_[i / 8] += popCountTable[((uint8_t*)a_)[i]]; |
|
|
return v_uint64x2(wasm_v128_load(b_)); |
|
|
} |
|
|
inline v_uint8x16 v_popcount(const v_int8x16& a) |
|
|
{ return v_popcount(v_reinterpret_as_u8(a)); } |
|
|
inline v_uint16x8 v_popcount(const v_int16x8& a) |
|
|
{ return v_popcount(v_reinterpret_as_u16(a)); } |
|
|
inline v_uint32x4 v_popcount(const v_int32x4& a) |
|
|
{ return v_popcount(v_reinterpret_as_u32(a)); } |
|
|
inline v_uint64x2 v_popcount(const v_int64x2& a) |
|
|
{ return v_popcount(v_reinterpret_as_u64(a)); } |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_CHECK_SIGNS(_Tpvec, suffix, scalarType) \ |
|
|
inline int v_signmask(const _Tpvec& a) \ |
|
|
{ \ |
|
|
_Tpvec::lane_type a_[_Tpvec::nlanes]; \ |
|
|
wasm_v128_store(a_, a.val); \ |
|
|
int mask = 0; \ |
|
|
for (int i = 0; i < _Tpvec::nlanes; i++) \ |
|
|
mask |= (reinterpret_int(a_[i]) < 0) << i; \ |
|
|
return mask; \ |
|
|
} \ |
|
|
inline bool v_check_all(const _Tpvec& a) \ |
|
|
{ return wasm_i8x16_all_true(wasm_##suffix##_lt(a.val, wasm_##suffix##_splat(0))); } \ |
|
|
inline bool v_check_any(const _Tpvec& a) \ |
|
|
{ return wasm_i8x16_any_true(wasm_##suffix##_lt(a.val, wasm_##suffix##_splat(0)));; } |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_CHECK_SIGNS(v_uint8x16, i8x16, schar) |
|
|
OPENCV_HAL_IMPL_WASM_CHECK_SIGNS(v_int8x16, i8x16, schar) |
|
|
OPENCV_HAL_IMPL_WASM_CHECK_SIGNS(v_uint16x8, i16x8, short) |
|
|
OPENCV_HAL_IMPL_WASM_CHECK_SIGNS(v_int16x8, i16x8, short) |
|
|
OPENCV_HAL_IMPL_WASM_CHECK_SIGNS(v_uint32x4, i32x4, int) |
|
|
OPENCV_HAL_IMPL_WASM_CHECK_SIGNS(v_int32x4, i32x4, int) |
|
|
OPENCV_HAL_IMPL_WASM_CHECK_SIGNS(v_float32x4, i32x4, float) |
|
|
OPENCV_HAL_IMPL_WASM_CHECK_SIGNS(v_float64x2, f64x2, double) |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_CHECK_ALL_ANY(_Tpvec, suffix, esuffix) \ |
|
|
inline bool v_check_all(const _Tpvec& a) \ |
|
|
{ \ |
|
|
v128_t masked = v_reinterpret_as_##esuffix(a).val; \ |
|
|
masked = wasm_i32x4_replace_lane(masked, 0, 0xffffffff); \ |
|
|
masked = wasm_i32x4_replace_lane(masked, 2, 0xffffffff); \ |
|
|
return wasm_i8x16_all_true(wasm_##suffix##_lt(masked, wasm_##suffix##_splat(0))); \ |
|
|
} \ |
|
|
inline bool v_check_any(const _Tpvec& a) \ |
|
|
{ \ |
|
|
v128_t masked = v_reinterpret_as_##esuffix(a).val; \ |
|
|
masked = wasm_i32x4_replace_lane(masked, 0, 0x0); \ |
|
|
masked = wasm_i32x4_replace_lane(masked, 2, 0x0); \ |
|
|
return wasm_i8x16_any_true(wasm_##suffix##_lt(masked, wasm_##suffix##_splat(0))); \ |
|
|
} \ |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_CHECK_ALL_ANY(v_int64x2, i32x4, s32) |
|
|
OPENCV_HAL_IMPL_WASM_CHECK_ALL_ANY(v_uint64x2, i32x4, u32) |
|
|
|
|
|
|
|
|
inline int v_scan_forward(const v_int8x16& a) { return trailingZeros32(v_signmask(v_reinterpret_as_s8(a))); } |
|
|
inline int v_scan_forward(const v_uint8x16& a) { return trailingZeros32(v_signmask(v_reinterpret_as_s8(a))); } |
|
|
inline int v_scan_forward(const v_int16x8& a) { return trailingZeros32(v_signmask(v_reinterpret_as_s8(a))) / 2; } |
|
|
inline int v_scan_forward(const v_uint16x8& a) { return trailingZeros32(v_signmask(v_reinterpret_as_s8(a))) / 2; } |
|
|
inline int v_scan_forward(const v_int32x4& a) { return trailingZeros32(v_signmask(v_reinterpret_as_s8(a))) / 4; } |
|
|
inline int v_scan_forward(const v_uint32x4& a) { return trailingZeros32(v_signmask(v_reinterpret_as_s8(a))) / 4; } |
|
|
inline int v_scan_forward(const v_float32x4& a) { return trailingZeros32(v_signmask(v_reinterpret_as_s8(a))) / 4; } |
|
|
inline int v_scan_forward(const v_int64x2& a) { return trailingZeros32(v_signmask(v_reinterpret_as_s8(a))) / 8; } |
|
|
inline int v_scan_forward(const v_uint64x2& a) { return trailingZeros32(v_signmask(v_reinterpret_as_s8(a))) / 8; } |
|
|
inline int v_scan_forward(const v_float64x2& a) { return trailingZeros32(v_signmask(v_reinterpret_as_s8(a))) / 8; } |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_SELECT(_Tpvec) \ |
|
|
inline _Tpvec v_select(const _Tpvec& mask, const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
return _Tpvec(wasm_v128_bitselect(a.val, b.val, mask.val)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_SELECT(v_uint8x16) |
|
|
OPENCV_HAL_IMPL_WASM_SELECT(v_int8x16) |
|
|
OPENCV_HAL_IMPL_WASM_SELECT(v_uint16x8) |
|
|
OPENCV_HAL_IMPL_WASM_SELECT(v_int16x8) |
|
|
OPENCV_HAL_IMPL_WASM_SELECT(v_uint32x4) |
|
|
OPENCV_HAL_IMPL_WASM_SELECT(v_int32x4) |
|
|
OPENCV_HAL_IMPL_WASM_SELECT(v_uint64x2) |
|
|
OPENCV_HAL_IMPL_WASM_SELECT(v_int64x2) |
|
|
OPENCV_HAL_IMPL_WASM_SELECT(v_float32x4) |
|
|
OPENCV_HAL_IMPL_WASM_SELECT(v_float64x2) |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_EXPAND(_Tpvec, _Tpwvec, _Tp, intrin) \ |
|
|
inline void v_expand(const _Tpvec& a, _Tpwvec& b0, _Tpwvec& b1) \ |
|
|
{ \ |
|
|
b0.val = intrin(a.val); \ |
|
|
b1.val = __CV_CAT(intrin, _high)(a.val); \ |
|
|
} \ |
|
|
inline _Tpwvec v_expand_low(const _Tpvec& a) \ |
|
|
{ return _Tpwvec(intrin(a.val)); } \ |
|
|
inline _Tpwvec v_expand_high(const _Tpvec& a) \ |
|
|
{ return _Tpwvec(__CV_CAT(intrin, _high)(a.val)); } \ |
|
|
inline _Tpwvec v_load_expand(const _Tp* ptr) \ |
|
|
{ \ |
|
|
v128_t a = wasm_v128_load(ptr); \ |
|
|
return _Tpwvec(intrin(a)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_EXPAND(v_uint8x16, v_uint16x8, uchar, v128_cvtu8x16_i16x8) |
|
|
OPENCV_HAL_IMPL_WASM_EXPAND(v_int8x16, v_int16x8, schar, v128_cvti8x16_i16x8) |
|
|
OPENCV_HAL_IMPL_WASM_EXPAND(v_uint16x8, v_uint32x4, ushort, v128_cvtu16x8_i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_EXPAND(v_int16x8, v_int32x4, short, v128_cvti16x8_i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_EXPAND(v_uint32x4, v_uint64x2, unsigned, v128_cvtu32x4_i64x2) |
|
|
OPENCV_HAL_IMPL_WASM_EXPAND(v_int32x4, v_int64x2, int, v128_cvti32x4_i64x2) |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_EXPAND_Q(_Tpvec, _Tp, intrin) \ |
|
|
inline _Tpvec v_load_expand_q(const _Tp* ptr) \ |
|
|
{ \ |
|
|
v128_t a = wasm_v128_load(ptr); \ |
|
|
return _Tpvec(intrin(a)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_EXPAND_Q(v_uint32x4, uchar, v128_cvtu8x16_i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_EXPAND_Q(v_int32x4, schar, v128_cvti8x16_i32x4) |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_UNPACKS(_Tpvec, suffix) \ |
|
|
inline void v_zip(const _Tpvec& a0, const _Tpvec& a1, _Tpvec& b0, _Tpvec& b1) \ |
|
|
{ \ |
|
|
b0.val = wasm_unpacklo_##suffix(a0.val, a1.val); \ |
|
|
b1.val = wasm_unpackhi_##suffix(a0.val, a1.val); \ |
|
|
} \ |
|
|
inline _Tpvec v_combine_low(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
return _Tpvec(wasm_unpacklo_i64x2(a.val, b.val)); \ |
|
|
} \ |
|
|
inline _Tpvec v_combine_high(const _Tpvec& a, const _Tpvec& b) \ |
|
|
{ \ |
|
|
return _Tpvec(wasm_unpackhi_i64x2(a.val, b.val)); \ |
|
|
} \ |
|
|
inline void v_recombine(const _Tpvec& a, const _Tpvec& b, _Tpvec& c, _Tpvec& d) \ |
|
|
{ \ |
|
|
c.val = wasm_unpacklo_i64x2(a.val, b.val); \ |
|
|
d.val = wasm_unpackhi_i64x2(a.val, b.val); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_UNPACKS(v_uint8x16, i8x16) |
|
|
OPENCV_HAL_IMPL_WASM_UNPACKS(v_int8x16, i8x16) |
|
|
OPENCV_HAL_IMPL_WASM_UNPACKS(v_uint16x8, i16x8) |
|
|
OPENCV_HAL_IMPL_WASM_UNPACKS(v_int16x8, i16x8) |
|
|
OPENCV_HAL_IMPL_WASM_UNPACKS(v_uint32x4, i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_UNPACKS(v_int32x4, i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_UNPACKS(v_float32x4, i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_UNPACKS(v_float64x2, i64x2) |
|
|
|
|
|
template<int s, typename _Tpvec> |
|
|
inline _Tpvec v_extract(const _Tpvec& a, const _Tpvec& b) |
|
|
{ |
|
|
return v_rotate_right<s>(a, b); |
|
|
} |
|
|
|
|
|
inline v_int32x4 v_round(const v_float32x4& a) |
|
|
{ |
|
|
v128_t h = wasm_f32x4_splat(0.5); |
|
|
return v_int32x4(wasm_i32x4_trunc_saturate_f32x4(wasm_f32x4_add(a.val, h))); |
|
|
} |
|
|
|
|
|
inline v_int32x4 v_floor(const v_float32x4& a) |
|
|
{ |
|
|
v128_t a1 = wasm_i32x4_trunc_saturate_f32x4(a.val); |
|
|
v128_t mask = wasm_f32x4_lt(a.val, wasm_f32x4_convert_i32x4(a1)); |
|
|
return v_int32x4(wasm_i32x4_add(a1, mask)); |
|
|
} |
|
|
|
|
|
inline v_int32x4 v_ceil(const v_float32x4& a) |
|
|
{ |
|
|
v128_t a1 = wasm_i32x4_trunc_saturate_f32x4(a.val); |
|
|
v128_t mask = wasm_f32x4_gt(a.val, wasm_f32x4_convert_i32x4(a1)); |
|
|
return v_int32x4(wasm_i32x4_sub(a1, mask)); |
|
|
} |
|
|
|
|
|
inline v_int32x4 v_trunc(const v_float32x4& a) |
|
|
{ return v_int32x4(wasm_i32x4_trunc_saturate_f32x4(a.val)); } |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_MATH_FUNC(func, cfunc) \ |
|
|
inline v_int32x4 func(const v_float64x2& a) \ |
|
|
{ \ |
|
|
double a_[2]; \ |
|
|
wasm_v128_store(a_, a.val); \ |
|
|
int c_[4]; \ |
|
|
c_[0] = cfunc(a_[0]); \ |
|
|
c_[1] = cfunc(a_[1]); \ |
|
|
c_[2] = 0; \ |
|
|
c_[3] = 0; \ |
|
|
return v_int32x4(wasm_v128_load(c_)); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_MATH_FUNC(v_round, cvRound) |
|
|
OPENCV_HAL_IMPL_WASM_MATH_FUNC(v_floor, cvFloor) |
|
|
OPENCV_HAL_IMPL_WASM_MATH_FUNC(v_ceil, cvCeil) |
|
|
OPENCV_HAL_IMPL_WASM_MATH_FUNC(v_trunc, int) |
|
|
|
|
|
inline v_int32x4 v_round(const v_float64x2& a, const v_float64x2& b) |
|
|
{ |
|
|
double a_[2], b_[2]; |
|
|
wasm_v128_store(a_, a.val); |
|
|
wasm_v128_store(b_, b.val); |
|
|
int c_[4]; |
|
|
c_[0] = cvRound(a_[0]); |
|
|
c_[1] = cvRound(a_[1]); |
|
|
c_[2] = cvRound(b_[0]); |
|
|
c_[3] = cvRound(b_[1]); |
|
|
return v_int32x4(wasm_v128_load(c_)); |
|
|
} |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_TRANSPOSE4x4(_Tpvec, suffix) \ |
|
|
inline void v_transpose4x4(const _Tpvec& a0, const _Tpvec& a1, \ |
|
|
const _Tpvec& a2, const _Tpvec& a3, \ |
|
|
_Tpvec& b0, _Tpvec& b1, \ |
|
|
_Tpvec& b2, _Tpvec& b3) \ |
|
|
{ \ |
|
|
v128_t t0 = wasm_unpacklo_##suffix(a0.val, a1.val); \ |
|
|
v128_t t1 = wasm_unpacklo_##suffix(a2.val, a3.val); \ |
|
|
v128_t t2 = wasm_unpackhi_##suffix(a0.val, a1.val); \ |
|
|
v128_t t3 = wasm_unpackhi_##suffix(a2.val, a3.val); \ |
|
|
\ |
|
|
b0.val = wasm_unpacklo_i64x2(t0, t1); \ |
|
|
b1.val = wasm_unpackhi_i64x2(t0, t1); \ |
|
|
b2.val = wasm_unpacklo_i64x2(t2, t3); \ |
|
|
b3.val = wasm_unpackhi_i64x2(t2, t3); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_TRANSPOSE4x4(v_uint32x4, i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_TRANSPOSE4x4(v_int32x4, i32x4) |
|
|
OPENCV_HAL_IMPL_WASM_TRANSPOSE4x4(v_float32x4, i32x4) |
|
|
|
|
|
|
|
|
inline void v_load_deinterleave(const uchar* ptr, v_uint8x16& a, v_uint8x16& b) |
|
|
{ |
|
|
v128_t t00 = wasm_v128_load(ptr); |
|
|
v128_t t01 = wasm_v128_load(ptr + 16); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(t00, t01, 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30); |
|
|
b.val = wasm_i8x16_shuffle(t00, t01, 1,3,5,7,9,11,13,15,17,19,21,23,25,27,29,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const uchar* ptr, v_uint8x16& a, v_uint8x16& b, v_uint8x16& c) |
|
|
{ |
|
|
v128_t t00 = wasm_v128_load(ptr); |
|
|
v128_t t01 = wasm_v128_load(ptr + 16); |
|
|
v128_t t02 = wasm_v128_load(ptr + 32); |
|
|
|
|
|
v128_t t10 = wasm_i8x16_shuffle(t00, t01, 0,3,6,9,12,15,18,21,24,27,30,1,2,4,5,7); |
|
|
v128_t t11 = wasm_i8x16_shuffle(t00, t01, 1,4,7,10,13,16,19,22,25,28,31,0,2,3,5,6); |
|
|
v128_t t12 = wasm_i8x16_shuffle(t00, t01, 2,5,8,11,14,17,20,23,26,29,0,1,3,4,6,7); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(t10, t02, 0,1,2,3,4,5,6,7,8,9,10,17,20,23,26,29); |
|
|
b.val = wasm_i8x16_shuffle(t11, t02, 0,1,2,3,4,5,6,7,8,9,10,18,21,24,27,30); |
|
|
c.val = wasm_i8x16_shuffle(t12, t02, 0,1,2,3,4,5,6,7,8,9,16,19,22,25,28,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const uchar* ptr, v_uint8x16& a, v_uint8x16& b, v_uint8x16& c, v_uint8x16& d) |
|
|
{ |
|
|
v128_t u0 = wasm_v128_load(ptr); |
|
|
v128_t u1 = wasm_v128_load(ptr + 16); |
|
|
v128_t u2 = wasm_v128_load(ptr + 32); |
|
|
v128_t u3 = wasm_v128_load(ptr + 48); |
|
|
|
|
|
v128_t v0 = wasm_i8x16_shuffle(u0, u1, 0,4,8,12,16,20,24,28,1,5,9,13,17,21,25,29); |
|
|
v128_t v1 = wasm_i8x16_shuffle(u2, u3, 0,4,8,12,16,20,24,28,1,5,9,13,17,21,25,29); |
|
|
v128_t v2 = wasm_i8x16_shuffle(u0, u1, 2,6,10,14,18,22,26,30,3,7,11,15,19,23,27,31); |
|
|
v128_t v3 = wasm_i8x16_shuffle(u2, u3, 2,6,10,14,18,22,26,30,3,7,11,15,19,23,27,31); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(v0, v1, 0,1,2,3,4,5,6,7,16,17,18,19,20,21,22,23); |
|
|
b.val = wasm_i8x16_shuffle(v0, v1, 8,9,10,11,12,13,14,15,24,25,26,27,28,29,30,31); |
|
|
c.val = wasm_i8x16_shuffle(v2, v3, 0,1,2,3,4,5,6,7,16,17,18,19,20,21,22,23); |
|
|
d.val = wasm_i8x16_shuffle(v2, v3, 8,9,10,11,12,13,14,15,24,25,26,27,28,29,30,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const ushort* ptr, v_uint16x8& a, v_uint16x8& b) |
|
|
{ |
|
|
v128_t v0 = wasm_v128_load(ptr); |
|
|
v128_t v1 = wasm_v128_load(ptr + 8); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(v0, v1, 0,1,4,5,8,9,12,13,16,17,20,21,24,25,28,29); |
|
|
b.val = wasm_i8x16_shuffle(v0, v1, 2,3,6,7,10,11,14,15,18,19,22,23,26,27,30,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const ushort* ptr, v_uint16x8& a, v_uint16x8& b, v_uint16x8& c) |
|
|
{ |
|
|
v128_t t00 = wasm_v128_load(ptr); |
|
|
v128_t t01 = wasm_v128_load(ptr + 8); |
|
|
v128_t t02 = wasm_v128_load(ptr + 16); |
|
|
|
|
|
v128_t t10 = wasm_i8x16_shuffle(t00, t01, 0,1,6,7,12,13,18,19,24,25,30,31,2,3,4,5); |
|
|
v128_t t11 = wasm_i8x16_shuffle(t00, t01, 2,3,8,9,14,15,20,21,26,27,0,1,4,5,6,7); |
|
|
v128_t t12 = wasm_i8x16_shuffle(t00, t01, 4,5,10,11,16,17,22,23,28,29,0,1,2,3,6,7); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(t10, t02, 0,1,2,3,4,5,6,7,8,9,10,11,20,21,26,27); |
|
|
b.val = wasm_i8x16_shuffle(t11, t02, 0,1,2,3,4,5,6,7,8,9,16,17,22,23,28,29); |
|
|
c.val = wasm_i8x16_shuffle(t12, t02, 0,1,2,3,4,5,6,7,8,9,18,19,24,25,30,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const ushort* ptr, v_uint16x8& a, v_uint16x8& b, v_uint16x8& c, v_uint16x8& d) |
|
|
{ |
|
|
v128_t u0 = wasm_v128_load(ptr); |
|
|
v128_t u1 = wasm_v128_load(ptr + 8); |
|
|
v128_t u2 = wasm_v128_load(ptr + 16); |
|
|
v128_t u3 = wasm_v128_load(ptr + 24); |
|
|
|
|
|
v128_t v0 = wasm_i8x16_shuffle(u0, u1, 0,1,8,9,16,17,24,25,2,3,10,11,18,19,26,27); |
|
|
v128_t v1 = wasm_i8x16_shuffle(u2, u3, 0,1,8,9,16,17,24,25,2,3,10,11,18,19,26,27); |
|
|
v128_t v2 = wasm_i8x16_shuffle(u0, u1, 4,5,12,13,20,21,28,29,6,7,14,15,22,23,30,31); |
|
|
v128_t v3 = wasm_i8x16_shuffle(u2, u3, 4,5,12,13,20,21,28,29,6,7,14,15,22,23,30,31); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(v0, v1, 0,1,2,3,4,5,6,7,16,17,18,19,20,21,22,23); |
|
|
b.val = wasm_i8x16_shuffle(v0, v1, 8,9,10,11,12,13,14,15,24,25,26,27,28,29,30,31); |
|
|
c.val = wasm_i8x16_shuffle(v2, v3, 0,1,2,3,4,5,6,7,16,17,18,19,20,21,22,23); |
|
|
d.val = wasm_i8x16_shuffle(v2, v3, 8,9,10,11,12,13,14,15,24,25,26,27,28,29,30,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const unsigned* ptr, v_uint32x4& a, v_uint32x4& b) |
|
|
{ |
|
|
v128_t v0 = wasm_v128_load(ptr); |
|
|
v128_t v1 = wasm_v128_load(ptr + 4); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(v0, v1, 0,1,2,3,8,9,10,11,16,17,18,19,24,25,26,27); |
|
|
b.val = wasm_i8x16_shuffle(v0, v1, 4,5,6,7,12,13,14,15,20,21,22,23,28,29,30,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const unsigned* ptr, v_uint32x4& a, v_uint32x4& b, v_uint32x4& c) |
|
|
{ |
|
|
v128_t t00 = wasm_v128_load(ptr); |
|
|
v128_t t01 = wasm_v128_load(ptr + 4); |
|
|
v128_t t02 = wasm_v128_load(ptr + 8); |
|
|
|
|
|
v128_t t10 = wasm_i8x16_shuffle(t00, t01, 0,1,2,3,12,13,14,15,24,25,26,27,4,5,6,7); |
|
|
v128_t t11 = wasm_i8x16_shuffle(t00, t01, 4,5,6,7,16,17,18,19,28,29,30,31,0,1,2,3); |
|
|
v128_t t12 = wasm_i8x16_shuffle(t00, t01, 8,9,10,11,20,21,22,23,0,1,2,3,4,5,6,7); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(t10, t02, 0,1,2,3,4,5,6,7,8,9,10,11,20,21,22,23); |
|
|
b.val = wasm_i8x16_shuffle(t11, t02, 0,1,2,3,4,5,6,7,8,9,10,11,24,25,26,27); |
|
|
c.val = wasm_i8x16_shuffle(t12, t02, 0,1,2,3,4,5,6,7,16,17,18,19,28,29,30,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const unsigned* ptr, v_uint32x4& a, v_uint32x4& b, v_uint32x4& c, v_uint32x4& d) |
|
|
{ |
|
|
v_uint32x4 s0(wasm_v128_load(ptr)); |
|
|
v_uint32x4 s1(wasm_v128_load(ptr + 4)); |
|
|
v_uint32x4 s2(wasm_v128_load(ptr + 8)); |
|
|
v_uint32x4 s3(wasm_v128_load(ptr + 12)); |
|
|
|
|
|
v_transpose4x4(s0, s1, s2, s3, a, b, c, d); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const float* ptr, v_float32x4& a, v_float32x4& b) |
|
|
{ |
|
|
v128_t v0 = wasm_v128_load(ptr); |
|
|
v128_t v1 = wasm_v128_load((ptr + 4)); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(v0, v1, 0,1,2,3,8,9,10,11,16,17,18,19,24,25,26,27); |
|
|
b.val = wasm_i8x16_shuffle(v0, v1, 4,5,6,7,12,13,14,15,20,21,22,23,28,29,30,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const float* ptr, v_float32x4& a, v_float32x4& b, v_float32x4& c) |
|
|
{ |
|
|
v128_t t00 = wasm_v128_load(ptr); |
|
|
v128_t t01 = wasm_v128_load(ptr + 4); |
|
|
v128_t t02 = wasm_v128_load(ptr + 8); |
|
|
|
|
|
v128_t t10 = wasm_i8x16_shuffle(t00, t01, 0,1,2,3,12,13,14,15,24,25,26,27,4,5,6,7); |
|
|
v128_t t11 = wasm_i8x16_shuffle(t00, t01, 4,5,6,7,16,17,18,19,28,29,30,31,0,1,2,3); |
|
|
v128_t t12 = wasm_i8x16_shuffle(t00, t01, 8,9,10,11,20,21,22,23,0,1,2,3,4,5,6,7); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(t10, t02, 0,1,2,3,4,5,6,7,8,9,10,11,20,21,22,23); |
|
|
b.val = wasm_i8x16_shuffle(t11, t02, 0,1,2,3,4,5,6,7,8,9,10,11,24,25,26,27); |
|
|
c.val = wasm_i8x16_shuffle(t12, t02, 0,1,2,3,4,5,6,7,16,17,18,19,28,29,30,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const float* ptr, v_float32x4& a, v_float32x4& b, v_float32x4& c, v_float32x4& d) |
|
|
{ |
|
|
v_float32x4 s0(wasm_v128_load(ptr)); |
|
|
v_float32x4 s1(wasm_v128_load(ptr + 4)); |
|
|
v_float32x4 s2(wasm_v128_load(ptr + 8)); |
|
|
v_float32x4 s3(wasm_v128_load(ptr + 12)); |
|
|
|
|
|
v_transpose4x4(s0, s1, s2, s3, a, b, c, d); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const uint64 *ptr, v_uint64x2& a, v_uint64x2& b) |
|
|
{ |
|
|
v128_t t0 = wasm_v128_load(ptr); |
|
|
v128_t t1 = wasm_v128_load(ptr + 2); |
|
|
|
|
|
a.val = wasm_unpacklo_i64x2(t0, t1); |
|
|
b.val = wasm_unpackhi_i64x2(t0, t1); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const uint64 *ptr, v_uint64x2& a, v_uint64x2& b, v_uint64x2& c) |
|
|
{ |
|
|
v128_t t0 = wasm_v128_load(ptr); |
|
|
v128_t t1 = wasm_v128_load(ptr + 2); |
|
|
v128_t t2 = wasm_v128_load(ptr + 4); |
|
|
|
|
|
a.val = wasm_i8x16_shuffle(t0, t1, 0,1,2,3,4,5,6,7,24,25,26,27,28,29,30,31); |
|
|
b.val = wasm_i8x16_shuffle(t0, t2, 8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23); |
|
|
c.val = wasm_i8x16_shuffle(t1, t2, 0,1,2,3,4,5,6,7,24,25,26,27,28,29,30,31); |
|
|
} |
|
|
|
|
|
inline void v_load_deinterleave(const uint64 *ptr, v_uint64x2& a, |
|
|
v_uint64x2& b, v_uint64x2& c, v_uint64x2& d) |
|
|
{ |
|
|
v128_t t0 = wasm_v128_load(ptr); |
|
|
v128_t t1 = wasm_v128_load(ptr + 2); |
|
|
v128_t t2 = wasm_v128_load(ptr + 4); |
|
|
v128_t t3 = wasm_v128_load(ptr + 6); |
|
|
|
|
|
a.val = wasm_unpacklo_i64x2(t0, t2); |
|
|
b.val = wasm_unpackhi_i64x2(t0, t2); |
|
|
c.val = wasm_unpacklo_i64x2(t1, t3); |
|
|
d.val = wasm_unpackhi_i64x2(t1, t3); |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
inline void v_store_interleave( uchar* ptr, const v_uint8x16& a, const v_uint8x16& b, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t v0 = wasm_unpacklo_i8x16(a.val, b.val); |
|
|
v128_t v1 = wasm_unpackhi_i8x16(a.val, b.val); |
|
|
|
|
|
wasm_v128_store(ptr, v0); |
|
|
wasm_v128_store(ptr + 16, v1); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave( uchar* ptr, const v_uint8x16& a, const v_uint8x16& b, |
|
|
const v_uint8x16& c, hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t t00 = wasm_i8x16_shuffle(a.val, b.val, 0,16,0,1,17,0,2,18,0,3,19,0,4,20,0,5); |
|
|
v128_t t01 = wasm_i8x16_shuffle(a.val, b.val, 21,0,6,22,0,7,23,0,8,24,0,9,25,0,10,26); |
|
|
v128_t t02 = wasm_i8x16_shuffle(a.val, b.val, 0,11,27,0,12,28,0,13,29,0,14,30,0,15,31,0); |
|
|
|
|
|
v128_t t10 = wasm_i8x16_shuffle(t00, c.val, 0,1,16,3,4,17,6,7,18,9,10,19,12,13,20,15); |
|
|
v128_t t11 = wasm_i8x16_shuffle(t01, c.val, 0,21,2,3,22,5,6,23,8,9,24,11,12,25,14,15); |
|
|
v128_t t12 = wasm_i8x16_shuffle(t02, c.val, 26,1,2,27,4,5,28,7,8,29,10,11,30,13,14,31); |
|
|
|
|
|
wasm_v128_store(ptr, t10); |
|
|
wasm_v128_store(ptr + 16, t11); |
|
|
wasm_v128_store(ptr + 32, t12); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave( uchar* ptr, const v_uint8x16& a, const v_uint8x16& b, |
|
|
const v_uint8x16& c, const v_uint8x16& d, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
v128_t u0 = wasm_unpacklo_i8x16(a.val, c.val); |
|
|
v128_t u1 = wasm_unpackhi_i8x16(a.val, c.val); |
|
|
v128_t u2 = wasm_unpacklo_i8x16(b.val, d.val); |
|
|
v128_t u3 = wasm_unpackhi_i8x16(b.val, d.val); |
|
|
|
|
|
v128_t v0 = wasm_unpacklo_i8x16(u0, u2); |
|
|
v128_t v1 = wasm_unpackhi_i8x16(u0, u2); |
|
|
v128_t v2 = wasm_unpacklo_i8x16(u1, u3); |
|
|
v128_t v3 = wasm_unpackhi_i8x16(u1, u3); |
|
|
|
|
|
wasm_v128_store(ptr, v0); |
|
|
wasm_v128_store(ptr + 16, v1); |
|
|
wasm_v128_store(ptr + 32, v2); |
|
|
wasm_v128_store(ptr + 48, v3); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave( ushort* ptr, const v_uint16x8& a, const v_uint16x8& b, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t v0 = wasm_unpacklo_i16x8(a.val, b.val); |
|
|
v128_t v1 = wasm_unpackhi_i16x8(a.val, b.val); |
|
|
|
|
|
wasm_v128_store(ptr, v0); |
|
|
wasm_v128_store(ptr + 8, v1); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave( ushort* ptr, const v_uint16x8& a, |
|
|
const v_uint16x8& b, const v_uint16x8& c, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t t00 = wasm_i8x16_shuffle(a.val, b.val, 0,1,16,17,0,0,2,3,18,19,0,0,4,5,20,21); |
|
|
v128_t t01 = wasm_i8x16_shuffle(a.val, b.val, 0,0,6,7,22,23,0,0,8,9,24,25,0,0,10,11); |
|
|
v128_t t02 = wasm_i8x16_shuffle(a.val, b.val, 26,27,0,0,12,13,28,29,0,0,14,15,30,31,0,0); |
|
|
|
|
|
v128_t t10 = wasm_i8x16_shuffle(t00, c.val, 0,1,2,3,16,17,6,7,8,9,18,19,12,13,14,15); |
|
|
v128_t t11 = wasm_i8x16_shuffle(t01, c.val, 20,21,2,3,4,5,22,23,8,9,10,11,24,25,14,15); |
|
|
v128_t t12 = wasm_i8x16_shuffle(t02, c.val, 0,1,26,27,4,5,6,7,28,29,10,11,12,13,30,31); |
|
|
|
|
|
wasm_v128_store(ptr, t10); |
|
|
wasm_v128_store(ptr + 8, t11); |
|
|
wasm_v128_store(ptr + 16, t12); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave( ushort* ptr, const v_uint16x8& a, const v_uint16x8& b, |
|
|
const v_uint16x8& c, const v_uint16x8& d, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
v128_t u0 = wasm_unpacklo_i16x8(a.val, c.val); |
|
|
v128_t u1 = wasm_unpackhi_i16x8(a.val, c.val); |
|
|
v128_t u2 = wasm_unpacklo_i16x8(b.val, d.val); |
|
|
v128_t u3 = wasm_unpackhi_i16x8(b.val, d.val); |
|
|
|
|
|
v128_t v0 = wasm_unpacklo_i16x8(u0, u2); |
|
|
v128_t v1 = wasm_unpackhi_i16x8(u0, u2); |
|
|
v128_t v2 = wasm_unpacklo_i16x8(u1, u3); |
|
|
v128_t v3 = wasm_unpackhi_i16x8(u1, u3); |
|
|
|
|
|
wasm_v128_store(ptr, v0); |
|
|
wasm_v128_store(ptr + 8, v1); |
|
|
wasm_v128_store(ptr + 16, v2); |
|
|
wasm_v128_store(ptr + 24, v3); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave( unsigned* ptr, const v_uint32x4& a, const v_uint32x4& b, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t v0 = wasm_unpacklo_i32x4(a.val, b.val); |
|
|
v128_t v1 = wasm_unpackhi_i32x4(a.val, b.val); |
|
|
|
|
|
wasm_v128_store(ptr, v0); |
|
|
wasm_v128_store(ptr + 4, v1); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave( unsigned* ptr, const v_uint32x4& a, const v_uint32x4& b, |
|
|
const v_uint32x4& c, hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t t00 = wasm_i8x16_shuffle(a.val, b.val, 0,1,2,3,16,17,18,19,0,0,0,0,4,5,6,7); |
|
|
v128_t t01 = wasm_i8x16_shuffle(a.val, b.val, 20,21,22,23,0,0,0,0,8,9,10,11,24,25,26,27); |
|
|
v128_t t02 = wasm_i8x16_shuffle(a.val, b.val, 0,0,0,0,12,13,14,15,28,29,30,31,0,0,0,0); |
|
|
|
|
|
v128_t t10 = wasm_i8x16_shuffle(t00, c.val, 0,1,2,3,4,5,6,7,16,17,18,19,12,13,14,15); |
|
|
v128_t t11 = wasm_i8x16_shuffle(t01, c.val, 0,1,2,3,20,21,22,23,8,9,10,11,12,13,14,15); |
|
|
v128_t t12 = wasm_i8x16_shuffle(t02, c.val, 24,25,26,27,4,5,6,7,8,9,10,11,28,29,30,31); |
|
|
|
|
|
wasm_v128_store(ptr, t10); |
|
|
wasm_v128_store(ptr + 4, t11); |
|
|
wasm_v128_store(ptr + 8, t12); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave(unsigned* ptr, const v_uint32x4& a, const v_uint32x4& b, |
|
|
const v_uint32x4& c, const v_uint32x4& d, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v_uint32x4 v0, v1, v2, v3; |
|
|
v_transpose4x4(a, b, c, d, v0, v1, v2, v3); |
|
|
|
|
|
wasm_v128_store(ptr, v0.val); |
|
|
wasm_v128_store(ptr + 4, v1.val); |
|
|
wasm_v128_store(ptr + 8, v2.val); |
|
|
wasm_v128_store(ptr + 12, v3.val); |
|
|
} |
|
|
|
|
|
|
|
|
inline void v_store_interleave(float* ptr, const v_float32x4& a, const v_float32x4& b, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t v0 = wasm_unpacklo_i32x4(a.val, b.val); |
|
|
v128_t v1 = wasm_unpackhi_i32x4(a.val, b.val); |
|
|
|
|
|
wasm_v128_store(ptr, v0); |
|
|
wasm_v128_store(ptr + 4, v1); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave(float* ptr, const v_float32x4& a, const v_float32x4& b, |
|
|
const v_float32x4& c, hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t t00 = wasm_i8x16_shuffle(a.val, b.val, 0,1,2,3,16,17,18,19,0,0,0,0,4,5,6,7); |
|
|
v128_t t01 = wasm_i8x16_shuffle(a.val, b.val, 20,21,22,23,0,0,0,0,8,9,10,11,24,25,26,27); |
|
|
v128_t t02 = wasm_i8x16_shuffle(a.val, b.val, 0,0,0,0,12,13,14,15,28,29,30,31,0,0,0,0); |
|
|
|
|
|
v128_t t10 = wasm_i8x16_shuffle(t00, c.val, 0,1,2,3,4,5,6,7,16,17,18,19,12,13,14,15); |
|
|
v128_t t11 = wasm_i8x16_shuffle(t01, c.val, 0,1,2,3,20,21,22,23,8,9,10,11,12,13,14,15); |
|
|
v128_t t12 = wasm_i8x16_shuffle(t02, c.val, 24,25,26,27,4,5,6,7,8,9,10,11,28,29,30,31); |
|
|
|
|
|
wasm_v128_store(ptr, t10); |
|
|
wasm_v128_store(ptr + 4, t11); |
|
|
wasm_v128_store(ptr + 8, t12); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave(float* ptr, const v_float32x4& a, const v_float32x4& b, |
|
|
const v_float32x4& c, const v_float32x4& d, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v_float32x4 v0, v1, v2, v3; |
|
|
v_transpose4x4(a, b, c, d, v0, v1, v2, v3); |
|
|
|
|
|
wasm_v128_store(ptr, v0.val); |
|
|
wasm_v128_store(ptr + 4, v1.val); |
|
|
wasm_v128_store(ptr + 8, v2.val); |
|
|
wasm_v128_store(ptr + 12, v3.val); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave(uint64 *ptr, const v_uint64x2& a, const v_uint64x2& b, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t v0 = wasm_unpacklo_i64x2(a.val, b.val); |
|
|
v128_t v1 = wasm_unpackhi_i64x2(a.val, b.val); |
|
|
|
|
|
wasm_v128_store(ptr, v0); |
|
|
wasm_v128_store(ptr + 2, v1); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave(uint64 *ptr, const v_uint64x2& a, const v_uint64x2& b, |
|
|
const v_uint64x2& c, hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t v0 = wasm_i8x16_shuffle(a.val, b.val, 0,1,2,3,4,5,6,7,16,17,18,19,20,21,22,23); |
|
|
v128_t v1 = wasm_i8x16_shuffle(a.val, c.val, 16,17,18,19,20,21,22,23,8,9,10,11,12,13,14,15); |
|
|
v128_t v2 = wasm_i8x16_shuffle(b.val, c.val, 8,9,10,11,12,13,14,15,24,25,26,27,28,29,30,31); |
|
|
|
|
|
wasm_v128_store(ptr, v0); |
|
|
wasm_v128_store(ptr + 2, v1); |
|
|
wasm_v128_store(ptr + 4, v2); |
|
|
} |
|
|
|
|
|
inline void v_store_interleave(uint64 *ptr, const v_uint64x2& a, const v_uint64x2& b, |
|
|
const v_uint64x2& c, const v_uint64x2& d, |
|
|
hal::StoreMode = hal::STORE_UNALIGNED) |
|
|
{ |
|
|
v128_t v0 = wasm_unpacklo_i64x2(a.val, b.val); |
|
|
v128_t v1 = wasm_unpacklo_i64x2(c.val, d.val); |
|
|
v128_t v2 = wasm_unpackhi_i64x2(a.val, b.val); |
|
|
v128_t v3 = wasm_unpackhi_i64x2(c.val, d.val); |
|
|
|
|
|
wasm_v128_store(ptr, v0); |
|
|
wasm_v128_store(ptr + 2, v1); |
|
|
wasm_v128_store(ptr + 4, v2); |
|
|
wasm_v128_store(ptr + 6, v3); |
|
|
} |
|
|
|
|
|
#define OPENCV_HAL_IMPL_WASM_LOADSTORE_INTERLEAVE(_Tpvec0, _Tp0, suffix0, _Tpvec1, _Tp1, suffix1) \ |
|
|
inline void v_load_deinterleave( const _Tp0* ptr, _Tpvec0& a0, _Tpvec0& b0 ) \ |
|
|
{ \ |
|
|
_Tpvec1 a1, b1; \ |
|
|
v_load_deinterleave((const _Tp1*)ptr, a1, b1); \ |
|
|
a0 = v_reinterpret_as_##suffix0(a1); \ |
|
|
b0 = v_reinterpret_as_##suffix0(b1); \ |
|
|
} \ |
|
|
inline void v_load_deinterleave( const _Tp0* ptr, _Tpvec0& a0, _Tpvec0& b0, _Tpvec0& c0 ) \ |
|
|
{ \ |
|
|
_Tpvec1 a1, b1, c1; \ |
|
|
v_load_deinterleave((const _Tp1*)ptr, a1, b1, c1); \ |
|
|
a0 = v_reinterpret_as_##suffix0(a1); \ |
|
|
b0 = v_reinterpret_as_##suffix0(b1); \ |
|
|
c0 = v_reinterpret_as_##suffix0(c1); \ |
|
|
} \ |
|
|
inline void v_load_deinterleave( const _Tp0* ptr, _Tpvec0& a0, _Tpvec0& b0, _Tpvec0& c0, _Tpvec0& d0 ) \ |
|
|
{ \ |
|
|
_Tpvec1 a1, b1, c1, d1; \ |
|
|
v_load_deinterleave((const _Tp1*)ptr, a1, b1, c1, d1); \ |
|
|
a0 = v_reinterpret_as_##suffix0(a1); \ |
|
|
b0 = v_reinterpret_as_##suffix0(b1); \ |
|
|
c0 = v_reinterpret_as_##suffix0(c1); \ |
|
|
d0 = v_reinterpret_as_##suffix0(d1); \ |
|
|
} \ |
|
|
inline void v_store_interleave( _Tp0* ptr, const _Tpvec0& a0, const _Tpvec0& b0, \ |
|
|
hal::StoreMode mode = hal::STORE_UNALIGNED ) \ |
|
|
{ \ |
|
|
_Tpvec1 a1 = v_reinterpret_as_##suffix1(a0); \ |
|
|
_Tpvec1 b1 = v_reinterpret_as_##suffix1(b0); \ |
|
|
v_store_interleave((_Tp1*)ptr, a1, b1, mode); \ |
|
|
} \ |
|
|
inline void v_store_interleave( _Tp0* ptr, const _Tpvec0& a0, const _Tpvec0& b0, \ |
|
|
const _Tpvec0& c0, hal::StoreMode mode = hal::STORE_UNALIGNED ) \ |
|
|
{ \ |
|
|
_Tpvec1 a1 = v_reinterpret_as_##suffix1(a0); \ |
|
|
_Tpvec1 b1 = v_reinterpret_as_##suffix1(b0); \ |
|
|
_Tpvec1 c1 = v_reinterpret_as_##suffix1(c0); \ |
|
|
v_store_interleave((_Tp1*)ptr, a1, b1, c1, mode); \ |
|
|
} \ |
|
|
inline void v_store_interleave( _Tp0* ptr, const _Tpvec0& a0, const _Tpvec0& b0, \ |
|
|
const _Tpvec0& c0, const _Tpvec0& d0, \ |
|
|
hal::StoreMode mode = hal::STORE_UNALIGNED ) \ |
|
|
{ \ |
|
|
_Tpvec1 a1 = v_reinterpret_as_##suffix1(a0); \ |
|
|
_Tpvec1 b1 = v_reinterpret_as_##suffix1(b0); \ |
|
|
_Tpvec1 c1 = v_reinterpret_as_##suffix1(c0); \ |
|
|
_Tpvec1 d1 = v_reinterpret_as_##suffix1(d0); \ |
|
|
v_store_interleave((_Tp1*)ptr, a1, b1, c1, d1, mode); \ |
|
|
} |
|
|
|
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INTERLEAVE(v_int8x16, schar, s8, v_uint8x16, uchar, u8) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INTERLEAVE(v_int16x8, short, s16, v_uint16x8, ushort, u16) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INTERLEAVE(v_int32x4, int, s32, v_uint32x4, unsigned, u32) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INTERLEAVE(v_int64x2, int64, s64, v_uint64x2, uint64, u64) |
|
|
OPENCV_HAL_IMPL_WASM_LOADSTORE_INTERLEAVE(v_float64x2, double, f64, v_uint64x2, uint64, u64) |
|
|
|
|
|
inline v_float32x4 v_cvt_f32(const v_int32x4& a) |
|
|
{ |
|
|
return v_float32x4(wasm_f32x4_convert_i32x4(a.val)); |
|
|
} |
|
|
|
|
|
inline v_float32x4 v_cvt_f32(const v_float64x2& a) |
|
|
{ |
|
|
double a_[2]; |
|
|
wasm_v128_store(a_, a.val); |
|
|
float c_[4]; |
|
|
c_[0] = (float)(a_[0]); |
|
|
c_[1] = (float)(a_[1]); |
|
|
c_[2] = 0; |
|
|
c_[3] = 0; |
|
|
return v_float32x4(wasm_v128_load(c_)); |
|
|
} |
|
|
|
|
|
inline v_float32x4 v_cvt_f32(const v_float64x2& a, const v_float64x2& b) |
|
|
{ |
|
|
double a_[2], b_[2]; |
|
|
wasm_v128_store(a_, a.val); |
|
|
wasm_v128_store(b_, b.val); |
|
|
float c_[4]; |
|
|
c_[0] = (float)(a_[0]); |
|
|
c_[1] = (float)(a_[1]); |
|
|
c_[2] = (float)(b_[0]); |
|
|
c_[3] = (float)(b_[1]); |
|
|
return v_float32x4(wasm_v128_load(c_)); |
|
|
} |
|
|
|
|
|
inline v_float64x2 v_cvt_f64(const v_int32x4& a) |
|
|
{ |
|
|
#ifdef __wasm_unimplemented_simd128__ |
|
|
v128_t p = v128_cvti32x4_i64x2(a.val); |
|
|
return v_float64x2(wasm_f64x2_convert_i64x2(p)); |
|
|
#else |
|
|
int a_[4]; |
|
|
wasm_v128_store(a_, a.val); |
|
|
double c_[2]; |
|
|
c_[0] = (double)(a_[0]); |
|
|
c_[1] = (double)(a_[1]); |
|
|
return v_float64x2(wasm_v128_load(c_)); |
|
|
#endif |
|
|
} |
|
|
|
|
|
inline v_float64x2 v_cvt_f64_high(const v_int32x4& a) |
|
|
{ |
|
|
#ifdef __wasm_unimplemented_simd128__ |
|
|
v128_t p = v128_cvti32x4_i64x2_high(a.val); |
|
|
return v_float64x2(wasm_f64x2_convert_i64x2(p)); |
|
|
#else |
|
|
int a_[4]; |
|
|
wasm_v128_store(a_, a.val); |
|
|
double c_[2]; |
|
|
c_[0] = (double)(a_[2]); |
|
|
c_[1] = (double)(a_[3]); |
|
|
return v_float64x2(wasm_v128_load(c_)); |
|
|
#endif |
|
|
} |
|
|
|
|
|
inline v_float64x2 v_cvt_f64(const v_float32x4& a) |
|
|
{ |
|
|
float a_[4]; |
|
|
wasm_v128_store(a_, a.val); |
|
|
double c_[2]; |
|
|
c_[0] = (double)(a_[0]); |
|
|
c_[1] = (double)(a_[1]); |
|
|
return v_float64x2(wasm_v128_load(c_)); |
|
|
} |
|
|
|
|
|
inline v_float64x2 v_cvt_f64_high(const v_float32x4& a) |
|
|
{ |
|
|
float a_[4]; |
|
|
wasm_v128_store(a_, a.val); |
|
|
double c_[2]; |
|
|
c_[0] = (double)(a_[2]); |
|
|
c_[1] = (double)(a_[3]); |
|
|
return v_float64x2(wasm_v128_load(c_)); |
|
|
} |
|
|
|
|
|
inline v_float64x2 v_cvt_f64(const v_int64x2& a) |
|
|
{ |
|
|
#ifdef __wasm_unimplemented_simd128__ |
|
|
return v_float64x2(wasm_f64x2_convert_i64x2(a.val)); |
|
|
#else |
|
|
int64 a_[2]; |
|
|
wasm_v128_store(a_, a.val); |
|
|
double c_[2]; |
|
|
c_[0] = (double)(a_[0]); |
|
|
c_[1] = (double)(a_[1]); |
|
|
return v_float64x2(wasm_v128_load(c_)); |
|
|
#endif |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
inline v_int8x16 v_lut(const schar* tab, const int* idx) |
|
|
{ |
|
|
return v_int8x16(tab[idx[0]], tab[idx[1]], tab[idx[ 2]], tab[idx[ 3]], tab[idx[ 4]], tab[idx[ 5]], tab[idx[ 6]], tab[idx[ 7]], |
|
|
tab[idx[8]], tab[idx[9]], tab[idx[10]], tab[idx[11]], tab[idx[12]], tab[idx[13]], tab[idx[14]], tab[idx[15]]); |
|
|
} |
|
|
inline v_int8x16 v_lut_pairs(const schar* tab, const int* idx) |
|
|
{ |
|
|
return v_int8x16(tab[idx[0]], tab[idx[0]+1], tab[idx[1]], tab[idx[1]+1], tab[idx[2]], tab[idx[2]+1], tab[idx[3]], tab[idx[3]+1], |
|
|
tab[idx[4]], tab[idx[4]+1], tab[idx[5]], tab[idx[5]+1], tab[idx[6]], tab[idx[6]+1], tab[idx[7]], tab[idx[7]+1]); |
|
|
} |
|
|
inline v_int8x16 v_lut_quads(const schar* tab, const int* idx) |
|
|
{ |
|
|
return v_int8x16(tab[idx[0]], tab[idx[0]+1], tab[idx[0]+2], tab[idx[0]+3], tab[idx[1]], tab[idx[1]+1], tab[idx[1]+2], tab[idx[1]+3], |
|
|
tab[idx[2]], tab[idx[2]+1], tab[idx[2]+2], tab[idx[2]+3], tab[idx[3]], tab[idx[3]+1], tab[idx[3]+2], tab[idx[3]+3]); |
|
|
} |
|
|
inline v_uint8x16 v_lut(const uchar* tab, const int* idx) { return v_reinterpret_as_u8(v_lut((const schar *)tab, idx)); } |
|
|
inline v_uint8x16 v_lut_pairs(const uchar* tab, const int* idx) { return v_reinterpret_as_u8(v_lut_pairs((const schar *)tab, idx)); } |
|
|
inline v_uint8x16 v_lut_quads(const uchar* tab, const int* idx) { return v_reinterpret_as_u8(v_lut_quads((const schar *)tab, idx)); } |
|
|
|
|
|
inline v_int16x8 v_lut(const short* tab, const int* idx) |
|
|
{ |
|
|
return v_int16x8(tab[idx[0]], tab[idx[1]], tab[idx[2]], tab[idx[3]], |
|
|
tab[idx[4]], tab[idx[5]], tab[idx[6]], tab[idx[7]]); |
|
|
} |
|
|
inline v_int16x8 v_lut_pairs(const short* tab, const int* idx) |
|
|
{ |
|
|
return v_int16x8(tab[idx[0]], tab[idx[0]+1], tab[idx[1]], tab[idx[1]+1], |
|
|
tab[idx[2]], tab[idx[2]+1], tab[idx[3]], tab[idx[3]+1]); |
|
|
} |
|
|
inline v_int16x8 v_lut_quads(const short* tab, const int* idx) |
|
|
{ |
|
|
return v_int16x8(tab[idx[0]], tab[idx[0]+1], tab[idx[0]+2], tab[idx[0]+3], |
|
|
tab[idx[1]], tab[idx[1]+1], tab[idx[1]+2], tab[idx[1]+3]); |
|
|
} |
|
|
inline v_uint16x8 v_lut(const ushort* tab, const int* idx) { return v_reinterpret_as_u16(v_lut((const short *)tab, idx)); } |
|
|
inline v_uint16x8 v_lut_pairs(const ushort* tab, const int* idx) { return v_reinterpret_as_u16(v_lut_pairs((const short *)tab, idx)); } |
|
|
inline v_uint16x8 v_lut_quads(const ushort* tab, const int* idx) { return v_reinterpret_as_u16(v_lut_quads((const short *)tab, idx)); } |
|
|
|
|
|
inline v_int32x4 v_lut(const int* tab, const int* idx) |
|
|
{ |
|
|
return v_int32x4(tab[idx[0]], tab[idx[1]], |
|
|
tab[idx[2]], tab[idx[3]]); |
|
|
} |
|
|
inline v_int32x4 v_lut_pairs(const int* tab, const int* idx) |
|
|
{ |
|
|
return v_int32x4(tab[idx[0]], tab[idx[0]+1], |
|
|
tab[idx[1]], tab[idx[1]+1]); |
|
|
} |
|
|
inline v_int32x4 v_lut_quads(const int* tab, const int* idx) |
|
|
{ |
|
|
return v_int32x4(wasm_v128_load(tab + idx[0])); |
|
|
} |
|
|
inline v_uint32x4 v_lut(const unsigned* tab, const int* idx) { return v_reinterpret_as_u32(v_lut((const int *)tab, idx)); } |
|
|
inline v_uint32x4 v_lut_pairs(const unsigned* tab, const int* idx) { return v_reinterpret_as_u32(v_lut_pairs((const int *)tab, idx)); } |
|
|
inline v_uint32x4 v_lut_quads(const unsigned* tab, const int* idx) { return v_reinterpret_as_u32(v_lut_quads((const int *)tab, idx)); } |
|
|
|
|
|
inline v_int64x2 v_lut(const int64_t* tab, const int* idx) |
|
|
{ |
|
|
return v_int64x2(tab[idx[0]], tab[idx[1]]); |
|
|
} |
|
|
inline v_int64x2 v_lut_pairs(const int64_t* tab, const int* idx) |
|
|
{ |
|
|
return v_int64x2(wasm_v128_load(tab + idx[0])); |
|
|
} |
|
|
inline v_uint64x2 v_lut(const uint64_t* tab, const int* idx) { return v_reinterpret_as_u64(v_lut((const int64_t *)tab, idx)); } |
|
|
inline v_uint64x2 v_lut_pairs(const uint64_t* tab, const int* idx) { return v_reinterpret_as_u64(v_lut_pairs((const int64_t *)tab, idx)); } |
|
|
|
|
|
inline v_float32x4 v_lut(const float* tab, const int* idx) |
|
|
{ |
|
|
return v_float32x4(tab[idx[0]], tab[idx[1]], tab[idx[2]], tab[idx[3]]); |
|
|
} |
|
|
inline v_float32x4 v_lut_pairs(const float* tab, const int* idx) { return v_reinterpret_as_f32(v_lut_pairs((const int *)tab, idx)); } |
|
|
inline v_float32x4 v_lut_quads(const float* tab, const int* idx) { return v_reinterpret_as_f32(v_lut_quads((const int *)tab, idx)); } |
|
|
|
|
|
inline v_float64x2 v_lut(const double* tab, const int* idx) |
|
|
{ |
|
|
return v_float64x2(tab[idx[0]], tab[idx[1]]); |
|
|
} |
|
|
inline v_float64x2 v_lut_pairs(const double* tab, const int* idx) |
|
|
{ |
|
|
return v_float64x2(wasm_v128_load(tab + idx[0])); |
|
|
} |
|
|
|
|
|
inline v_int32x4 v_lut(const int* tab, const v_int32x4& idxvec) |
|
|
{ |
|
|
return v_int32x4(tab[wasm_i32x4_extract_lane(idxvec.val, 0)], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 1)], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 2)], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 3)]); |
|
|
} |
|
|
|
|
|
inline v_uint32x4 v_lut(const unsigned* tab, const v_int32x4& idxvec) |
|
|
{ |
|
|
return v_reinterpret_as_u32(v_lut((const int *)tab, idxvec)); |
|
|
} |
|
|
|
|
|
inline v_float32x4 v_lut(const float* tab, const v_int32x4& idxvec) |
|
|
{ |
|
|
return v_float32x4(tab[wasm_i32x4_extract_lane(idxvec.val, 0)], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 1)], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 2)], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 3)]); |
|
|
} |
|
|
|
|
|
inline v_float64x2 v_lut(const double* tab, const v_int32x4& idxvec) |
|
|
{ |
|
|
return v_float64x2(tab[wasm_i32x4_extract_lane(idxvec.val, 0)], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 1)]); |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
inline void v_lut_deinterleave(const float* tab, const v_int32x4& idxvec, v_float32x4& x, v_float32x4& y) |
|
|
{ |
|
|
x = v_float32x4(tab[wasm_i32x4_extract_lane(idxvec.val, 0)], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 1)], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 2)], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 3)]); |
|
|
y = v_float32x4(tab[wasm_i32x4_extract_lane(idxvec.val, 0)+1], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 1)+1], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 2)+1], |
|
|
tab[wasm_i32x4_extract_lane(idxvec.val, 3)+1]); |
|
|
} |
|
|
|
|
|
inline void v_lut_deinterleave(const double* tab, const v_int32x4& idxvec, v_float64x2& x, v_float64x2& y) |
|
|
{ |
|
|
v128_t xy0 = wasm_v128_load(tab + wasm_i32x4_extract_lane(idxvec.val, 0)); |
|
|
v128_t xy1 = wasm_v128_load(tab + wasm_i32x4_extract_lane(idxvec.val, 1)); |
|
|
x.val = wasm_unpacklo_i64x2(xy0, xy1); |
|
|
y.val = wasm_unpacklo_i64x2(xy0, xy1); |
|
|
} |
|
|
|
|
|
inline v_int8x16 v_interleave_pairs(const v_int8x16& vec) |
|
|
{ |
|
|
return v_int8x16(wasm_i8x16_shuffle(vec.val, vec.val, 0,2,1,3,4,6,5,7,8,10,9,11,12,14,13,15)); |
|
|
} |
|
|
inline v_uint8x16 v_interleave_pairs(const v_uint8x16& vec) { return v_reinterpret_as_u8(v_interleave_pairs(v_reinterpret_as_s8(vec))); } |
|
|
inline v_int8x16 v_interleave_quads(const v_int8x16& vec) |
|
|
{ |
|
|
return v_int8x16(wasm_i8x16_shuffle(vec.val, vec.val, 0,4,1,5,2,6,3,7,8,12,9,13,10,14,11,15)); |
|
|
} |
|
|
inline v_uint8x16 v_interleave_quads(const v_uint8x16& vec) { return v_reinterpret_as_u8(v_interleave_quads(v_reinterpret_as_s8(vec))); } |
|
|
|
|
|
inline v_int16x8 v_interleave_pairs(const v_int16x8& vec) |
|
|
{ |
|
|
return v_int16x8(wasm_i8x16_shuffle(vec.val, vec.val, 0,1,4,5,2,3,6,7,8,9,12,13,10,11,14,15)); |
|
|
} |
|
|
inline v_uint16x8 v_interleave_pairs(const v_uint16x8& vec) { return v_reinterpret_as_u16(v_interleave_pairs(v_reinterpret_as_s16(vec))); } |
|
|
inline v_int16x8 v_interleave_quads(const v_int16x8& vec) |
|
|
{ |
|
|
return v_int16x8(wasm_i8x16_shuffle(vec.val, vec.val, 0,1,8,9,2,3,10,11,4,5,12,13,6,7,14,15)); |
|
|
} |
|
|
inline v_uint16x8 v_interleave_quads(const v_uint16x8& vec) { return v_reinterpret_as_u16(v_interleave_quads(v_reinterpret_as_s16(vec))); } |
|
|
|
|
|
inline v_int32x4 v_interleave_pairs(const v_int32x4& vec) |
|
|
{ |
|
|
return v_int32x4(wasm_i8x16_shuffle(vec.val, vec.val, 0,1,2,3,8,9,10,11,4,5,6,7,12,13,14,15)); |
|
|
} |
|
|
inline v_uint32x4 v_interleave_pairs(const v_uint32x4& vec) { return v_reinterpret_as_u32(v_interleave_pairs(v_reinterpret_as_s32(vec))); } |
|
|
inline v_float32x4 v_interleave_pairs(const v_float32x4& vec) |
|
|
{ |
|
|
return v_float32x4(wasm_i8x16_shuffle(vec.val, vec.val, 0,1,2,3,8,9,10,11,4,5,6,7,12,13,14,15)); |
|
|
} |
|
|
|
|
|
inline v_int8x16 v_pack_triplets(const v_int8x16& vec) |
|
|
{ |
|
|
return v_int8x16(wasm_i8x16_shuffle(vec.val, vec.val, 0,1,2,4,5,6,8,9,10,12,13,14,16,16,16,16)); |
|
|
} |
|
|
inline v_uint8x16 v_pack_triplets(const v_uint8x16& vec) { return v_reinterpret_as_u8(v_pack_triplets(v_reinterpret_as_s8(vec))); } |
|
|
|
|
|
inline v_int16x8 v_pack_triplets(const v_int16x8& vec) |
|
|
{ |
|
|
return v_int16x8(wasm_i8x16_shuffle(vec.val, vec.val, 0,1,2,3,4,5,8,9,10,11,12,13,14,15,6,7)); |
|
|
} |
|
|
inline v_uint16x8 v_pack_triplets(const v_uint16x8& vec) { return v_reinterpret_as_u16(v_pack_triplets(v_reinterpret_as_s16(vec))); } |
|
|
|
|
|
inline v_int32x4 v_pack_triplets(const v_int32x4& vec) { return vec; } |
|
|
inline v_uint32x4 v_pack_triplets(const v_uint32x4& vec) { return vec; } |
|
|
inline v_float32x4 v_pack_triplets(const v_float32x4& vec) { return vec; } |
|
|
|
|
|
template<int i, typename _Tp> |
|
|
inline typename _Tp::lane_type v_extract_n(const _Tp& a) |
|
|
{ |
|
|
return v_rotate_right<i>(a).get0(); |
|
|
} |
|
|
|
|
|
template<int i> |
|
|
inline v_uint32x4 v_broadcast_element(const v_uint32x4& a) |
|
|
{ |
|
|
return v_setall_u32(v_extract_n<i>(a)); |
|
|
} |
|
|
template<int i> |
|
|
inline v_int32x4 v_broadcast_element(const v_int32x4& a) |
|
|
{ |
|
|
return v_setall_s32(v_extract_n<i>(a)); |
|
|
} |
|
|
template<int i> |
|
|
inline v_float32x4 v_broadcast_element(const v_float32x4& a) |
|
|
{ |
|
|
return v_setall_f32(v_extract_n<i>(a)); |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
inline v_float32x4 v_load_expand(const hfloat* ptr) |
|
|
{ |
|
|
float a[4]; |
|
|
for (int i = 0; i < 4; i++) |
|
|
a[i] = ptr[i]; |
|
|
return v_float32x4(wasm_v128_load(a)); |
|
|
} |
|
|
|
|
|
inline void v_pack_store(hfloat* ptr, const v_float32x4& v) |
|
|
{ |
|
|
double v_[4]; |
|
|
wasm_v128_store(v_, v.val); |
|
|
ptr[0] = hfloat(v_[0]); |
|
|
ptr[1] = hfloat(v_[1]); |
|
|
ptr[2] = hfloat(v_[2]); |
|
|
ptr[3] = hfloat(v_[3]); |
|
|
} |
|
|
|
|
|
inline void v_cleanup() {} |
|
|
|
|
|
#include "intrin_math.hpp" |
|
|
inline v_float32x4 v_exp(const v_float32x4& x) { return v_exp_default_32f<v_float32x4, v_int32x4>(x); } |
|
|
inline v_float32x4 v_log(const v_float32x4& x) { return v_log_default_32f<v_float32x4, v_int32x4>(x); } |
|
|
inline void v_sincos(const v_float32x4& x, v_float32x4& s, v_float32x4& c) { v_sincos_default_32f<v_float32x4, v_int32x4>(x, s, c); } |
|
|
inline v_float32x4 v_sin(const v_float32x4& x) { return v_sin_default_32f<v_float32x4, v_int32x4>(x); } |
|
|
inline v_float32x4 v_cos(const v_float32x4& x) { return v_cos_default_32f<v_float32x4, v_int32x4>(x); } |
|
|
inline v_float32x4 v_erf(const v_float32x4& x) { return v_erf_default_32f<v_float32x4, v_int32x4>(x); } |
|
|
|
|
|
inline v_float64x2 v_exp(const v_float64x2& x) { return v_exp_default_64f<v_float64x2, v_int64x2>(x); } |
|
|
inline v_float64x2 v_log(const v_float64x2& x) { return v_log_default_64f<v_float64x2, v_int64x2>(x); } |
|
|
inline void v_sincos(const v_float64x2& x, v_float64x2& s, v_float64x2& c) { v_sincos_default_64f<v_float64x2, v_int64x2>(x, s, c); } |
|
|
inline v_float64x2 v_sin(const v_float64x2& x) { return v_sin_default_64f<v_float64x2, v_int64x2>(x); } |
|
|
inline v_float64x2 v_cos(const v_float64x2& x) { return v_cos_default_64f<v_float64x2, v_int64x2>(x); } |
|
|
|
|
|
CV_CPU_OPTIMIZATION_HAL_NAMESPACE_END |
|
|
|
|
|
|
|
|
|
|
|
} |
|
|
|
|
|
#endif |
|
|
|