|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#ifndef OPENCV_CUDA_EMULATION_HPP_ |
|
|
#define OPENCV_CUDA_EMULATION_HPP_ |
|
|
|
|
|
#include "common.hpp" |
|
|
#include "warp_reduce.hpp" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
namespace cv { namespace cuda { namespace device |
|
|
{ |
|
|
struct Emulation |
|
|
{ |
|
|
|
|
|
static __device__ __forceinline__ int syncthreadsOr(int pred) |
|
|
{ |
|
|
#if defined (__CUDA_ARCH__) && (__CUDA_ARCH__ < 200) |
|
|
|
|
|
return 0; |
|
|
#else |
|
|
return __syncthreads_or(pred); |
|
|
#endif |
|
|
} |
|
|
|
|
|
template<int CTA_SIZE> |
|
|
static __forceinline__ __device__ int Ballot(int predicate) |
|
|
{ |
|
|
#if defined (__CUDA_ARCH__) && (__CUDA_ARCH__ >= 200) |
|
|
return __ballot(predicate); |
|
|
#else |
|
|
__shared__ volatile int cta_buffer[CTA_SIZE]; |
|
|
|
|
|
int tid = threadIdx.x; |
|
|
cta_buffer[tid] = predicate ? (1 << (tid & 31)) : 0; |
|
|
return warp_reduce(cta_buffer); |
|
|
#endif |
|
|
} |
|
|
|
|
|
struct smem |
|
|
{ |
|
|
enum { TAG_MASK = (1U << ( (sizeof(unsigned int) << 3) - 5U)) - 1U }; |
|
|
|
|
|
template<typename T> |
|
|
static __device__ __forceinline__ T atomicInc(T* address, T val) |
|
|
{ |
|
|
#if defined (__CUDA_ARCH__) && (__CUDA_ARCH__ < 120) |
|
|
T count; |
|
|
unsigned int tag = threadIdx.x << ( (sizeof(unsigned int) << 3) - 5U); |
|
|
do |
|
|
{ |
|
|
count = *address & TAG_MASK; |
|
|
count = tag | (count + 1); |
|
|
*address = count; |
|
|
} while (*address != count); |
|
|
|
|
|
return (count & TAG_MASK) - 1; |
|
|
#else |
|
|
return ::atomicInc(address, val); |
|
|
#endif |
|
|
} |
|
|
|
|
|
template<typename T> |
|
|
static __device__ __forceinline__ T atomicAdd(T* address, T val) |
|
|
{ |
|
|
#if defined (__CUDA_ARCH__) && (__CUDA_ARCH__ < 120) |
|
|
T count; |
|
|
unsigned int tag = threadIdx.x << ( (sizeof(unsigned int) << 3) - 5U); |
|
|
do |
|
|
{ |
|
|
count = *address & TAG_MASK; |
|
|
count = tag | (count + val); |
|
|
*address = count; |
|
|
} while (*address != count); |
|
|
|
|
|
return (count & TAG_MASK) - val; |
|
|
#else |
|
|
return ::atomicAdd(address, val); |
|
|
#endif |
|
|
} |
|
|
|
|
|
template<typename T> |
|
|
static __device__ __forceinline__ T atomicMin(T* address, T val) |
|
|
{ |
|
|
#if defined (__CUDA_ARCH__) && (__CUDA_ARCH__ < 120) |
|
|
T count = ::min(*address, val); |
|
|
do |
|
|
{ |
|
|
*address = count; |
|
|
} while (*address > count); |
|
|
|
|
|
return count; |
|
|
#else |
|
|
return ::atomicMin(address, val); |
|
|
#endif |
|
|
} |
|
|
}; |
|
|
|
|
|
struct glob |
|
|
{ |
|
|
static __device__ __forceinline__ int atomicAdd(int* address, int val) |
|
|
{ |
|
|
return ::atomicAdd(address, val); |
|
|
} |
|
|
static __device__ __forceinline__ unsigned int atomicAdd(unsigned int* address, unsigned int val) |
|
|
{ |
|
|
return ::atomicAdd(address, val); |
|
|
} |
|
|
static __device__ __forceinline__ float atomicAdd(float* address, float val) |
|
|
{ |
|
|
#if __CUDA_ARCH__ >= 200 |
|
|
return ::atomicAdd(address, val); |
|
|
#else |
|
|
int* address_as_i = (int*) address; |
|
|
int old = *address_as_i, assumed; |
|
|
do { |
|
|
assumed = old; |
|
|
old = ::atomicCAS(address_as_i, assumed, |
|
|
__float_as_int(val + __int_as_float(assumed))); |
|
|
} while (assumed != old); |
|
|
return __int_as_float(old); |
|
|
#endif |
|
|
} |
|
|
static __device__ __forceinline__ double atomicAdd(double* address, double val) |
|
|
{ |
|
|
#if __CUDA_ARCH__ >= 130 |
|
|
unsigned long long int* address_as_ull = (unsigned long long int*) address; |
|
|
unsigned long long int old = *address_as_ull, assumed; |
|
|
do { |
|
|
assumed = old; |
|
|
old = ::atomicCAS(address_as_ull, assumed, |
|
|
__double_as_longlong(val + __longlong_as_double(assumed))); |
|
|
} while (assumed != old); |
|
|
return __longlong_as_double(old); |
|
|
#else |
|
|
CV_UNUSED(address); |
|
|
CV_UNUSED(val); |
|
|
return 0.0; |
|
|
#endif |
|
|
} |
|
|
|
|
|
static __device__ __forceinline__ int atomicMin(int* address, int val) |
|
|
{ |
|
|
return ::atomicMin(address, val); |
|
|
} |
|
|
static __device__ __forceinline__ float atomicMin(float* address, float val) |
|
|
{ |
|
|
#if __CUDA_ARCH__ >= 120 |
|
|
int* address_as_i = (int*) address; |
|
|
int old = *address_as_i, assumed; |
|
|
do { |
|
|
assumed = old; |
|
|
old = ::atomicCAS(address_as_i, assumed, |
|
|
__float_as_int(::fminf(val, __int_as_float(assumed)))); |
|
|
} while (assumed != old); |
|
|
return __int_as_float(old); |
|
|
#else |
|
|
CV_UNUSED(address); |
|
|
CV_UNUSED(val); |
|
|
return 0.0f; |
|
|
#endif |
|
|
} |
|
|
static __device__ __forceinline__ double atomicMin(double* address, double val) |
|
|
{ |
|
|
#if __CUDA_ARCH__ >= 130 |
|
|
unsigned long long int* address_as_ull = (unsigned long long int*) address; |
|
|
unsigned long long int old = *address_as_ull, assumed; |
|
|
do { |
|
|
assumed = old; |
|
|
old = ::atomicCAS(address_as_ull, assumed, |
|
|
__double_as_longlong(::fmin(val, __longlong_as_double(assumed)))); |
|
|
} while (assumed != old); |
|
|
return __longlong_as_double(old); |
|
|
#else |
|
|
CV_UNUSED(address); |
|
|
CV_UNUSED(val); |
|
|
return 0.0; |
|
|
#endif |
|
|
} |
|
|
|
|
|
static __device__ __forceinline__ int atomicMax(int* address, int val) |
|
|
{ |
|
|
return ::atomicMax(address, val); |
|
|
} |
|
|
static __device__ __forceinline__ float atomicMax(float* address, float val) |
|
|
{ |
|
|
#if __CUDA_ARCH__ >= 120 |
|
|
int* address_as_i = (int*) address; |
|
|
int old = *address_as_i, assumed; |
|
|
do { |
|
|
assumed = old; |
|
|
old = ::atomicCAS(address_as_i, assumed, |
|
|
__float_as_int(::fmaxf(val, __int_as_float(assumed)))); |
|
|
} while (assumed != old); |
|
|
return __int_as_float(old); |
|
|
#else |
|
|
CV_UNUSED(address); |
|
|
CV_UNUSED(val); |
|
|
return 0.0f; |
|
|
#endif |
|
|
} |
|
|
static __device__ __forceinline__ double atomicMax(double* address, double val) |
|
|
{ |
|
|
#if __CUDA_ARCH__ >= 130 |
|
|
unsigned long long int* address_as_ull = (unsigned long long int*) address; |
|
|
unsigned long long int old = *address_as_ull, assumed; |
|
|
do { |
|
|
assumed = old; |
|
|
old = ::atomicCAS(address_as_ull, assumed, |
|
|
__double_as_longlong(::fmax(val, __longlong_as_double(assumed)))); |
|
|
} while (assumed != old); |
|
|
return __longlong_as_double(old); |
|
|
#else |
|
|
CV_UNUSED(address); |
|
|
CV_UNUSED(val); |
|
|
return 0.0; |
|
|
#endif |
|
|
} |
|
|
}; |
|
|
}; |
|
|
}}} |
|
|
|
|
|
|
|
|
|
|
|
#endif |
|
|
|