| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | #include "bias_arm.h" |
| |
|
| | #if __ARM_NEON |
| | #include <arm_neon.h> |
| | #endif |
| |
|
| | namespace ncnn { |
| |
|
| | int Bias_arm::forward_inplace(Mat& bottom_top_blob, const Option& opt) const |
| | { |
| | int w = bottom_top_blob.w; |
| | int h = bottom_top_blob.h; |
| | int d = bottom_top_blob.d; |
| | int channels = bottom_top_blob.c; |
| | int size = w * h * d; |
| |
|
| | const float* bias_ptr = bias_data; |
| | #pragma omp parallel for num_threads(opt.num_threads) |
| | for (int q = 0; q < channels; q++) |
| | { |
| | float* ptr = bottom_top_blob.channel(q); |
| |
|
| | float bias = bias_ptr[q]; |
| |
|
| | #if __ARM_NEON |
| | int nn = size >> 2; |
| | int remain = size - (nn << 2); |
| | #else |
| | int remain = size; |
| | #endif |
| |
|
| | #if __ARM_NEON |
| | float32x4_t _bias = vdupq_n_f32(bias); |
| | for (; nn > 0; nn--) |
| | { |
| | float32x4_t _p = vld1q_f32(ptr); |
| | float32x4_t _outp = vaddq_f32(_p, _bias); |
| | vst1q_f32(ptr, _outp); |
| |
|
| | ptr += 4; |
| | } |
| | #endif |
| |
|
| | for (; remain > 0; remain--) |
| | { |
| | *ptr = *ptr + bias; |
| |
|
| | ptr++; |
| | } |
| | } |
| |
|
| | return 0; |
| | } |
| |
|
| | } |
| |
|