| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| #include "bias_x86_avx.h" |
|
|
| #if __SSE2__ |
| #include <emmintrin.h> |
| #if __AVX__ |
| #include <immintrin.h> |
| #endif |
| #endif |
|
|
| namespace ncnn { |
|
|
| int Bias_x86_avx::forward_inplace(Mat& bottom_top_blob, const Option& opt) const |
| { |
| int w = bottom_top_blob.w; |
| int h = bottom_top_blob.h; |
| int d = bottom_top_blob.d; |
| int channels = bottom_top_blob.c; |
| int size = w * h * d; |
|
|
| const float* bias_ptr = bias_data; |
| #pragma omp parallel for num_threads(opt.num_threads) |
| for (int q = 0; q < channels; q++) |
| { |
| float* ptr = bottom_top_blob.channel(q); |
|
|
| float bias = bias_ptr[q]; |
|
|
| int i = 0; |
| #if __SSE2__ |
| #if __AVX__ |
| { |
| __m256 _bias256 = _mm256_set1_ps(bias); |
| for (; i + 7 < size; i += 8) |
| { |
| __m256 _p = _mm256_loadu_ps(ptr); |
| __m256 _outp = _mm256_add_ps(_p, _bias256); |
| _mm256_storeu_ps(ptr, _outp); |
|
|
| ptr += 8; |
| } |
| } |
| #endif |
| { |
| __m128 _bias = _mm_set1_ps(bias); |
| for (; i + 3 < size; i += 4) |
| { |
| __m128 _p = _mm_loadu_ps(ptr); |
| __m128 _outp = _mm_add_ps(_p, _bias); |
| _mm_storeu_ps(ptr, _outp); |
|
|
| ptr += 4; |
| } |
| } |
| #endif |
|
|
| for (; i < size; i++) |
| { |
| *ptr = *ptr + bias; |
|
|
| ptr++; |
| } |
| } |
|
|
| return 0; |
| } |
|
|
| } |
|
|