ncnn / src /layer /x86 /bias_x86.cpp
camenduru's picture
thanks to ncnn ❤
be903e2
// Tencent is pleased to support the open source community by making ncnn available.
//
// Copyright (C) 2017 THL A29 Limited, a Tencent company. All rights reserved.
//
// Licensed under the BSD 3-Clause License (the "License"); you may not use this file except
// in compliance with the License. You may obtain a copy of the License at
//
// https://opensource.org/licenses/BSD-3-Clause
//
// Unless required by applicable law or agreed to in writing, software distributed
// under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
// CONDITIONS OF ANY KIND, either express or implied. See the License for the
// specific language governing permissions and limitations under the License.
#include "bias_x86.h"
#if __SSE2__
#include <emmintrin.h>
#if __AVX__
#include <immintrin.h>
#endif // __AVX__
#endif // __SSE2__
namespace ncnn {
int Bias_x86::forward_inplace(Mat& bottom_top_blob, const Option& opt) const
{
int w = bottom_top_blob.w;
int h = bottom_top_blob.h;
int d = bottom_top_blob.d;
int channels = bottom_top_blob.c;
int size = w * h * d;
const float* bias_ptr = bias_data;
#pragma omp parallel for num_threads(opt.num_threads)
for (int q = 0; q < channels; q++)
{
float* ptr = bottom_top_blob.channel(q);
float bias = bias_ptr[q];
int i = 0;
#if __SSE2__
#if __AVX__
{
__m256 _bias256 = _mm256_set1_ps(bias);
for (; i + 7 < size; i += 8)
{
__m256 _p = _mm256_loadu_ps(ptr);
__m256 _outp = _mm256_add_ps(_p, _bias256);
_mm256_storeu_ps(ptr, _outp);
ptr += 8;
}
}
#endif // __AVX__
{
__m128 _bias = _mm_set1_ps(bias);
for (; i + 3 < size; i += 4)
{
__m128 _p = _mm_loadu_ps(ptr);
__m128 _outp = _mm_add_ps(_p, _bias);
_mm_storeu_ps(ptr, _outp);
ptr += 4;
}
}
#endif // __SSE2__
for (; i < size; i++)
{
*ptr = *ptr + bias;
ptr++;
}
}
return 0;
}
} // namespace ncnn