File size: 3,585 Bytes
be903e2 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 | // Tencent is pleased to support the open source community by making ncnn available.
//
// Copyright (C) 2022 THL A29 Limited, a Tencent company. All rights reserved.
//
// Licensed under the BSD 3-Clause License (the "License"); you may not use this file except
// in compliance with the License. You may obtain a copy of the License at
//
// https://opensource.org/licenses/BSD-3-Clause
//
// Unless required by applicable law or agreed to in writing, software distributed
// under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
// CONDITIONS OF ANY KIND, either express or implied. See the License for the
// specific language governing permissions and limitations under the License.
#include "clip_arm.h"
#ifdef __ARM_NEON
#include <arm_neon.h>
#endif // __ARM_NEON
namespace ncnn {
#if __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
int Clip_arm::forward_inplace_fp16s(Mat& bottom_top_blob, const Option& opt) const
{
int w = bottom_top_blob.w;
int h = bottom_top_blob.h;
int d = bottom_top_blob.d;
int channels = bottom_top_blob.c;
int elempack = bottom_top_blob.elempack;
int size = w * h * d * elempack;
#pragma omp parallel for num_threads(opt.num_threads)
for (int q = 0; q < channels; q++)
{
__fp16* ptr = bottom_top_blob.channel(q);
__fp16 min_fp16 = min;
__fp16 max_fp16 = max;
float16x8_t _min = vdupq_n_f16(min_fp16);
float16x8_t _max = vdupq_n_f16(max_fp16);
int i = 0;
for (; i + 31 < size; i += 32)
{
asm volatile(
"prfm pldl1keep, [%0, #512] \n"
"ld1 {v0.8h, v1.8h, v2.8h, v3.8h}, [%0] \n"
"fmax v0.8h, v0.8h, %2.8h \n"
"fmax v1.8h, v1.8h, %2.8h \n"
"fmax v2.8h, v2.8h, %2.8h \n"
"fmax v3.8h, v3.8h, %2.8h \n"
"fmin v0.8h, v0.8h, %3.8h \n"
"fmin v1.8h, v1.8h, %3.8h \n"
"fmin v2.8h, v2.8h, %3.8h \n"
"fmin v3.8h, v3.8h, %3.8h \n"
"st1 {v0.8h, v1.8h, v2.8h, v3.8h}, [%0], #64 \n"
: "=r"(ptr) // %0
: "0"(ptr),
"w"(_min), // %2
"w"(_max) // %3
: "memory", "v0", "v1", "v2", "v3");
}
for (; i + 15 < size; i += 16)
{
float16x8_t _p0 = vld1q_f16(ptr);
float16x8_t _p1 = vld1q_f16(ptr + 8);
_p0 = vmaxq_f16(_p0, _min);
_p1 = vmaxq_f16(_p1, _min);
_p0 = vminq_f16(_p0, _max);
_p1 = vminq_f16(_p1, _max);
vst1q_f16(ptr, _p0);
vst1q_f16(ptr + 8, _p1);
ptr += 16;
}
for (; i + 7 < size; i += 8)
{
float16x8_t _p = vld1q_f16(ptr);
_p = vmaxq_f16(_p, _min);
_p = vminq_f16(_p, _max);
vst1q_f16(ptr, _p);
ptr += 8;
}
for (; i + 3 < size; i += 4)
{
float16x4_t _p = vld1_f16(ptr);
_p = vmax_f16(_p, vget_low_f16(_min));
_p = vmin_f16(_p, vget_low_f16(_max));
vst1_f16(ptr, _p);
ptr += 4;
}
for (; i < size; i++)
{
__fp16 v = *ptr;
if (v < min_fp16)
v = min_fp16;
if (v > max_fp16)
v = max_fp16;
*ptr = v;
ptr++;
}
}
return 0;
}
#endif // __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
} // namespace ncnn
|