// Tencent is pleased to support the open source community by making ncnn available. // // Copyright (C) 2019 THL A29 Limited, a Tencent company. All rights reserved. // // Licensed under the BSD 3-Clause License (the "License"); you may not use this file except // in compliance with the License. You may obtain a copy of the License at // // https://opensource.org/licenses/BSD-3-Clause // // Unless required by applicable law or agreed to in writing, software distributed // under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR // CONDITIONS OF ANY KIND, either express or implied. See the License for the // specific language governing permissions and limitations under the License. #include "cast_x86_fma.h" #if __SSE2__ #include #if __AVX__ #include #endif // __AVX__ #endif // __SSE2__ #include "x86_usability.h" #include "cpu.h" namespace ncnn { #include "cast_fp16.h" #include "cast_bf16.h" Cast_x86_fma::Cast_x86_fma() { support_packing = true; } int Cast_x86_fma::forward(const Mat& bottom_blob, Mat& top_blob, const Option& opt) const { if (type_from == type_to) { top_blob = bottom_blob; return 0; } int w = bottom_blob.w; int h = bottom_blob.h; int d = bottom_blob.d; int channels = bottom_blob.c; int dims = bottom_blob.dims; size_t elemsize = bottom_blob.elemsize; int elempack = bottom_blob.elempack; size_t out_elemsize = elemsize; if (type_to == 1) { if (type_from == 3) { Cast::forward(bottom_blob, top_blob, opt); } // float32 out_elemsize = 4 * elempack; } else if (type_to == 2) { // float16 out_elemsize = 2 * elempack; } else if (type_to == 3) { // int8 out_elemsize = elempack; } else if (type_to == 4) { // bfloat16 out_elemsize = 2 * elempack; } if (dims == 1) { top_blob.create(w, out_elemsize, elempack, opt.blob_allocator); } else if (dims == 2) { top_blob.create(w, h, out_elemsize, elempack, opt.blob_allocator); } else if (dims == 3) { top_blob.create(w, h, channels, out_elemsize, elempack, opt.blob_allocator); } else if (dims == 4) { top_blob.create(w, h, d, channels, out_elemsize, elempack, opt.blob_allocator); } if (top_blob.empty()) return -100; int size = w * h * d * elempack; if (type_from == 1 && type_to == 2) { cast_fp32_to_fp16_sse(bottom_blob, top_blob, opt); } if (type_from == 2 && type_to == 1) { cast_fp16_to_fp32_sse(bottom_blob, top_blob, opt); } if (type_from == 3 && type_to == 1) { #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { const signed char* ptr = bottom_blob.channel(q); float* outptr = top_blob.channel(q); for (int i = 0; i < size; i++) { outptr[i] = (float)ptr[i]; } } } if (type_from == 1 && type_to == 4) { cast_fp32_to_bf16_sse(bottom_blob, top_blob, opt); } if (type_from == 4 && type_to == 1) { cast_bf16_to_fp32_sse(bottom_blob, top_blob, opt); } return 0; } } // namespace ncnn