/*! * Modified from Deformable DETR */ #include #include #include #include #include #include #include #include #include #include #define CUDA_KERNEL_LOOP(i, n) \ for (int i = blockIdx.x * blockDim.x + threadIdx.x; \ i < (n); \ i += blockDim.x * gridDim.x) #define CUDA_NUM_THREADS 512 #define MAX_POINT 32 inline int GET_BLOCKS(const int N, const int num_threads) { return (N + num_threads - 1) / num_threads; } __device__ void ms_deform_attn_col2im_bilinear(const float *&bottom_data, const int &height, const int &width, const int &channels, const float &h, const float &w, const int &c, const float &top_grad, const float &attn_weight, const float *&grad_value, float *&grad_sampling_loc, float *&grad_attn_weight) { const int h_low = floor(h); const int w_low = floor(w); const int h_high = h_low + 1; const int w_high = w_low + 1; const float lh = h - h_low; const float lw = w - w_low; const float hh = 1 - lh, hw = 1 - lw; const int w_stride = channels; const int h_stride = width * w_stride; const int h_low_ptr_offset = h_low * h_stride; const int h_high_ptr_offset = h_low_ptr_offset + h_stride; const int w_low_ptr_offset = w_low * w_stride; const int w_high_ptr_offset = w_low_ptr_offset + w_stride; const float w1 = hh * hw, w2 = hh * lw, w3 = lh * hw, w4 = lh * lw; const float top_grad_value = top_grad * attn_weight; float grad_h_weight = 0, grad_w_weight = 0; float *grad_ptr; float v1 = 0; if (h_low >= 0 && w_low >= 0) { const int ptr1 = h_low_ptr_offset + w_low_ptr_offset + c; grad_ptr = const_cast(grad_value + ptr1); v1 = bottom_data[ptr1]; grad_h_weight -= hw * v1; grad_w_weight -= hh * v1; atomicAdd(grad_ptr, w1 * top_grad_value); } float v2 = 0; if (h_low >= 0 && w_high <= width - 1) { const int ptr2 = h_low_ptr_offset + w_high_ptr_offset + c; grad_ptr = const_cast(grad_value + ptr2); v2 = bottom_data[ptr2]; grad_h_weight -= lw * v2; grad_w_weight += hh * v2; atomicAdd(grad_ptr, w2 * top_grad_value); } float v3 = 0; if (h_high <= height - 1 && w_low >= 0) { const int ptr3 = h_high_ptr_offset + w_low_ptr_offset + c; grad_ptr = const_cast(grad_value + ptr3); v3 = bottom_data[ptr3]; grad_h_weight += hw * v3; grad_w_weight -= lh * v3; atomicAdd(grad_ptr, w3 * top_grad_value); } float v4 = 0; if (h_high <= height - 1 && w_high <= width - 1) { const int ptr4 = h_high_ptr_offset + w_high_ptr_offset + c; grad_ptr = const_cast(grad_value + ptr4); v4 = bottom_data[ptr4]; grad_h_weight += lw * v4; grad_w_weight += lh * v4; atomicAdd(grad_ptr, w4 * top_grad_value); } const float val = (w1 * v1 + w2 * v2 + w3 * v3 + w4 * v4); atomicAdd(grad_attn_weight, top_grad * val); atomicAdd(grad_sampling_loc, (width - 1) * grad_w_weight * top_grad_value); atomicAdd(grad_sampling_loc + 1, (height - 1) * grad_h_weight * top_grad_value); } // global_memory_way __global__ void ms_deformable_col2im_gpu_kernel_gm_c2345( const float *grad_col, const float *feat_c2, const float *feat_c3, const float *feat_c4, const float *feat_c5, const int h_c2, const int w_c2, const int h_c3, const int w_c3, const int h_c4, const int w_c4, const int h_c5, const int w_c5, const float *data_sampling_loc, const float *data_attn_weight, const int batch_size, const int channels, const int num_views, const int num_query, const int num_point, float *grad_value_c2, float *grad_value_c3, float *grad_value_c4, float *grad_value_c5, float *grad_sampling_loc, float *grad_attn_weight) { CUDA_KERNEL_LOOP(index, batch_size * num_query * channels * num_point) { // n: bs x query x channels int _temp = index; const int p_col = _temp % num_point; _temp /= num_point; const int c_col = _temp % channels; _temp /= channels; const int sampling_index = _temp; _temp /= num_query; const int b_col = _temp; const float top_grad = grad_col[index]; // Sampling location in range [0, 1] int data_loc_ptr = sampling_index * num_point * 3 + p_col * 3; const float loc_w = data_sampling_loc[data_loc_ptr]; const float loc_h = data_sampling_loc[data_loc_ptr + 1]; const int loc_v = round(data_sampling_loc[data_loc_ptr + 2] * (num_views - 1)); // Attn weights int data_weight_ptr = sampling_index * num_point * 4 + p_col * 4; const float weight_c2 = data_attn_weight[data_weight_ptr]; const float weight_c3 = data_attn_weight[data_weight_ptr + 1]; const float weight_c4 = data_attn_weight[data_weight_ptr + 2]; const float weight_c5 = data_attn_weight[data_weight_ptr + 3]; // const float h_im = loc_h * spatial_h - 0.5; // align_corners = False // const float w_im = loc_w * spatial_w - 0.5; // C2 Feature float h_im = loc_h * (h_c2 - 1); // align_corners = True float w_im = loc_w * (w_c2 - 1); float *grad_location_ptr = grad_sampling_loc + data_loc_ptr; float *grad_weights_ptr = grad_attn_weight + data_weight_ptr; if (h_im > -1 && w_im > -1 && h_im < h_c2 && w_im < w_c2) { const float *feat_c2_ptr = feat_c2 + b_col * num_views * h_c2 * w_c2 * channels + loc_v * h_c2 * w_c2 * channels; const float *grad_c2_ptr = grad_value_c2 + b_col * num_views * h_c2 * w_c2 * channels + loc_v * h_c2 * w_c2 * channels; ms_deform_attn_col2im_bilinear(feat_c2_ptr, h_c2, w_c2, channels, h_im, w_im, c_col, top_grad, weight_c2, grad_c2_ptr, grad_location_ptr, grad_weights_ptr); } grad_weights_ptr += 1; // C3 Feature h_im = loc_h * (h_c3 - 1); // align_corners = True w_im = loc_w * (w_c3 - 1); if (h_im > -1 && w_im > -1 && h_im < h_c3 && w_im < w_c3) { const float *feat_c3_ptr = feat_c3 + b_col * num_views * h_c3 * w_c3 * channels + loc_v * h_c3 * w_c3 * channels; const float *grad_c3_ptr = grad_value_c3 + b_col * num_views * h_c3 * w_c3 * channels + loc_v * h_c3 * w_c3 * channels; ms_deform_attn_col2im_bilinear(feat_c3_ptr, h_c3, w_c3, channels, h_im, w_im, c_col, top_grad, weight_c3, grad_c3_ptr, grad_location_ptr, grad_weights_ptr); } grad_weights_ptr += 1; // C4 Feature h_im = loc_h * (h_c4 - 1); // align_corners = True w_im = loc_w * (w_c4 - 1); if (h_im > -1 && w_im > -1 && h_im < h_c4 && w_im < w_c4) { const float *feat_c4_ptr = feat_c4 + b_col * num_views * h_c4 * w_c4 * channels + loc_v * h_c4 * w_c4 * channels; const float *grad_c4_ptr = grad_value_c4 + b_col * num_views * h_c4 * w_c4 * channels + loc_v * h_c4 * w_c4 * channels; ms_deform_attn_col2im_bilinear(feat_c4_ptr, h_c4, w_c4, channels, h_im, w_im, c_col, top_grad, weight_c4, grad_c4_ptr, grad_location_ptr, grad_weights_ptr); } grad_weights_ptr += 1; // C5 Feature h_im = loc_h * (h_c5 - 1); // align_corners = True w_im = loc_w * (w_c5 - 1); if (h_im > -1 && w_im > -1 && h_im < h_c5 && w_im < w_c5) { const float *feat_c5_ptr = feat_c5 + b_col * num_views * h_c5 * w_c5 * channels + loc_v * h_c5 * w_c5 * channels; const float *grad_c5_ptr = grad_value_c5 + b_col * num_views * h_c5 * w_c5 * channels + loc_v * h_c5 * w_c5 * channels; ms_deform_attn_col2im_bilinear(feat_c5_ptr, h_c5, w_c5, channels, h_im, w_im, c_col, top_grad, weight_c5, grad_c5_ptr, grad_location_ptr, grad_weights_ptr); } } } __global__ void ms_deformable_col2im_gpu_kernel_gm_c23456( const float *grad_col, const float *feat_c2, const float *feat_c3, const float *feat_c4, const float *feat_c5, const float *feat_c6, const int h_c2, const int w_c2, const int h_c3, const int w_c3, const int h_c4, const int w_c4, const int h_c5, const int w_c5, const int h_c6, const int w_c6, const float *data_sampling_loc, const float *data_attn_weight, const int batch_size, const int channels, const int num_views, const int num_query, const int num_point, float *grad_value_c2, float *grad_value_c3, float *grad_value_c4, float *grad_value_c5, float *grad_value_c6, float *grad_sampling_loc, float *grad_attn_weight) { CUDA_KERNEL_LOOP(index, batch_size * num_query * channels * num_point) { // n: bs x query x channels int _temp = index; const int p_col = _temp % num_point; _temp /= num_point; const int c_col = _temp % channels; _temp /= channels; const int sampling_index = _temp; _temp /= num_query; const int b_col = _temp; const float top_grad = grad_col[index]; // Sampling location in range [0, 1] int data_loc_ptr = sampling_index * num_point * 3 + p_col * 3; const float loc_w = data_sampling_loc[data_loc_ptr]; const float loc_h = data_sampling_loc[data_loc_ptr + 1]; const int loc_v = round(data_sampling_loc[data_loc_ptr + 2] * (num_views - 1)); // Attn weights int data_weight_ptr = sampling_index * num_point * 5 + p_col * 5; const float weight_c2 = data_attn_weight[data_weight_ptr]; const float weight_c3 = data_attn_weight[data_weight_ptr + 1]; const float weight_c4 = data_attn_weight[data_weight_ptr + 2]; const float weight_c5 = data_attn_weight[data_weight_ptr + 3]; const float weight_c6 = data_attn_weight[data_weight_ptr + 4]; // const float h_im = loc_h * spatial_h - 0.5; // align_corners = False // const float w_im = loc_w * spatial_w - 0.5; // C2 Feature float h_im = loc_h * (h_c2 - 1); // align_corners = True float w_im = loc_w * (w_c2 - 1); float *grad_location_ptr = grad_sampling_loc + data_loc_ptr; float *grad_weights_ptr = grad_attn_weight + data_weight_ptr; if (h_im > -1 && w_im > -1 && h_im < h_c2 && w_im < w_c2) { const float *feat_c2_ptr = feat_c2 + b_col * num_views * h_c2 * w_c2 * channels + loc_v * h_c2 * w_c2 * channels; const float *grad_c2_ptr = grad_value_c2 + b_col * num_views * h_c2 * w_c2 * channels + loc_v * h_c2 * w_c2 * channels; ms_deform_attn_col2im_bilinear(feat_c2_ptr, h_c2, w_c2, channels, h_im, w_im, c_col, top_grad, weight_c2, grad_c2_ptr, grad_location_ptr, grad_weights_ptr); } grad_weights_ptr += 1; // C3 Feature h_im = loc_h * (h_c3 - 1); // align_corners = True w_im = loc_w * (w_c3 - 1); if (h_im > -1 && w_im > -1 && h_im < h_c3 && w_im < w_c3) { const float *feat_c3_ptr = feat_c3 + b_col * num_views * h_c3 * w_c3 * channels + loc_v * h_c3 * w_c3 * channels; const float *grad_c3_ptr = grad_value_c3 + b_col * num_views * h_c3 * w_c3 * channels + loc_v * h_c3 * w_c3 * channels; ms_deform_attn_col2im_bilinear(feat_c3_ptr, h_c3, w_c3, channels, h_im, w_im, c_col, top_grad, weight_c3, grad_c3_ptr, grad_location_ptr, grad_weights_ptr); } grad_weights_ptr += 1; // C4 Feature h_im = loc_h * (h_c4 - 1); // align_corners = True w_im = loc_w * (w_c4 - 1); if (h_im > -1 && w_im > -1 && h_im < h_c4 && w_im < w_c4) { const float *feat_c4_ptr = feat_c4 + b_col * num_views * h_c4 * w_c4 * channels + loc_v * h_c4 * w_c4 * channels; const float *grad_c4_ptr = grad_value_c4 + b_col * num_views * h_c4 * w_c4 * channels + loc_v * h_c4 * w_c4 * channels; ms_deform_attn_col2im_bilinear(feat_c4_ptr, h_c4, w_c4, channels, h_im, w_im, c_col, top_grad, weight_c4, grad_c4_ptr, grad_location_ptr, grad_weights_ptr); } grad_weights_ptr += 1; // C5 Feature h_im = loc_h * (h_c5 - 1); // align_corners = True w_im = loc_w * (w_c5 - 1); if (h_im > -1 && w_im > -1 && h_im < h_c5 && w_im < w_c5) { const float *feat_c5_ptr = feat_c5 + b_col * num_views * h_c5 * w_c5 * channels + loc_v * h_c5 * w_c5 * channels; const float *grad_c5_ptr = grad_value_c5 + b_col * num_views * h_c5 * w_c5 * channels + loc_v * h_c5 * w_c5 * channels; ms_deform_attn_col2im_bilinear(feat_c5_ptr, h_c5, w_c5, channels, h_im, w_im, c_col, top_grad, weight_c5, grad_c5_ptr, grad_location_ptr, grad_weights_ptr); } grad_weights_ptr += 1; // C6 Feature h_im = loc_h * (h_c6 - 1); // align_corners = True w_im = loc_w * (w_c6 - 1); if (h_im > -1 && w_im > -1 && h_im < h_c6 && w_im < w_c6) { const float *feat_c6_ptr = feat_c6 + b_col * num_views * h_c6 * w_c6 * channels + loc_v * h_c6 * w_c6 * channels; const float *grad_c6_ptr = grad_value_c6 + b_col * num_views * h_c6 * w_c6 * channels + loc_v * h_c6 * w_c6 * channels; ms_deform_attn_col2im_bilinear(feat_c6_ptr, h_c6, w_c6, channels, h_im, w_im, c_col, top_grad, weight_c6, grad_c6_ptr, grad_location_ptr, grad_weights_ptr); } } } void ms_deformable_col2im_cuda_c2345( const float *grad_col, const float *feat_c2, const float *feat_c3, const float *feat_c4, const float *feat_c5, const int h_c2, const int w_c2, const int h_c3, const int w_c3, const int h_c4, const int w_c4, const int h_c5, const int w_c5, const float *data_sampling_loc, const float *data_attn_weight, const int batch_size, const int channels, const int num_views, const int num_query, const int num_point, float *grad_value_c2, float *grad_value_c3, float *grad_value_c4, float *grad_value_c5, float *grad_sampling_loc, float *grad_attn_weight) { const int num_kernels = batch_size * num_query * channels * num_point; const int num_threads = (channels * num_point > CUDA_NUM_THREADS) ? CUDA_NUM_THREADS : channels * num_point; ms_deformable_col2im_gpu_kernel_gm_c2345 <<>>( grad_col, feat_c2, feat_c3, feat_c4, feat_c5, h_c2, w_c2, h_c3, w_c3, h_c4, w_c4, h_c5, w_c5, data_sampling_loc, data_attn_weight, batch_size, channels, num_views, num_query, num_point, grad_value_c2, grad_value_c3, grad_value_c4, grad_value_c5, grad_sampling_loc, grad_attn_weight); cudaError_t err = cudaGetLastError(); if (err != cudaSuccess) { printf("error in ms_deformable_col2im_cuda_c2345: %s\n", cudaGetErrorString(err)); } } void ms_deformable_col2im_cuda_c23456( const float *grad_col, const float *feat_c2, const float *feat_c3, const float *feat_c4, const float *feat_c5, const float *feat_c6, const int h_c2, const int w_c2, const int h_c3, const int w_c3, const int h_c4, const int w_c4, const int h_c5, const int w_c5, const int h_c6, const int w_c6, const float *data_sampling_loc, const float *data_attn_weight, const int batch_size, const int channels, const int num_views, const int num_query, const int num_point, float *grad_value_c2, float *grad_value_c3, float *grad_value_c4, float *grad_value_c5, float *grad_value_c6, float *grad_sampling_loc, float *grad_attn_weight) { const int num_kernels = batch_size * num_query * channels * num_point; const int num_threads = (channels * num_point > CUDA_NUM_THREADS) ? CUDA_NUM_THREADS : channels * num_point; ms_deformable_col2im_gpu_kernel_gm_c23456 <<>>( grad_col, feat_c2, feat_c3, feat_c4, feat_c5, feat_c6, h_c2, w_c2, h_c3, w_c3, h_c4, w_c4, h_c5, w_c5, h_c6, w_c6, data_sampling_loc, data_attn_weight, batch_size, channels, num_views, num_query, num_point, grad_value_c2, grad_value_c3, grad_value_c4, grad_value_c5, grad_value_c6, grad_sampling_loc, grad_attn_weight); cudaError_t err = cudaGetLastError(); if (err != cudaSuccess) { printf("error in ms_deformable_col2im_cuda_c23456: %s\n", cudaGetErrorString(err)); } }