| #version 450 |
| #extension GL_EXT_control_flow_attributes : enable |
| #extension GL_KHR_memory_scope_semantics : enable |
| #pragma use_vulkan_memory_model |
|
|
| #include "types.glsl" |
|
|
| layout(constant_id = 0) const int BLOCK_SIZE = 1024; |
| layout(constant_id = 1) const int WG_UNROLL_FACTOR = 2; |
| #define ASC 0 |
|
|
| layout(local_size_x_id = 0, local_size_y = 1, local_size_z = 1) in; |
|
|
| layout (binding = 0) readonly buffer A {A_TYPE data_a[];}; |
| layout (binding = 1) workgroupcoherent buffer B {ivec2 tmp_idx[];}; |
| layout (binding = 2) workgroupcoherent buffer D {int data_d[];}; |
|
|
| layout (push_constant) uniform parameter { |
| uint ncols; |
| uint ncols_padded; |
| uint ncols_padded_log2; |
| uint nrows; |
| uint order; |
| uint outer_start; |
| uint outer_end; |
| uint inner_start; |
| uint inner_end; |
| } p; |
|
|
| void argsort(bool needs_bounds_check, const uint row) { |
| |
| int col = int(gl_GlobalInvocationID.x); |
| col = (col % BLOCK_SIZE) + (col / BLOCK_SIZE) * BLOCK_SIZE * WG_UNROLL_FACTOR; |
|
|
| const uint row_offset = row * p.ncols; |
| uint idx_offset = row * p.ncols_padded; |
|
|
| bool need_barrier = false; |
|
|
| |
| if (p.outer_start == 0 && p.inner_start == 0) { |
| [[unroll]] for (int u = 0; u < WG_UNROLL_FACTOR; ++u) { |
| uint c = u*BLOCK_SIZE + col; |
| if (c < p.ncols_padded) { |
| ivec2 v = ivec2(c, floatBitsToInt(data_a[row_offset + c])); |
| tmp_idx[idx_offset + c] = v; |
| } |
| } |
| need_barrier = true; |
| } |
|
|
| [[unroll]] for (uint outer_idx = p.outer_start, k = (2 << outer_idx); outer_idx < p.outer_end; k *= 2, outer_idx++) { |
| uint inner_end = min(p.inner_end, outer_idx + 1); |
| for (uint j = k >> (p.inner_start + 1), inner_idx = p.inner_start; inner_idx < inner_end; j /= 2, inner_idx++) { |
| if (need_barrier) { |
| controlBarrier(gl_ScopeWorkgroup, gl_ScopeWorkgroup, gl_StorageSemanticsBuffer, gl_SemanticsAcquireRelease); |
| } |
| need_barrier = true; |
| [[unroll]] for (int u = 0; u < WG_UNROLL_FACTOR; ++u) { |
| int c = u*BLOCK_SIZE + col; |
| const int ixj = int(c ^ j); |
|
|
| if (ixj < c) { |
| continue; |
| } |
|
|
| int idx_0 = (c & k) == 0 ? c : ixj; |
| int idx_1 = (c & k) == 0 ? ixj : c; |
|
|
| ivec2 sh_idx_0 = tmp_idx[idx_offset + idx_0]; |
| ivec2 sh_idx_1 = tmp_idx[idx_offset + idx_1]; |
| bool idx_0_oob = needs_bounds_check ? sh_idx_0.x >= p.ncols : false; |
| bool idx_1_oob = needs_bounds_check ? sh_idx_1.x >= p.ncols : false; |
|
|
| if ((idx_0_oob || |
| (!idx_1_oob && intBitsToFloat(sh_idx_0.y) > intBitsToFloat(sh_idx_1.y)))) { |
| tmp_idx[idx_offset + idx_0] = sh_idx_1; |
| tmp_idx[idx_offset + idx_1] = sh_idx_0; |
| } |
| } |
| } |
| } |
|
|
| if (p.outer_end == p.ncols_padded_log2 && |
| p.inner_end >= p.ncols_padded_log2 + 1) { |
| controlBarrier(gl_ScopeWorkgroup, gl_ScopeWorkgroup, gl_StorageSemanticsBuffer, gl_SemanticsAcquireRelease); |
| [[unroll]] for (int u = 0; u < WG_UNROLL_FACTOR; ++u) { |
| uint c = u*BLOCK_SIZE + col; |
| if (c < p.ncols) { |
| if (p.order == ASC) { |
| data_d[row_offset + c] = tmp_idx[idx_offset + c].x; |
| } else { |
| data_d[row_offset + p.ncols - c - 1] = tmp_idx[idx_offset + c].x; |
| } |
| } |
| } |
| } |
| } |
|
|
| void main() { |
| if (p.ncols == p.ncols_padded) { |
| uint row = gl_WorkGroupID.y; |
| while (row < p.nrows) { |
| argsort(false, row); |
| row += gl_WorkGroupSize.y * gl_NumWorkGroups.y; |
| } |
| } else { |
| uint row = gl_WorkGroupID.y; |
| while (row < p.nrows) { |
| argsort(true, row); |
| row += gl_WorkGroupSize.y * gl_NumWorkGroups.y; |
| } |
| } |
| } |
|
|