|
|
#include "convert.hpp" |
|
|
#include "dequantize.hpp" |
|
|
#include "presets.hpp" |
|
|
|
|
|
template <int qk, int qr, dequantize_kernel_t dequantize_kernel, typename dst_t> |
|
|
static void dequantize_block(const void * __restrict__ vx, dst_t * __restrict__ y, const int64_t k, |
|
|
const sycl::nd_item<3> &item_ct1) { |
|
|
const int64_t i = 2 * (item_ct1.get_local_range(2) * item_ct1.get_group(2) + |
|
|
item_ct1.get_local_id(2)); |
|
|
|
|
|
if (i >= k) { |
|
|
return; |
|
|
} |
|
|
|
|
|
const int64_t ib = i/qk; |
|
|
const int64_t iqs = (i%qk)/qr; |
|
|
const int64_t iybs = i - i%qk; |
|
|
const int64_t y_offset = qr == 1 ? 1 : qk/2; |
|
|
|
|
|
|
|
|
dfloat2 v; |
|
|
dequantize_kernel(vx, ib, iqs, v); |
|
|
|
|
|
y[iybs + iqs + 0] = v.x(); |
|
|
y[iybs + iqs + y_offset] = v.y(); |
|
|
} |
|
|
|
|
|
template <int qk, int qr, dequantize_kernel_t dequantize_kernel, typename dst_t> |
|
|
static void dequantize_block_sycl(const void *__restrict__ vx, |
|
|
dst_t *__restrict__ y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t num_blocks = (k + 2*SYCL_DEQUANTIZE_BLOCK_SIZE - 1) / (2*SYCL_DEQUANTIZE_BLOCK_SIZE); |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
sycl_parallel_for( |
|
|
stream, |
|
|
sycl::nd_range<3>(sycl::range<3>(1, 1, num_blocks) * sycl::range<3>(1, 1, SYCL_DEQUANTIZE_BLOCK_SIZE), |
|
|
sycl::range<3>(1, 1, SYCL_DEQUANTIZE_BLOCK_SIZE)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block<qk, qr, dequantize_kernel>(vx, y, k, item_ct1); }); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_q2_K_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
#if QK_K == 256 |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_parallel_for( |
|
|
stream, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 64), sycl::range<3>(1, 1, 64)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q2_K(vx, y, item_ct1); }); |
|
|
} |
|
|
#else |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_parallel_for( |
|
|
stream, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q2_K(vx, y, item_ct1); }); |
|
|
} |
|
|
|
|
|
#endif |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_q3_K_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
#if QK_K == 256 |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_parallel_for( |
|
|
stream, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 64), sycl::range<3>(1, 1, 64)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q3_K(vx, y, item_ct1); }); |
|
|
} |
|
|
#else |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_parallel_for( |
|
|
stream, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q3_K(vx, y, item_ct1); }); |
|
|
} |
|
|
#endif |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_q4_0_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb32 = k / 32; |
|
|
const int64_t nb = (k + 255) / 256; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_parallel_for( |
|
|
stream, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q4_0(vx, y, nb32, item_ct1); }); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_q4_0_sycl_reorder(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
|
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
int constexpr WARP_K = WARP_SIZE * QK4_0; |
|
|
const int n_warp = (k + WARP_K - 1) / WARP_K; |
|
|
GGML_ASSERT(k % 2 == 0); |
|
|
sycl_parallel_for(stream, |
|
|
sycl::nd_range<3>(sycl::range<3>(1, 1, n_warp) * sycl::range<3>(1, 1, WARP_SIZE), |
|
|
sycl::range<3>(1, 1, WARP_SIZE)), |
|
|
[=](sycl::nd_item<3> item_ct1) [[sycl::reqd_sub_group_size(WARP_SIZE)]] { |
|
|
dequantize_block_q4_0_reorder(vx, y, k, item_ct1); |
|
|
}); |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_q4_1_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb32 = k / 32; |
|
|
const int64_t nb = (k + 255) / 256; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_parallel_for( |
|
|
stream, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q4_1(vx, y, nb32, item_ct1); }); |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_q4_K_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl::local_accessor<uint8_t, 1> scale_local_acc(sycl::range<1>(12), cgh); |
|
|
sycl_parallel_for( |
|
|
cgh, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { |
|
|
dequantize_block_q4_K(vx, y, get_pointer(scale_local_acc), item_ct1); |
|
|
}); |
|
|
}); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_q4_K_sycl_reorder(const void * vx, dst_t * y, const int64_t k, dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
const size_t local_size = 32; |
|
|
const size_t global_size = nb * local_size; |
|
|
|
|
|
dpct::has_capability_or_fail(stream->get_device(), { sycl::aspect::fp16 }); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl::local_accessor<uint8_t, 1> scale_local_acc(sycl::range<1>(12), cgh); |
|
|
|
|
|
sycl_parallel_for<1>(cgh, sycl::nd_range<1>(sycl::range<1>(global_size), sycl::range<1>(local_size)), |
|
|
[=](sycl::nd_item<1> item_ct1) { |
|
|
dequantize_block_q4_K_reorder(vx, y, get_pointer(scale_local_acc), item_ct1, nb); |
|
|
}); |
|
|
}); |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_q5_K_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
#if QK_K == 256 |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_parallel_for( |
|
|
stream, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 64), sycl::range<3>(1, 1, 64)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q5_K(vx, y, item_ct1); }); |
|
|
} |
|
|
#else |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_parallel_for( |
|
|
stream, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q5_K(vx, y, item_ct1); }); |
|
|
} |
|
|
|
|
|
#endif |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_q6_K_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
#if QK_K == 256 |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_parallel_for( |
|
|
stream, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 64), sycl::range<3>(1, 1, 64)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q6_K(vx, y, item_ct1); }); |
|
|
} |
|
|
#else |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_parallel_for( |
|
|
stream, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q6_K(vx, y, item_ct1); }); |
|
|
} |
|
|
|
|
|
#endif |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_q6_K_sycl_reorder(const void * vx, dst_t * y, const int64_t k, dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
|
|
|
dpct::has_capability_or_fail(stream->get_device(), { sycl::aspect::fp16 }); |
|
|
|
|
|
sycl_parallel_for(stream, |
|
|
sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 64), sycl::range<3>(1, 1, 64)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_q6_K_reorder(vx, y, item_ct1, nb); }); |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_iq1_s_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl_parallel_for( |
|
|
cgh, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_iq1_s(vx, y, item_ct1, iq1s_grid_gpu); }); |
|
|
}); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_iq1_m_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl_parallel_for( |
|
|
cgh, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_iq1_m(vx, y, item_ct1, iq1s_grid_gpu); }); |
|
|
}); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_iq2_xxs_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl_parallel_for( |
|
|
cgh, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { |
|
|
dequantize_block_iq2_xxs(vx, y, item_ct1, iq2xxs_grid, ksigns_iq2xs, kmask_iq2xs); |
|
|
}); |
|
|
}); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_iq2_xs_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl_parallel_for( |
|
|
cgh, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { |
|
|
dequantize_block_iq2_xs(vx, y, item_ct1, iq2xs_grid, ksigns_iq2xs, kmask_iq2xs); |
|
|
}); |
|
|
}); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_iq2_s_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl_parallel_for( |
|
|
cgh, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_iq2_s(vx, y, item_ct1); }); |
|
|
}); |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_iq3_xxs_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl_parallel_for( |
|
|
cgh, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { |
|
|
dequantize_block_iq3_xxs(vx, y, item_ct1, iq3xxs_grid, ksigns_iq2xs, kmask_iq2xs); |
|
|
}); |
|
|
}); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_iq3_s_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = k / QK_K; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl_parallel_for( |
|
|
cgh, sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_iq3_s(vx, y, item_ct1, kmask_iq2xs, iq3s_grid); }); |
|
|
}); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_iq4_xs_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = (k + QK_K - 1) / QK_K; |
|
|
#if QK_K == 64 |
|
|
dequantize_row_iq4_nl_sycl(vx, y, k, stream); |
|
|
#else |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl_parallel_for( |
|
|
cgh, |
|
|
sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_iq4_xs(vx, y, item_ct1); }); |
|
|
}); |
|
|
} |
|
|
#endif |
|
|
} |
|
|
|
|
|
template <typename dst_t> |
|
|
static void dequantize_row_iq4_nl_sycl(const void *vx, dst_t *y, const int64_t k, |
|
|
dpct::queue_ptr stream) { |
|
|
const int64_t nb = (k + QK_K - 1) / QK_K; |
|
|
{ |
|
|
dpct::has_capability_or_fail(stream->get_device(), |
|
|
{sycl::aspect::fp16}); |
|
|
|
|
|
sycl_launch(stream, [&](sycl::handler & cgh) { |
|
|
sycl_parallel_for( |
|
|
cgh, |
|
|
sycl::nd_range<3>(sycl::range<3>(1, 1, nb) * sycl::range<3>(1, 1, 32), sycl::range<3>(1, 1, 32)), |
|
|
[=](sycl::nd_item<3> item_ct1) { dequantize_block_iq4_nl(vx, y, item_ct1); }); |
|
|
}); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename src_t, typename dst_t> |
|
|
static void convert_unary_nc(const void * __restrict__ vx, dst_t * __restrict__ y, const int64_t ne00, const int64_t ne01, |
|
|
const int64_t ne02, const int64_t s01, const int64_t s02, const int64_t s03, |
|
|
const sycl::nd_item<3> & item_ct1) { |
|
|
|
|
|
const int64_t work_group_size = item_ct1.get_local_range(2); |
|
|
const int64_t global_id = item_ct1.get_local_id(2) + work_group_size * item_ct1.get_group(2); |
|
|
|
|
|
const int64_t i01 = item_ct1.get_group(1); |
|
|
const int64_t i02 = item_ct1.get_group(0) % ne02; |
|
|
const int64_t i03 = item_ct1.get_group(0) / ne02; |
|
|
|
|
|
|
|
|
const src_t * x = static_cast<const src_t *>(vx); |
|
|
const int64_t ix = i03 * s03 + i02 * s02 + i01 * s01; |
|
|
const int64_t iy = ((i03 * ne02 + i02) * ne01 + i01) * ne00; |
|
|
|
|
|
#pragma unroll |
|
|
for (int64_t i00 = global_id; i00 < ne00; i00 += work_group_size * item_ct1.get_group_range(2)) { |
|
|
y[iy + i00] = static_cast<dst_t>(x[ix + i00]); |
|
|
} |
|
|
} |
|
|
|
|
|
template <typename src_t, typename dst_t> |
|
|
static void convert_unary_nc_sycl(const void * __restrict__ vx, dst_t * __restrict__ y, |
|
|
const int64_t ne00, const int64_t ne01, const int64_t ne02, const int64_t ne03, |
|
|
const int64_t s01, const int64_t s02, const int64_t s03, dpct::queue_ptr queue) { |
|
|
dpct::has_capability_or_fail(queue->get_device(), { sycl::aspect::fp16 }); |
|
|
|
|
|
sycl::range<3> global_size(ne02 * ne03, ne01, ceil_div(ne00, SYCL_DEQUANTIZE_BLOCK_SIZE)); |
|
|
|
|
|
|
|
|
|
|
|
int64_t downsized_workgroup = downsample_sycl_global_range(global_size[0], SYCL_DEQUANTIZE_BLOCK_SIZE); |
|
|
sycl::range<3> workgroup_size(1, 1, downsized_workgroup); |
|
|
|
|
|
queue->parallel_for(sycl::nd_range<3>(global_size * workgroup_size, workgroup_size), [=](sycl::nd_item<3> item_ct1) { |
|
|
convert_unary_nc<src_t>(vx, y, ne00, ne01, ne02, s01, s02, s03, item_ct1); |
|
|
}); |
|
|
} |
|
|
|
|
|
template <typename src_t, typename dst_t> |
|
|
static void convert_unary_sycl(const void * vx, dst_t * y, const int64_t k, dpct::queue_ptr queue) { |
|
|
convert_unary_nc_sycl<src_t>(vx, y, k, 1, 1, 1, k, k, k, queue); |
|
|
} |
|
|
|
|
|
to_fp16_sycl_t ggml_get_to_fp16_sycl(ggml_type type, ggml_tensor * dst) { |
|
|
switch (type) { |
|
|
case GGML_TYPE_Q4_0: |
|
|
if (dst->src[0]->extra && |
|
|
((ggml_tensor_extra_gpu*)dst->src[0]->extra)->optimized_feature.reorder) { |
|
|
return dequantize_row_q4_0_sycl_reorder; |
|
|
} else { |
|
|
return dequantize_block_sycl<QK4_0, QR4_0, dequantize_q4_0>; |
|
|
} |
|
|
case GGML_TYPE_Q4_1: |
|
|
return dequantize_block_sycl<QK4_1, QR4_1, dequantize_q4_1>; |
|
|
case GGML_TYPE_Q5_0: |
|
|
return dequantize_block_sycl<QK5_0, QR5_0, dequantize_q5_0>; |
|
|
case GGML_TYPE_Q5_1: |
|
|
return dequantize_block_sycl<QK5_1, QR5_1, dequantize_q5_1>; |
|
|
case GGML_TYPE_Q8_0: |
|
|
return dequantize_block_sycl<QK8_0, QR8_0, dequantize_q8_0>; |
|
|
case GGML_TYPE_Q2_K: |
|
|
return dequantize_row_q2_K_sycl; |
|
|
case GGML_TYPE_Q3_K: |
|
|
return dequantize_row_q3_K_sycl; |
|
|
case GGML_TYPE_Q4_K: |
|
|
if (dst->src[0]->extra && ((ggml_tensor_extra_gpu *) dst->src[0]->extra)->optimized_feature.reorder) { |
|
|
return dequantize_row_q4_K_sycl_reorder; |
|
|
} else { |
|
|
return dequantize_row_q4_K_sycl; |
|
|
} |
|
|
case GGML_TYPE_Q5_K: |
|
|
return dequantize_row_q5_K_sycl; |
|
|
case GGML_TYPE_Q6_K: |
|
|
if (dst->src[0]->extra && ((ggml_tensor_extra_gpu *) dst->src[0]->extra)->optimized_feature.reorder) { |
|
|
return dequantize_row_q6_K_sycl_reorder; |
|
|
} else { |
|
|
return dequantize_row_q6_K_sycl; |
|
|
} |
|
|
case GGML_TYPE_IQ1_S: |
|
|
return dequantize_row_iq1_s_sycl; |
|
|
case GGML_TYPE_IQ1_M: |
|
|
return dequantize_row_iq1_m_sycl; |
|
|
case GGML_TYPE_IQ2_XXS: |
|
|
return dequantize_row_iq2_xxs_sycl; |
|
|
case GGML_TYPE_IQ2_XS: |
|
|
return dequantize_row_iq2_xs_sycl; |
|
|
case GGML_TYPE_IQ2_S: |
|
|
return dequantize_row_iq2_s_sycl; |
|
|
case GGML_TYPE_IQ3_XXS: |
|
|
return dequantize_row_iq3_xxs_sycl; |
|
|
case GGML_TYPE_IQ3_S: |
|
|
return dequantize_row_iq3_s_sycl; |
|
|
case GGML_TYPE_IQ4_XS: |
|
|
return dequantize_row_iq4_xs_sycl; |
|
|
case GGML_TYPE_IQ4_NL: |
|
|
return dequantize_row_iq4_nl_sycl; |
|
|
case GGML_TYPE_F32: |
|
|
return convert_unary_sycl<float>; |
|
|
default: |
|
|
return nullptr; |
|
|
} |
|
|
} |
|
|
|
|
|
to_fp32_sycl_t ggml_get_to_fp32_sycl(ggml_type type, ggml_tensor *dst) { |
|
|
switch (type) { |
|
|
case GGML_TYPE_Q4_0: |
|
|
if (dst->src[0]->extra && |
|
|
((ggml_tensor_extra_gpu*)dst->src[0]->extra)->optimized_feature.reorder) { |
|
|
return dequantize_row_q4_0_sycl_reorder; |
|
|
} else { |
|
|
return dequantize_row_q4_0_sycl; |
|
|
} |
|
|
case GGML_TYPE_Q4_1: |
|
|
return dequantize_row_q4_1_sycl; |
|
|
case GGML_TYPE_Q5_0: |
|
|
return dequantize_block_sycl<QK5_0, QR5_0, dequantize_q5_0>; |
|
|
case GGML_TYPE_Q5_1: |
|
|
return dequantize_block_sycl<QK5_1, QR5_1, dequantize_q5_1>; |
|
|
case GGML_TYPE_Q8_0: |
|
|
return dequantize_block_sycl<QK8_0, QR8_0, dequantize_q8_0>; |
|
|
case GGML_TYPE_Q2_K: |
|
|
return dequantize_row_q2_K_sycl; |
|
|
case GGML_TYPE_Q3_K: |
|
|
return dequantize_row_q3_K_sycl; |
|
|
case GGML_TYPE_Q4_K: |
|
|
if (dst->src[0]->extra && |
|
|
((ggml_tensor_extra_gpu*)dst->src[0]->extra)->optimized_feature.reorder) { |
|
|
return dequantize_row_q4_K_sycl_reorder; |
|
|
} else { |
|
|
return dequantize_row_q4_K_sycl; |
|
|
} |
|
|
case GGML_TYPE_Q5_K: |
|
|
return dequantize_row_q5_K_sycl; |
|
|
case GGML_TYPE_Q6_K: |
|
|
if (dst->src[0]->extra && ((ggml_tensor_extra_gpu *) dst->src[0]->extra)->optimized_feature.reorder) { |
|
|
return dequantize_row_q6_K_sycl_reorder; |
|
|
} else { |
|
|
return dequantize_row_q6_K_sycl; |
|
|
} |
|
|
case GGML_TYPE_IQ1_S: |
|
|
return dequantize_row_iq1_s_sycl; |
|
|
case GGML_TYPE_IQ1_M: |
|
|
return dequantize_row_iq1_m_sycl; |
|
|
case GGML_TYPE_IQ2_XXS: |
|
|
return dequantize_row_iq2_xxs_sycl; |
|
|
case GGML_TYPE_IQ2_XS: |
|
|
return dequantize_row_iq2_xs_sycl; |
|
|
case GGML_TYPE_IQ2_S: |
|
|
return dequantize_row_iq2_s_sycl; |
|
|
case GGML_TYPE_IQ3_XXS: |
|
|
return dequantize_row_iq3_xxs_sycl; |
|
|
case GGML_TYPE_IQ3_S: |
|
|
return dequantize_row_iq3_s_sycl; |
|
|
case GGML_TYPE_IQ4_XS: |
|
|
return dequantize_row_iq4_xs_sycl; |
|
|
case GGML_TYPE_IQ4_NL: |
|
|
return dequantize_row_iq4_nl_sycl; |
|
|
case GGML_TYPE_F16: |
|
|
return convert_unary_sycl<sycl::half>; |
|
|
default: |
|
|
return nullptr; |
|
|
} |
|
|
} |
|
|
|
|
|
to_fp16_nc_sycl_t get_to_fp16_nc_sycl(ggml_type type) { |
|
|
switch (type) { |
|
|
case GGML_TYPE_F32: |
|
|
return convert_unary_nc_sycl<float>; |
|
|
default: |
|
|
return nullptr; |
|
|
} |
|
|
} |
|
|
|