| #include "ggml-metal.h" |
|
|
| #include "ggml-impl.h" |
| #include "ggml-backend-impl.h" |
|
|
| #include "ggml-metal-device.h" |
| #include "ggml-metal-context.h" |
| #include "ggml-metal-ops.h" |
|
|
| #include <mutex> |
| #include <string> |
|
|
| #define GGML_METAL_NAME "MTL" |
| #define GGML_METAL_MAX_DEVICES 16 |
|
|
| |
| |
| static int g_devices = 1; |
|
|
| |
| |
| |
|
|
| |
|
|
| static void ggml_backend_metal_buffer_shared_free_buffer(ggml_backend_buffer_t buffer) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(ggml_metal_buffer_is_shared(ctx)); |
|
|
| ggml_metal_buffer_free(ctx); |
| } |
|
|
| static void * ggml_backend_metal_buffer_shared_get_base(ggml_backend_buffer_t buffer) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(ggml_metal_buffer_is_shared(ctx)); |
|
|
| return ggml_metal_buffer_get_base(ctx); |
| } |
|
|
| static void ggml_backend_metal_buffer_shared_memset_tensor(ggml_backend_buffer_t buffer, ggml_tensor * tensor, uint8_t value, size_t offset, size_t size) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(ggml_metal_buffer_is_shared(ctx)); |
|
|
| ggml_metal_buffer_memset_tensor(ctx, tensor, value, offset, size); |
| } |
|
|
| static void ggml_backend_metal_buffer_shared_set_tensor(ggml_backend_buffer_t buffer, ggml_tensor * tensor, const void * data, size_t offset, size_t size) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(ggml_metal_buffer_is_shared(ctx)); |
|
|
| ggml_metal_buffer_set_tensor(ctx, tensor, data, offset, size); |
| } |
|
|
| static void ggml_backend_metal_buffer_shared_get_tensor(ggml_backend_buffer_t buffer, const ggml_tensor * tensor, void * data, size_t offset, size_t size) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(ggml_metal_buffer_is_shared(ctx)); |
|
|
| ggml_metal_buffer_get_tensor(ctx, tensor, data, offset, size); |
| } |
|
|
| static bool ggml_backend_metal_buffer_shared_cpy_tensor(ggml_backend_buffer_t buffer, const ggml_tensor * src, ggml_tensor * dst) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(ggml_metal_buffer_is_shared(ctx)); |
|
|
| GGML_UNUSED(buffer); |
| GGML_UNUSED(src); |
| GGML_UNUSED(dst); |
|
|
| return false; |
| } |
|
|
| static void ggml_backend_metal_buffer_shared_clear(ggml_backend_buffer_t buffer, uint8_t value) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(ggml_metal_buffer_is_shared(ctx)); |
|
|
| ggml_metal_buffer_clear(ctx, value); |
| } |
|
|
| static ggml_backend_buffer_i ggml_backend_metal_buffer_shared_i = { |
| ggml_backend_metal_buffer_shared_free_buffer, |
| ggml_backend_metal_buffer_shared_get_base, |
| NULL, |
| ggml_backend_metal_buffer_shared_memset_tensor, |
| ggml_backend_metal_buffer_shared_set_tensor, |
| ggml_backend_metal_buffer_shared_get_tensor, |
| ggml_backend_metal_buffer_shared_cpy_tensor, |
| ggml_backend_metal_buffer_shared_clear, |
| NULL, |
| }; |
|
|
| |
|
|
| static void ggml_backend_metal_buffer_private_free_buffer(ggml_backend_buffer_t buffer) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(!ggml_metal_buffer_is_shared(ctx)); |
|
|
| ggml_metal_buffer_free(ctx); |
| } |
|
|
| static void * ggml_backend_metal_buffer_private_get_base(ggml_backend_buffer_t buffer) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(!ggml_metal_buffer_is_shared(ctx)); |
|
|
| return ggml_metal_buffer_get_base(ctx); |
| } |
|
|
| static void ggml_backend_metal_buffer_private_memset_tensor(ggml_backend_buffer_t buffer, ggml_tensor * tensor, uint8_t value, size_t offset, size_t size) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(!ggml_metal_buffer_is_shared(ctx)); |
|
|
| ggml_metal_buffer_memset_tensor(ctx, tensor, value, offset, size); |
| } |
|
|
| static void ggml_backend_metal_buffer_private_set_tensor(ggml_backend_buffer_t buffer, ggml_tensor * tensor, const void * data, size_t offset, size_t size) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(!ggml_metal_buffer_is_shared(ctx)); |
|
|
| ggml_metal_buffer_set_tensor(ctx, tensor, data, offset, size); |
| } |
|
|
| static void ggml_backend_metal_buffer_private_get_tensor(ggml_backend_buffer_t buffer, const ggml_tensor * tensor, void * data, size_t offset, size_t size) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(!ggml_metal_buffer_is_shared(ctx)); |
|
|
| ggml_metal_buffer_get_tensor(ctx, tensor, data, offset, size); |
| } |
|
|
| static bool ggml_backend_metal_buffer_private_cpy_tensor(ggml_backend_buffer_t buffer, const ggml_tensor * src, ggml_tensor * dst) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(!ggml_metal_buffer_is_shared(ctx)); |
|
|
| GGML_UNUSED(buffer); |
| GGML_UNUSED(src); |
| GGML_UNUSED(dst); |
|
|
| return false; |
| } |
|
|
| static void ggml_backend_metal_buffer_private_clear(ggml_backend_buffer_t buffer, uint8_t value) { |
| ggml_metal_buffer_t ctx = (ggml_metal_buffer_t)buffer->context; |
|
|
| GGML_ASSERT(!ggml_metal_buffer_is_shared(ctx)); |
|
|
| ggml_metal_buffer_clear(ctx, value); |
| } |
|
|
| static ggml_backend_buffer_i ggml_backend_metal_buffer_private_i = { |
| ggml_backend_metal_buffer_private_free_buffer, |
| ggml_backend_metal_buffer_private_get_base, |
| NULL, |
| ggml_backend_metal_buffer_private_memset_tensor, |
| ggml_backend_metal_buffer_private_set_tensor, |
| ggml_backend_metal_buffer_private_get_tensor, |
| ggml_backend_metal_buffer_private_cpy_tensor, |
| ggml_backend_metal_buffer_private_clear, |
| NULL, |
| }; |
|
|
| static bool ggml_backend_buffer_is_metal(ggml_backend_buffer_t buffer) { |
| return buffer->iface.free_buffer == ggml_backend_metal_buffer_shared_free_buffer || |
| buffer->iface.free_buffer == ggml_backend_metal_buffer_private_free_buffer; |
| } |
|
|
| |
| |
| |
|
|
| struct ggml_backend_metal_buffer_type { |
| int device; |
| std::string name; |
| }; |
|
|
| struct ggml_backend_metal_buffer_type_deleter { |
| void operator()(ggml_backend_metal_buffer_type * ctx) const { |
| delete ctx; |
| } |
| }; |
|
|
| typedef std::unique_ptr<ggml_backend_metal_buffer_type, ggml_backend_metal_buffer_type_deleter> ggml_backend_metal_buffer_type_ptr; |
|
|
| |
| static ggml_backend_buffer_t ggml_backend_metal_buffer_type_alloc_buffer(ggml_backend_buffer_type_t buft, size_t size, bool shared) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)buft->device->context; |
| ggml_metal_buffer_t res = ggml_metal_buffer_init(ctx_dev, size, shared); |
|
|
| ggml_backend_buffer_i buf_i = ggml_metal_buffer_is_shared(res) |
| ? ggml_backend_metal_buffer_shared_i |
| : ggml_backend_metal_buffer_private_i; |
|
|
| return ggml_backend_buffer_init(buft, buf_i, res, size); |
| } |
|
|
| static size_t ggml_backend_metal_buffer_type_get_alloc_size(ggml_backend_buffer_type_t buft, const ggml_tensor * tensor) { |
| size_t res = ggml_nbytes(tensor); |
|
|
| |
| switch (tensor->op) { |
| case GGML_OP_MUL_MAT_ID: |
| { |
| res += ggml_metal_op_mul_mat_id_extra_tpe(tensor); |
| res += ggml_metal_op_mul_mat_id_extra_ids(tensor); |
| } break; |
| case GGML_OP_FLASH_ATTN_EXT: |
| { |
| res += ggml_metal_op_flash_attn_ext_extra_pad(tensor); |
| res += ggml_metal_op_flash_attn_ext_extra_blk(tensor); |
| res += ggml_metal_op_flash_attn_ext_extra_tmp(tensor); |
| } break; |
| case GGML_OP_CUMSUM: |
| case GGML_OP_ARGSORT: |
| { |
| res *= 2; |
| } break; |
| case GGML_OP_TOP_K: |
| { |
| res = 2*sizeof(int32_t)*ggml_nelements(tensor->src[0]); |
| } break; |
| default: |
| break; |
| } |
|
|
| return res; |
|
|
| GGML_UNUSED(buft); |
| } |
|
|
| |
|
|
| static const char * ggml_backend_metal_buffer_type_shared_get_name(ggml_backend_buffer_type_t buft) { |
| ggml_backend_metal_buffer_type * ctx = (ggml_backend_metal_buffer_type *)buft->context; |
|
|
| return ctx->name.c_str(); |
| } |
|
|
| static ggml_backend_buffer_t ggml_backend_metal_buffer_type_shared_alloc_buffer(ggml_backend_buffer_type_t buft, size_t size) { |
| return ggml_backend_metal_buffer_type_alloc_buffer(buft, size, true); |
| } |
|
|
| static size_t ggml_backend_metal_buffer_type_shared_get_alignment(ggml_backend_buffer_type_t buft) { |
| return 32; |
|
|
| GGML_UNUSED(buft); |
| } |
|
|
| static size_t ggml_backend_metal_buffer_type_shared_get_max_size(ggml_backend_buffer_type_t buft) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)buft->device->context; |
|
|
| return ggml_metal_device_get_props(ctx_dev)->max_buffer_size; |
| } |
|
|
| static size_t ggml_backend_metal_buffer_type_shared_get_alloc_size(ggml_backend_buffer_type_t buft, const ggml_tensor * tensor) { |
| return ggml_backend_metal_buffer_type_get_alloc_size(buft, tensor); |
| } |
|
|
| static bool ggml_backend_metal_buffer_type_shared_is_host(ggml_backend_buffer_type_t buft) { |
| return false; |
|
|
| GGML_UNUSED(buft); |
| } |
|
|
| static ggml_backend_buffer_type_t ggml_backend_metal_buffer_type_shared(int device) { |
| static std::mutex mutex; |
| std::lock_guard<std::mutex> lock(mutex); |
|
|
| static std::vector<ggml_backend_buffer_type> bufts; |
| static std::vector<ggml_backend_metal_buffer_type_ptr> ctxs; |
|
|
| static bool initialized = false; |
| if (!initialized) { |
| bufts.reserve(g_devices); |
| ctxs.reserve(g_devices); |
|
|
| for (int i = 0; i < g_devices; ++i) { |
| ggml_backend_metal_buffer_type * raw_ctx = |
| new ggml_backend_metal_buffer_type { |
| i, |
| GGML_METAL_NAME + std::to_string(i), |
| }; |
| ctxs.emplace_back(raw_ctx); |
|
|
| ggml_backend_buffer_type buft = { |
| { |
| ggml_backend_metal_buffer_type_shared_get_name, |
| ggml_backend_metal_buffer_type_shared_alloc_buffer, |
| ggml_backend_metal_buffer_type_shared_get_alignment, |
| ggml_backend_metal_buffer_type_shared_get_max_size, |
| ggml_backend_metal_buffer_type_shared_get_alloc_size, |
| ggml_backend_metal_buffer_type_shared_is_host, |
| }, |
| ggml_backend_reg_dev_get(ggml_backend_metal_reg(), i), |
| raw_ctx, |
| }; |
|
|
| bufts.emplace_back(buft); |
| } |
|
|
| initialized = true; |
| } |
|
|
| return &bufts[device]; |
| } |
|
|
| |
|
|
| static const char * ggml_backend_metal_buffer_type_private_get_name(ggml_backend_buffer_type_t buft) { |
| ggml_backend_metal_buffer_type * ctx = (ggml_backend_metal_buffer_type *)buft->context; |
|
|
| return ctx->name.c_str(); |
| } |
|
|
| static ggml_backend_buffer_t ggml_backend_metal_buffer_type_private_alloc_buffer(ggml_backend_buffer_type_t buft, size_t size) { |
| return ggml_backend_metal_buffer_type_alloc_buffer(buft, size, false); |
| } |
|
|
| static size_t ggml_backend_metal_buffer_type_private_get_alignment(ggml_backend_buffer_type_t buft) { |
| return 32; |
|
|
| GGML_UNUSED(buft); |
| } |
|
|
| static size_t ggml_backend_metal_buffer_type_private_get_max_size(ggml_backend_buffer_type_t buft) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)buft->device->context; |
|
|
| return ggml_metal_device_get_props(ctx_dev)->max_buffer_size; |
| } |
|
|
| static size_t ggml_backend_metal_buffer_type_private_get_alloc_size(ggml_backend_buffer_type_t buft, const ggml_tensor * tensor) { |
| return ggml_backend_metal_buffer_type_get_alloc_size(buft, tensor); |
| } |
|
|
| static bool ggml_backend_metal_buffer_type_private_is_host(ggml_backend_buffer_type_t buft) { |
| return false; |
|
|
| GGML_UNUSED(buft); |
| } |
|
|
| static ggml_backend_buffer_type_t ggml_backend_metal_buffer_type_private(int device) { |
| static std::mutex mutex; |
| std::lock_guard<std::mutex> lock(mutex); |
|
|
| static std::vector<ggml_backend_buffer_type> bufts; |
| static std::vector<ggml_backend_metal_buffer_type_ptr> ctxs; |
|
|
| static bool initialized = false; |
| if (!initialized) { |
| bufts.reserve(g_devices); |
| ctxs.reserve(g_devices); |
|
|
| for (int i = 0; i < g_devices; ++i) { |
| ggml_backend_metal_buffer_type * raw_ctx = new ggml_backend_metal_buffer_type{ |
| i, |
| GGML_METAL_NAME + std::to_string(i) + "_Private" |
| }; |
| ctxs.emplace_back(raw_ctx); |
|
|
| ggml_backend_buffer_type buft = { |
| { |
| ggml_backend_metal_buffer_type_private_get_name, |
| ggml_backend_metal_buffer_type_private_alloc_buffer, |
| ggml_backend_metal_buffer_type_private_get_alignment, |
| ggml_backend_metal_buffer_type_private_get_max_size, |
| ggml_backend_metal_buffer_type_private_get_alloc_size, |
| ggml_backend_metal_buffer_type_private_is_host, |
| }, |
| ggml_backend_reg_dev_get(ggml_backend_metal_reg(), i), |
| raw_ctx, |
| }; |
|
|
| bufts.emplace_back(buft); |
| } |
|
|
| initialized = true; |
| } |
|
|
| return &bufts[device]; |
| } |
|
|
| |
|
|
| static const char * ggml_backend_metal_buffer_type_mapped_get_name(ggml_backend_buffer_type_t buft) { |
| ggml_backend_metal_buffer_type * ctx = (ggml_backend_metal_buffer_type *)buft->context; |
|
|
| return ctx->name.c_str(); |
| } |
|
|
| static ggml_backend_buffer_t ggml_backend_metal_buffer_type_mapped_alloc_buffer(ggml_backend_buffer_type_t buft, size_t size) { |
| |
| return ggml_backend_metal_buffer_type_alloc_buffer(buft, size, true); |
| } |
|
|
| static size_t ggml_backend_metal_buffer_type_mapped_get_alignment(ggml_backend_buffer_type_t buft) { |
| return 32; |
|
|
| GGML_UNUSED(buft); |
| } |
|
|
| static size_t ggml_backend_metal_buffer_type_mapped_get_max_size(ggml_backend_buffer_type_t buft) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)buft->device->context; |
|
|
| return ggml_metal_device_get_props(ctx_dev)->max_buffer_size; |
| } |
|
|
| static size_t ggml_backend_metal_buffer_type_mapped_get_alloc_size(ggml_backend_buffer_type_t buft, const ggml_tensor * tensor) { |
| return ggml_backend_metal_buffer_type_get_alloc_size(buft, tensor); |
| } |
|
|
| static bool ggml_backend_metal_buffer_type_mapped_is_host(ggml_backend_buffer_type_t buft) { |
| return false; |
|
|
| GGML_UNUSED(buft); |
| } |
|
|
| static ggml_backend_buffer_type_t ggml_backend_metal_buffer_type_mapped(int device) { |
| static std::mutex mutex; |
| std::lock_guard<std::mutex> lock(mutex); |
|
|
| static std::vector<ggml_backend_buffer_type> bufts; |
| static std::vector<ggml_backend_metal_buffer_type_ptr> ctxs; |
|
|
| static bool initialized = false; |
| if (!initialized) { |
| bufts.reserve(g_devices); |
| ctxs.reserve(g_devices); |
|
|
| for (int i = 0; i < g_devices; ++i) { |
| ggml_backend_metal_buffer_type * raw_ctx = new ggml_backend_metal_buffer_type{ |
| i, |
| GGML_METAL_NAME + std::to_string(i) + "_Mapped" |
| }; |
| ctxs.emplace_back(raw_ctx); |
|
|
| |
| |
| ggml_backend_buffer_type buft = { |
| { |
| ggml_backend_metal_buffer_type_mapped_get_name, |
| ggml_backend_metal_buffer_type_mapped_alloc_buffer, |
| ggml_backend_metal_buffer_type_mapped_get_alignment, |
| ggml_backend_metal_buffer_type_mapped_get_max_size, |
| ggml_backend_metal_buffer_type_mapped_get_alloc_size, |
| ggml_backend_metal_buffer_type_mapped_is_host, |
| }, |
| ggml_backend_reg_dev_get(ggml_backend_metal_reg(), i), |
| raw_ctx, |
| }; |
|
|
| bufts.emplace_back(buft); |
| } |
|
|
| initialized = true; |
| } |
|
|
| return &bufts[device]; |
| } |
|
|
| |
|
|
| static const char * ggml_backend_metal_name(ggml_backend_t backend) { |
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| return ggml_metal_get_name(ctx); |
| } |
|
|
| static void ggml_backend_metal_free(ggml_backend_t backend) { |
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| |
| ggml_metal_synchronize(ctx); |
|
|
| ggml_metal_free(ctx); |
|
|
| free(backend); |
| } |
|
|
| static void ggml_backend_metal_synchronize(ggml_backend_t backend) { |
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| ggml_metal_synchronize(ctx); |
| } |
|
|
| static void ggml_backend_metal_set_tensor_async(ggml_backend_t backend, ggml_tensor * tensor, const void * data, size_t offset, size_t size) { |
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| ggml_metal_set_tensor_async(ctx, tensor, data, offset, size); |
| } |
|
|
| static void ggml_backend_metal_get_tensor_async(ggml_backend_t backend, const ggml_tensor * tensor, void * data, size_t offset, size_t size) { |
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| ggml_metal_get_tensor_async(ctx, tensor, data, offset, size); |
| } |
|
|
| static bool ggml_backend_metal_cpy_tensor_async(ggml_backend_t backend_src, ggml_backend_t backend_dst, const ggml_tensor * src, ggml_tensor * dst) { |
| if (!ggml_backend_is_metal(backend_src) || !ggml_backend_is_metal(backend_dst)) { |
| return false; |
| } |
|
|
| if (!ggml_backend_buffer_is_metal(src->buffer) || !ggml_backend_buffer_is_metal(dst->buffer)) { |
| return false; |
| } |
|
|
| ggml_metal_t ctx_src = (ggml_metal_t)backend_src->context; |
| ggml_metal_t ctx_dst = (ggml_metal_t)backend_dst->context; |
|
|
| |
| |
|
|
| |
| |
|
|
| return ggml_metal_cpy_tensor_async(ctx_src, ctx_dst, src, dst); |
| } |
|
|
| static enum ggml_status ggml_backend_metal_graph_compute(ggml_backend_t backend, ggml_cgraph * cgraph) { |
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| return ggml_metal_graph_compute(ctx, cgraph); |
| } |
|
|
| static void ggml_backend_metal_event_record(ggml_backend_t backend, ggml_backend_event_t event) { |
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
| ggml_metal_event_t ev = (ggml_metal_event_t)event->context; |
|
|
| ggml_metal_event_record(ctx, ev); |
| } |
|
|
| static void ggml_backend_metal_event_wait(ggml_backend_t backend, ggml_backend_event_t event) { |
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
| ggml_metal_event_t ev = (ggml_metal_event_t)event->context; |
|
|
| ggml_metal_event_wait(ctx, ev); |
| } |
|
|
| static void ggml_backend_metal_graph_optimize(ggml_backend_t backend, ggml_cgraph * cgraph) { |
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| ggml_metal_graph_optimize(ctx, cgraph); |
| } |
|
|
| static void ggml_backend_metal_set_n_cb(ggml_backend_t backend, int n_cb) { |
| GGML_ASSERT(ggml_backend_is_metal(backend)); |
|
|
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| ggml_metal_set_n_cb(ctx, n_cb); |
| } |
|
|
| static ggml_backend_i ggml_backend_metal_i = { |
| ggml_backend_metal_name, |
| ggml_backend_metal_free, |
| ggml_backend_metal_set_tensor_async, |
| ggml_backend_metal_get_tensor_async, |
| ggml_backend_metal_cpy_tensor_async, |
| ggml_backend_metal_synchronize, |
| NULL, |
| NULL, |
| NULL, |
| NULL, |
| ggml_backend_metal_graph_compute, |
| ggml_backend_metal_event_record, |
| ggml_backend_metal_event_wait, |
| ggml_backend_metal_graph_optimize, |
| }; |
|
|
| static ggml_guid_t ggml_backend_metal_guid(void) { |
| static ggml_guid guid = { 0x81, 0xa1, 0x8b, 0x1e, 0x71, 0xec, 0x79, 0xed, 0x2b, 0x85, 0xdc, 0x8a, 0x61, 0x98, 0x30, 0xe6 }; |
| return &guid; |
| } |
|
|
| ggml_backend_t ggml_backend_metal_init(void) { |
| ggml_backend_dev_t dev = ggml_backend_reg_dev_get(ggml_backend_metal_reg(), 0); |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| ggml_metal_t ctx = ggml_metal_init(ctx_dev); |
| if (ctx == NULL) { |
| GGML_LOG_ERROR("%s: error: failed to allocate context\n", __func__); |
| return NULL; |
| } |
|
|
| ggml_backend_t backend = (ggml_backend_t) malloc(sizeof(ggml_backend)); |
|
|
| *backend = { |
| ggml_backend_metal_guid(), |
| ggml_backend_metal_i, |
| dev, |
| ctx, |
| }; |
|
|
| ggml_backend_metal_set_n_cb(backend, 1); |
|
|
| return backend; |
| } |
|
|
| bool ggml_backend_is_metal(ggml_backend_t backend) { |
| return backend != NULL && ggml_guid_matches(backend->guid, ggml_backend_metal_guid()); |
| } |
|
|
| void ggml_backend_metal_set_abort_callback(ggml_backend_t backend, ggml_abort_callback abort_callback, void * user_data) { |
| GGML_ASSERT(ggml_backend_is_metal(backend)); |
|
|
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| ggml_metal_set_abort_callback(ctx, abort_callback, user_data); |
| } |
|
|
| bool ggml_backend_metal_supports_family(ggml_backend_t backend, int family) { |
| GGML_ASSERT(ggml_backend_is_metal(backend)); |
|
|
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| return ggml_metal_supports_family(ctx, family); |
| } |
|
|
| void ggml_backend_metal_capture_next_compute(ggml_backend_t backend) { |
| GGML_ASSERT(ggml_backend_is_metal(backend)); |
|
|
| ggml_metal_t ctx = (ggml_metal_t)backend->context; |
|
|
| ggml_metal_capture_next_compute(ctx); |
| } |
|
|
| |
|
|
| static const char * ggml_backend_metal_device_get_name(ggml_backend_dev_t dev) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| const ggml_metal_device_props * props_dev = ggml_metal_device_get_props(ctx_dev); |
|
|
| return props_dev->name; |
| } |
|
|
| static const char * ggml_backend_metal_device_get_description(ggml_backend_dev_t dev) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| return ggml_metal_device_get_props(ctx_dev)->desc; |
| } |
|
|
| static void ggml_backend_metal_device_get_memory(ggml_backend_dev_t dev, size_t * free, size_t * total) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| ggml_metal_device_get_memory(ctx_dev, free, total); |
| } |
|
|
| static enum ggml_backend_dev_type ggml_backend_metal_device_get_type(ggml_backend_dev_t dev) { |
| return GGML_BACKEND_DEVICE_TYPE_GPU; |
|
|
| GGML_UNUSED(dev); |
| } |
|
|
| static void ggml_backend_metal_device_get_props(ggml_backend_dev_t dev, ggml_backend_dev_props * props) { |
| props->name = ggml_backend_metal_device_get_name(dev); |
| props->description = ggml_backend_metal_device_get_description(dev); |
| props->type = ggml_backend_metal_device_get_type(dev); |
|
|
| ggml_backend_metal_device_get_memory(dev, &props->memory_free, &props->memory_total); |
|
|
| props->caps = { |
| true, |
| false, |
| true, |
| true, |
| }; |
| } |
|
|
| static ggml_backend_t ggml_backend_metal_device_init_backend(ggml_backend_dev_t dev, const char * params) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| ggml_metal_t ctx = ggml_metal_init(ctx_dev); |
| if (ctx == NULL) { |
| GGML_LOG_ERROR("%s: error: failed to allocate context\n", __func__); |
| return NULL; |
| } |
|
|
| ggml_backend_t backend = (ggml_backend_t) malloc(sizeof(ggml_backend)); |
|
|
| *backend = { |
| ggml_backend_metal_guid(), |
| ggml_backend_metal_i, |
| dev, |
| ctx, |
| }; |
|
|
| ggml_backend_metal_set_n_cb(backend, 1); |
|
|
| return backend; |
|
|
| GGML_UNUSED(params); |
| } |
|
|
| static ggml_backend_buffer_type_t ggml_backend_metal_device_get_buffer_type(ggml_backend_dev_t dev) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| const ggml_metal_device_props * props_dev = ggml_metal_device_get_props(ctx_dev); |
|
|
| return props_dev->use_shared_buffers ? ggml_backend_metal_buffer_type_shared(props_dev->device) : ggml_backend_metal_buffer_type_private(props_dev->device); |
| } |
|
|
| static ggml_backend_buffer_t ggml_backend_metal_device_buffer_mapped(ggml_backend_dev_t dev, void * ptr, size_t size, size_t max_tensor_size) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| ggml_metal_buffer_t res = ggml_metal_buffer_map(ctx_dev, ptr, size, max_tensor_size); |
|
|
| const ggml_metal_device_props * props_dev = ggml_metal_device_get_props(ctx_dev); |
|
|
| return ggml_backend_buffer_init(ggml_backend_metal_buffer_type_mapped(props_dev->device), ggml_backend_metal_buffer_shared_i, res, size); |
| } |
|
|
| static bool ggml_backend_metal_device_supports_op(ggml_backend_dev_t dev, const ggml_tensor * op) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| return ggml_metal_device_supports_op(ctx_dev, op); |
| } |
|
|
| static bool ggml_backend_metal_device_supports_buft(ggml_backend_dev_t dev, ggml_backend_buffer_type_t buft) { |
| return |
| buft->device == dev && ( |
| buft->iface.get_name == ggml_backend_metal_buffer_type_shared_get_name || |
| buft->iface.get_name == ggml_backend_metal_buffer_type_private_get_name || |
| buft->iface.get_name == ggml_backend_metal_buffer_type_mapped_get_name); |
|
|
| GGML_UNUSED(dev); |
| } |
|
|
| static int64_t get_op_batch_size(const ggml_tensor * op) { |
| switch (op->op) { |
| case GGML_OP_MUL_MAT: |
| return op->ne[1]; |
| case GGML_OP_MUL_MAT_ID: |
| return op->ne[2]; |
| default: |
| return ggml_nrows(op); |
| } |
| } |
|
|
| static bool ggml_backend_metal_device_offload_op(ggml_backend_dev_t dev, const ggml_tensor * op) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| return (op->op == GGML_OP_MUL_MAT || |
| op->op == GGML_OP_MUL_MAT_ID) && |
| get_op_batch_size(op) >= ggml_metal_device_get_props(ctx_dev)->op_offload_min_batch_size; |
| } |
|
|
| static ggml_backend_event_t ggml_backend_metal_device_event_new(ggml_backend_dev_t dev) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| ggml_metal_event_t event = ggml_metal_device_event_init(ctx_dev); |
| GGML_ASSERT(event); |
|
|
| ggml_backend_event_t ev = new ggml_backend_event { |
| dev, |
| event, |
| }; |
|
|
| return ev; |
| } |
|
|
| static void ggml_backend_metal_device_event_free(ggml_backend_dev_t dev, ggml_backend_event_t event) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| ggml_metal_event_t ev = (ggml_metal_event_t)event->context; |
|
|
| ggml_metal_device_event_free(ctx_dev, ev); |
|
|
| delete event; |
| } |
|
|
| static void ggml_backend_metal_device_event_synchronize(ggml_backend_dev_t dev, ggml_backend_event_t event) { |
| ggml_metal_device_t ctx_dev = (ggml_metal_device_t)dev->context; |
|
|
| ggml_metal_event_t evt = (ggml_metal_event_t)event->context; |
|
|
| ggml_metal_device_event_synchronize(ctx_dev, evt); |
| } |
|
|
| static ggml_backend_device_i ggml_backend_metal_device_i = { |
| ggml_backend_metal_device_get_name, |
| ggml_backend_metal_device_get_description, |
| ggml_backend_metal_device_get_memory, |
| ggml_backend_metal_device_get_type, |
| ggml_backend_metal_device_get_props, |
| ggml_backend_metal_device_init_backend, |
| ggml_backend_metal_device_get_buffer_type, |
| NULL, |
| ggml_backend_metal_device_buffer_mapped, |
| ggml_backend_metal_device_supports_op, |
| ggml_backend_metal_device_supports_buft, |
| ggml_backend_metal_device_offload_op, |
| ggml_backend_metal_device_event_new, |
| ggml_backend_metal_device_event_free, |
| ggml_backend_metal_device_event_synchronize, |
| }; |
|
|
| |
|
|
| struct ggml_backend_metal_reg { |
| std::vector<ggml_backend_dev_t> devices; |
| }; |
|
|
| typedef struct ggml_backend_metal_reg * ggml_backend_metal_reg_t; |
|
|
| static ggml_backend_metal_reg_t ggml_backend_metal_reg_init(void) { |
| ggml_backend_metal_reg_t ctx = new struct ggml_backend_metal_reg; |
|
|
| return ctx; |
| } |
|
|
| static void ggml_backend_metal_reg_free(ggml_backend_metal_reg_t ctx) { |
| delete ctx; |
| } |
|
|
| struct ggml_backend_metal_reg_deleter { |
| void operator()(ggml_backend_metal_reg_t ctx) { |
| ggml_backend_metal_reg_free(ctx); |
| } |
| }; |
|
|
| typedef std::unique_ptr<struct ggml_backend_metal_reg, ggml_backend_metal_reg_deleter> ggml_backend_metal_reg_ptr; |
|
|
| static const char * ggml_backend_metal_reg_get_name(ggml_backend_reg_t reg) { |
| return GGML_METAL_NAME; |
|
|
| GGML_UNUSED(reg); |
| } |
|
|
| static size_t ggml_backend_metal_reg_device_count(ggml_backend_reg_t reg) { |
| ggml_backend_metal_reg_t ctx = (ggml_backend_metal_reg_t)reg->context; |
| return ctx->devices.size(); |
| } |
|
|
| static ggml_backend_dev_t ggml_backend_metal_reg_device_get(ggml_backend_reg_t reg, size_t index) { |
| ggml_backend_metal_reg_t ctx = (ggml_backend_metal_reg_t)reg->context; |
| GGML_ASSERT(index < ctx->devices.size()); |
| return ctx->devices[index]; |
| } |
|
|
| static ggml_backend_feature g_ggml_backend_metal_features[] = { |
| #if defined(GGML_METAL_EMBED_LIBRARY) |
| { "EMBED_LIBRARY", "1" }, |
| #endif |
| { NULL, NULL }, |
| }; |
|
|
| static ggml_backend_feature * ggml_backend_metal_get_features(ggml_backend_reg_t reg) { |
| return g_ggml_backend_metal_features; |
|
|
| GGML_UNUSED(reg); |
| } |
|
|
| static void * ggml_backend_metal_get_proc_address(ggml_backend_reg_t reg, const char * name) { |
| if (strcmp(name, "ggml_backend_get_features") == 0) { |
| return (void *)ggml_backend_metal_get_features; |
| } |
|
|
| return NULL; |
|
|
| GGML_UNUSED(reg); |
| } |
|
|
| static ggml_backend_reg_i ggml_backend_metal_reg_i = { |
| ggml_backend_metal_reg_get_name, |
| ggml_backend_metal_reg_device_count, |
| ggml_backend_metal_reg_device_get, |
| ggml_backend_metal_get_proc_address, |
| }; |
|
|
| static ggml_backend_dev_t ggml_backend_metal_device_init(ggml_backend_reg_t reg, int device) { |
| return new ggml_backend_device { |
| ggml_backend_metal_device_i, |
| reg, |
| ggml_metal_device_get(device), |
| }; |
| } |
|
|
| static void ggml_backend_metal_device_free(ggml_backend_dev_t dev) { |
| delete dev; |
| } |
|
|
| struct ggml_backend_device_deleter { |
| void operator()(ggml_backend_dev_t ctx) { |
| ggml_backend_metal_device_free(ctx); |
| } |
| }; |
|
|
| typedef std::unique_ptr<ggml_backend_device, ggml_backend_device_deleter> ggml_backend_device_ptr; |
|
|
| ggml_backend_reg_t ggml_backend_metal_reg(void) { |
| static ggml_backend_reg reg; |
| static bool initialized = false; |
|
|
| { |
| static std::mutex mutex; |
| std::lock_guard<std::mutex> lock(mutex); |
|
|
| const char * env = getenv("GGML_METAL_DEVICES"); |
| if (env) { |
| g_devices = atoi(env); |
| } |
|
|
| static std::vector<ggml_backend_device_ptr> devs; |
|
|
| if (!initialized) { |
| static ggml_backend_metal_reg_ptr reg_ctx(ggml_backend_metal_reg_init()); |
|
|
| for (int i = 0; i < g_devices; ++i) { |
| auto * dev = ggml_backend_metal_device_init(®, i); |
| devs.emplace_back(dev); |
|
|
| reg_ctx->devices.push_back(dev); |
| } |
|
|
| reg = { |
| GGML_BACKEND_API_VERSION, |
| ggml_backend_metal_reg_i, |
| reg_ctx.get(), |
| }; |
| } |
|
|
| initialized = true; |
| } |
|
|
| return ® |
| } |
|
|
| GGML_BACKEND_DL_IMPL(ggml_backend_metal_reg) |
|
|